FAST  3.2.0
Framework for Heterogeneous Medical Image Computing and Visualization
TensorRTEngine.hpp
Go to the documentation of this file.
1 #pragma once
2 
4 #include <TensorRTExport.hpp>
5 
6 namespace nvinfer1 {
7 class ICudaEngine;
8 class IExecutionContext;
9 }
10 
11 namespace fast {
12 
13 class INFERENCEENGINETENSORRT_EXPORT TensorRTEngine : public InferenceEngine {
15  public:
16  void run() override;
17  void load() override;
18  ImageOrdering getPreferredImageOrdering() const override;
19  std::string getName() const override;
20  std::vector<ModelFormat> getSupportedModelFormats() const {
22  };
24  return ModelFormat::ONNX;
25  };
26  ~TensorRTEngine() override;
27  void setMaxBatchSize(int maxBathSize);
28  int getMaxBatchSize() const;
30  private:
31  nvinfer1::ICudaEngine* m_engine = nullptr;
32  nvinfer1::IExecutionContext* m_context = nullptr;
33  std::size_t m_maxWorkspaceSize = 512*1024*1024; // in bytes
34  ImageOrdering m_dimensionOrdering;
35 };
36 
37 DEFINE_INFERENCE_ENGINE(TensorRTEngine, INFERENCEENGINETENSORRT_EXPORT)
38 
39 }
InferenceEngine
Definition: OpenVINOEngine.hpp:6
fast::TensorRTEngine
Definition: TensorRTEngine.hpp:13
fast::ModelFormat::UFF
@ UFF
fast
Definition: AffineTransformation.hpp:7
fast::TensorRTEngine::getPreferredModelFormat
ModelFormat getPreferredModelFormat() const
Definition: TensorRTEngine.hpp:23
nvinfer1
Definition: TensorRTEngine.hpp:6
FAST_OBJECT
#define FAST_OBJECT(className)
Definition: Object.hpp:9
fast::ModelFormat
ModelFormat
Definition: InferenceEngine.hpp:48
fast::TensorRTEngine::getSupportedModelFormats
std::vector< ModelFormat > getSupportedModelFormats() const
Definition: TensorRTEngine.hpp:20
DEFINE_INFERENCE_ENGINE
#define DEFINE_INFERENCE_ENGINE(classType, exportStatement)
Definition: InferenceEngine.hpp:10
fast::ModelFormat::ONNX
@ ONNX
InferenceEngine.hpp
fast::ImageOrdering
ImageOrdering
Definition: InferenceEngine.hpp:21