TensorRT  7.2.1.6
NVIDIA TensorRT
Looking for a C++ dev who knows TensorRT?
I'm looking for work. Hire me!
nvinfer1::IInt8MinMaxCalibrator Class Referenceabstract

MinMax Calibrator. More...

Inheritance diagram for nvinfer1::IInt8MinMaxCalibrator:
Collaboration diagram for nvinfer1::IInt8MinMaxCalibrator:

Public Member Functions

CalibrationAlgoType getAlgorithm () override
 Signal that this is the MinMax Calibrator. More...
 
virtual ~IInt8MinMaxCalibrator ()
 
virtual int32_t getBatchSize () const =0
 Get the batch size used for calibration batches. More...
 
virtual bool getBatch (void *bindings[], const char *names[], int32_t nbBindings)=0
 Get a batch of input for calibration. More...
 
virtual const void * readCalibrationCache (std::size_t &length)=0
 Load a calibration cache. More...
 
virtual void writeCalibrationCache (const void *ptr, std::size_t length)=0
 Save a calibration cache. More...
 

Detailed Description

MinMax Calibrator.

This is the preferred calibrator for NLP tasks. It supports per activation tensor scaling.

Constructor & Destructor Documentation

◆ ~IInt8MinMaxCalibrator()

virtual nvinfer1::IInt8MinMaxCalibrator::~IInt8MinMaxCalibrator ( )
inlinevirtual

Member Function Documentation

◆ getAlgorithm()

CalibrationAlgoType nvinfer1::IInt8MinMaxCalibrator::getAlgorithm ( )
inlineoverridevirtual

Signal that this is the MinMax Calibrator.

Implements nvinfer1::IInt8Calibrator.

◆ getBatchSize()

virtual int32_t nvinfer1::IInt8Calibrator::getBatchSize ( ) const
pure virtualinherited

Get the batch size used for calibration batches.

Returns
The batch size.

Implemented in sample::anonymous_namespace{sampleEngines.cpp}::RndInt8Calibrator.

◆ getBatch()

virtual bool nvinfer1::IInt8Calibrator::getBatch ( void *  bindings[],
const char *  names[],
int32_t  nbBindings 
)
pure virtualinherited

Get a batch of input for calibration.

The batch size of the input must match the batch size returned by getBatchSize().

Parameters
bindingsAn array of pointers to device memory that must be updated to point to device memory containing each network input data.
namesThe names of the network input for each pointer in the binding array.
nbBindingsThe number of pointers in the bindings array.
Returns
False if there are no more batches for calibration.
See also
getBatchSize()

◆ readCalibrationCache()

virtual const void* nvinfer1::IInt8Calibrator::readCalibrationCache ( std::size_t &  length)
pure virtualinherited

Load a calibration cache.

Calibration is potentially expensive, so it can be useful to generate the calibration data once, then use it on subsequent builds of the network. The cache includes the regression cutoff and quantile values used to generate it, and will not be used if these do not batch the settings of the current calibrator. However, the network should also be recalibrated if its structure changes, or the input data set changes, and it is the responsibility of the application to ensure this.

Parameters
lengthThe length of the cached data, that should be set by the called function. If there is no data, this should be zero.
Returns
A pointer to the cache, or nullptr if there is no data.

◆ writeCalibrationCache()

virtual void nvinfer1::IInt8Calibrator::writeCalibrationCache ( const void *  ptr,
std::size_t  length 
)
pure virtualinherited

Save a calibration cache.

Parameters
ptrA pointer to the data to cache.
lengthThe length in bytes of the data to cache.
See also
readCalibrationCache()

The documentation for this class was generated from the following file: