TensorRT  7.2.1.6
NVIDIA TensorRT
Looking for a C++ dev who knows TensorRT?
I'm looking for work. Hire me!
bert::Fused_multihead_attention_params Struct Reference
Collaboration diagram for bert::Fused_multihead_attention_params:

Public Attributes

void * qkv_ptr
 
void * packed_mask_ptr
 
void * o_ptr
 
int64_t qkv_stride_in_bytes
 
int64_t packed_mask_stride_in_bytes
 
int64_t o_stride_in_bytes
 
int b
 
int h
 
int s
 
int d
 
uint32_t scale_bmm1
 
uint32_t scale_softmax
 
uint32_t scale_bmm2
 
bool enable_i2f_trick
 

Member Data Documentation

◆ qkv_ptr

void* bert::Fused_multihead_attention_params::qkv_ptr

◆ packed_mask_ptr

void* bert::Fused_multihead_attention_params::packed_mask_ptr

◆ o_ptr

void* bert::Fused_multihead_attention_params::o_ptr

◆ qkv_stride_in_bytes

int64_t bert::Fused_multihead_attention_params::qkv_stride_in_bytes

◆ packed_mask_stride_in_bytes

int64_t bert::Fused_multihead_attention_params::packed_mask_stride_in_bytes

◆ o_stride_in_bytes

int64_t bert::Fused_multihead_attention_params::o_stride_in_bytes

◆ b

int bert::Fused_multihead_attention_params::b

◆ h

int bert::Fused_multihead_attention_params::h

◆ s

int bert::Fused_multihead_attention_params::s

◆ d

int bert::Fused_multihead_attention_params::d

◆ scale_bmm1

uint32_t bert::Fused_multihead_attention_params::scale_bmm1

◆ scale_softmax

uint32_t bert::Fused_multihead_attention_params::scale_softmax

◆ scale_bmm2

uint32_t bert::Fused_multihead_attention_params::scale_bmm2

◆ enable_i2f_trick

bool bert::Fused_multihead_attention_params::enable_i2f_trick

The documentation for this struct was generated from the following file: