Cuda¶
Dynamic Link Libs¶
hl_dso_loader.h¶
Functions
-
void
GetCublasDsoHandle
(void **dso_handle)¶ load the DSO of CUBLAS
- Parameters
**dso_handle
-dso handler
-
void
GetCudnnDsoHandle
(void **dso_handle)¶ load the DSO of CUDNN
- Parameters
**dso_handle
-dso handler
-
void
GetCudartDsoHandle
(void **dso_handle)¶ load the DSO of CUDA Run Time
- Parameters
**dso_handle
-dso handler
-
void
GetCurandDsoHandle
(void **dso_handle)¶ load the DSO of CURAND
- Parameters
**dso_handle
-dso handler
GPU Resources¶
hl_cuda.ph¶
Defines
-
HL_CUDA_PH_
¶
Typedefs
-
typedef struct _global_device_resources *
global_device_resources
¶
-
typedef struct _thread_device_resources *
thread_device_resources
¶
-
typedef struct _hl_device_prop *
hl_device_prop
¶
Functions
-
void
hl_create_thread_resources
(int device, thread_device_resources device_res)¶ thread device resource allocation.
create cuda stream and cuda event, allocate gpu memory and host page-lock memory for threads.
- Parameters
device
-device number.
device_res
-device properties.
-
void
hl_create_global_resources
(hl_device_prop device_prop)¶ global device resource allocation.
create cuda stream, initialize cublas, curand and cudnn.
- Parameters
device_prop
-device properties.
-
struct
_global_device_resources
¶ global device resources.
- Parameters
*stream
-device global stream.
handle
-devcie cublas handle.
gen
-device curand generator.
cudnn_handle
-cudnn handle.
*gen_mutex
-gen lock.
-
struct
_thread_device_resources
¶
-
struct
_hl_device_prop
¶
hl_cuda.h¶
Typedefs
-
typedef struct _hl_event_st *
hl_event_t
¶ HPPL event.
Functions
-
int
hl_get_cuda_lib_version
()¶ return cuda runtime api version.
-
void
hl_start
()¶ HPPL strat(Initialize all GPU).
-
void
hl_specify_devices_start
(int *device, int number)¶ HPPL start(Initialize the specific GPU).
- Parameters
device
-device id(0, 1......). if device is NULL, will start all GPU.
number
-number of devices.
-
bool
hl_device_can_access_peer
(int device, int peerDevice)¶ Queries if a device may directly access a peer device’s memory.
- Return
- Returns true if device is capable of directly accessing memory from peerDevice and false otherwise.
- Parameters
device
-Device from which allocations on peerDevice are to be directly accessed.
peerDevice
-Device on which the allocations to be directly accessed by device reside.
-
void
hl_device_enable_peer_access
(int peerDevice)¶ Enables direct access to memory allocations on a peer device.
- Parameters
peerDevice
-Peer device to enable direct access to from the current device
-
void
hl_init
(int device)¶ Init a work thread.
- Parameters
device
-device id.
-
void
hl_fini
()¶ Finish a work thread.
-
void
hl_set_sync_flag
(bool flag)¶ Set synchronous/asynchronous flag.
- Note
- This setting is only valid for the current worker thread.
- Parameters
flag
-true(default), set synchronous flag. false, set asynchronous flag.
-
bool
hl_get_sync_flag
()¶ Get synchronous/asynchronous flag.
- Return
- Synchronous call true. Asynchronous call false.
-
int
hl_get_device_count
()¶ Returns the number of compute-capable devices.
-
void
hl_set_device
(int device)¶ Set device to be used.
- Parameters
device
-device id.
-
int
hl_get_device
()¶ Returns which device is currently being used.
- Return
- device device id.
-
void *
hl_malloc_device
(size_t size)¶ Allocate device memory.
- Return
- dest_d pointer to device memory.
- Parameters
size
-size in bytes to copy.
-
void
hl_free_mem_device
(void *dest_d)¶ Free device memory.
- Parameters
dest_d
-pointer to device memory.
-
void *
hl_malloc_host
(size_t size)¶ Allocate host page-lock memory.
- Return
- dest_h pointer to host memory.
- Parameters
size
-size in bytes to copy.
-
void
hl_free_mem_host
(void *dest_h)¶ Free host page-lock memory.
- Parameters
dest_h
-pointer to host memory.
-
void
hl_memcpy
(void *dst, void *src, size_t size)¶ Copy data.
- Parameters
dst
-dst memory address(host or device).
src
-src memory address(host or device).
size
-size in bytes to copy.
-
void
hl_memset_device
(void *dest_d, int value, size_t size)¶ Set device memory to a value.
- Parameters
dest_d
-pointer to device memory.
value
-value to set for each byte of specified memory.
size
-size in bytes to set.
-
void
hl_memcpy_host2device
(void *dest_d, void *src_h, size_t size)¶ Copy host memory to device memory.
- Parameters
dest_d
-dst memory address.
src_h
-src memory address.
size
-size in bytes to copy.
-
void
hl_memcpy_device2host
(void *dest_h, void *src_d, size_t size)¶ Copy device memory to host memory.
- Parameters
dest_h
-dst memory address.
src_d
-src memory address.
size
-size in bytes to copy.
-
void
hl_memcpy_device2device
(void *dest_d, void *src_d, size_t size)¶ Copy device memory to device memory.
- Parameters
dest_d
-dst memory address.
src_d
-src memory address.
size
-size in bytes to copy.
-
void
hl_rand
(real *dest_d, size_t num)¶ Generate uniformly distributed floats (0, 1.0].
- Parameters
dest_d
-pointer to device memory to store results.
num
-number of floats to generate.
-
void
hl_srand
(unsigned int seed)¶ Set the seed value of the random number generator.
- Parameters
seed
-seed value.
-
void
hl_memcpy_async
(void *dst, void *src, size_t size, hl_stream_t stream)¶ Copy data.
- Parameters
dst
-dst memory address(host or device).
src
-src memory address(host or device).
size
-size in bytes to copy.
stream
-stream id.
-
void
hl_stream_synchronize
(hl_stream_t stream)¶ Waits for stream tasks to complete.
- Parameters
stream
-stream id.
-
void
hl_create_event
(hl_event_t *event)¶ Creates an event object.
- Parameters
event
-New event.
-
void
hl_destroy_event
(hl_event_t event)¶ Destroys an event object.
- Parameters
event
-Event to destroy.
-
float
hl_event_elapsed_time
(hl_event_t start, hl_event_t end)¶ Computes the elapsed time between events.
- Return
- time Time between start and end in ms.
- Parameters
start
-Starting event.
end
-Ending event.
-
void
hl_stream_record_event
(hl_stream_t stream, hl_event_t event)¶ Records an event.
- Parameters
stream
-Stream in which to insert event.
event
-Event waiting to be recorded as completed.
-
void
hl_stream_wait_event
(hl_stream_t stream, hl_event_t event)¶ Make a compute stream wait on an event.
- Parameters
stream
-Stream in which to insert event.
event
-Event to wait on.
-
void
hl_event_synchronize
(hl_event_t event)¶ Wait for an event to complete.
- Parameters
event
-event to wait for.
-
void
hl_set_device_flags_block
()¶ Sets block flags to be used for device executions.
- Note
- This interface needs to be called before hl_start.
-
const char *
hl_get_device_error_string
()¶ Returns the last error string from a cuda runtime call.
-
const char *
hl_get_device_error_string
(size_t err)¶ Returns the last error string from a cuda runtime call.
- See
- hl_get_device_last_error()
- Parameters
err
-error number.
-
int
hl_get_device_last_error
()¶ Returns the last error number.
- Return
- error number.
- See
- hl_get_device_error_string()
-
bool
hl_cuda_event_is_ready
(hl_event_t event)¶ check cuda event is ready
- Return
- true cuda event is ready. false cuda event is not ready.
- Parameters
event
-cuda event to query.
-
void
hl_device_synchronize
()¶ hppl device synchronization.
CUDA Wrapper¶
hl_cuda_cublas.h¶
Functions
-
void
hl_matrix_transpose
(real *A_d, real *C_d, int dimM, int dimN, int lda, int ldc)¶ Matrix transpose: C_d = T(A_d)
- Parameters
A_d
-input matrix (M x N).
C_d
-output matrix (N x M).
dimM
-matrix height.
dimN
-matrix width.
lda
-the first dimension of A_d.
ldc
-the first dimension of C_d.
-
void
hl_matrix_transpose
(real *A_d, real *C_d, int dimM, int dimN)¶
-
void
hl_matrix_mul
(real *A_d, hl_trans_op_t transa, real *B_d, hl_trans_op_t transb, real *C_d, int dimM, int dimN, int dimK, real alpha, real beta, int lda, int ldb, int ldc)¶ C_d = alpha*(op(A_d) * op(B_d)) + beta*C_d.
- Parameters
A_d
-input.
transa
-operation op(A) that is non-or transpose.
B_d
-input.
transb
-operation op(B) that is non-or transpose.
C_d
-output.
dimM
-matrix height of op(A) & C
dimN
-matrix width of op(B) & C
dimK
-width of op(A) & height of op(B)
alpha
-scalar used for multiplication.
beta
-scalar used for multiplication.
lda
-the first dimension of A_d.
ldb
-the first dimension of B_d.
ldc
-the first dimension of C_d.
-
void
hl_matrix_mul
(real *A_d, hl_trans_op_t transa, real *B_d, hl_trans_op_t transb, real *C_d, int dimM, int dimN, int dimK, real alpha, real beta)¶ C_d = alpha*(op(A_d) * op(B_d)) + beta*C_d.
- Parameters
A_d
-input.
transa
-operation op(A) that is non-or transpose.
B_d
-input.
transb
-operation op(B) that is non-or transpose.
C_d
-output.
dimM
-matrix height of op(A) & C
dimN
-matrix width of op(B) & C
dimK
-width of op(A) & height of op(B)
alpha
-scalar used for multiplication.
beta
-scalar used for multiplication.
-
void
hl_matrix_mul_vector
(real *A_d, hl_trans_op_t trans, real *B_d, real *C_d, int dimM, int dimN, real alpha, real beta, int lda, int incb, int incc)¶ This function performs the matrix-vector multiplication. C_d = alpha*op(A_d)*B_d + beta*C_d.
- Parameters
A_d
-matrix.
trans
-operation op(A) that is non-or transpose.
B_d
-vector with dimN(dimM) elements if trans==HPPL_OP_N(HPPL_OP_T).
C_d
-vector with dimM(dimN) elements if trans==HPPL_OP_N(HPPL_OP_T).
dimM
-number of rows of matrix A_d.
dimN
-number of columns of matrix A_d.
alpha
-scalar used for multiplication.
beta
-scalar used for multiplication.
lda
-the first dimension of A_d.
incb
-increase B_d size for compaction.
incc
-increase C_d size for compaction.
-
void
hl_matrix_mul_vector
(real *A_d, hl_trans_op_t trans, real *B_d, real *C_d, int dimM, int dimN, real alpha, real beta)¶ This function performs the matrix-vector multiplication. C_d = alpha*op(A_d)*B_d + beta*C_d.
- Parameters
A_d
-matrix.
trans
-operation op(A) that is non-or transpose.
B_d
-vector with dimN(dimM) elements if trans==HPPL_OP_N(HPPL_OP_T).
C_d
-vector with dimM(dimN) elements if trans==HPPL_OP_N(HPPL_OP_T).
dimM
-number of rows of matrix A_d.
dimN
-number of columns of matrix A_d.
alpha
-scalar used for multiplication.
beta
-scalar used for multiplication.
hl_cuda_cudnn.h¶
Typedefs
-
typedef struct _hl_tensor_descriptor *
hl_tensor_descriptor
¶ hppl image descriptor.
-
typedef struct _hl_pooling_descriptor *
hl_pooling_descriptor
¶ hppl pooling descriptor.
-
typedef struct _hl_filter_descriptor *
hl_filter_descriptor
¶ hppl filter descriptor.
-
typedef struct _hl_convolution_descriptor *
hl_convolution_descriptor
¶ hppl filter descriptor.
Enums
Functions
-
int
hl_get_cudnn_lib_version
()¶ return cudnn lib version
-
void
hl_create_tensor_descriptor
(hl_tensor_descriptor *image_desc)¶ create image descriptor.
- Parameters
image_desc
-image descriptor.
-
void
hl_tensor_reshape
(hl_tensor_descriptor image_desc, int batch_size, int feature_maps, int height, int width)¶ reshape image descriptor.
- Parameters
image_desc
-image descriptor.
batch_size
-input batch size.
feature_maps
-image feature maps.
height
-image height.
width
-image width.
-
void
hl_tensor_reshape
(hl_tensor_descriptor image_desc, int batch_size, int feature_maps, int height, int width, int nStride, int cStride, int hStride, int wStride)¶ reshape image descriptor.
- Parameters
image_desc
-image descriptor.
batch_size
-input batch size.
feature_maps
-image feature maps.
height
-image height.
width
-image width.
nStride
-stride between two consecutive images.
cStride
-stride between two consecutive feature maps.
hStride
-stride between two consecutive rows.
wStride
-stride between two consecutive columns.
-
void
hl_destroy_tensor_descriptor
(hl_tensor_descriptor image_desc)¶ destroy image descriptor.
- Parameters
image_desc
-hppl image descriptor.
-
void
hl_create_pooling_descriptor
(hl_pooling_descriptor *pooling_desc, hl_pooling_mode_t mode, int height, int width, int height_padding, int width_padding, int stride_height, int stride_width)¶ create pooling descriptor.
- Parameters
pooling_desc
-pooling descriptor.
mode
-pooling mode.
height
-height of the pooling window.
width
-width of the pooling window.
height_padding
-padding height.
width_padding
-padding width.
stride_height
-pooling vertical stride.
stride_width
-pooling horizontal stride.
-
void
hl_destroy_pooling_descriptor
(hl_pooling_descriptor pooling_desc)¶ destroy pooling descriptor.
- Parameters
pooling_desc
-hppl pooling descriptor.
-
void
hl_pooling_forward
(hl_tensor_descriptor input, real *input_image, hl_tensor_descriptor output, real *output_image, hl_pooling_descriptor pooling)¶ pooling forward(calculate output image).
- Parameters
input
-input image descriptor.
input_image
-input image data.
output
-output image descriptor.
output_image
-output image data.
pooling
-pooling descriptor.
-
void
hl_pooling_backward
(hl_tensor_descriptor input, real *input_image, real *input_image_grad, hl_tensor_descriptor output, real *output_image, real *output_image_grad, hl_pooling_descriptor pooling)¶ pooling backward(calculate input image gradient).
- Parameters
input
-input image descriptor.
input_image
-input image data.
input_image_grad
-input image gradient data.
output
-output image descriptor.
output_image
-output image data.
output_image_grad
-output image gradient data.
pooling
-pooling descriptor.
-
void
hl_create_filter_descriptor
(hl_filter_descriptor *filter, int input_feature_maps, int output_feature_maps, int height, int width)¶ create filter descriptor.
- Parameters
filter
-filter descriptor.
input_feature_maps
-input image feature maps.
output_feature_maps
-output image feature maps.
height
-filter height.
width
-filter width.
-
void
hl_conv_workspace
(hl_tensor_descriptor input, hl_tensor_descriptor output, hl_filter_descriptor filter, hl_convolution_descriptor conv, int *convFwdAlgo, size_t *fwdLimitBytes, int *convBwdDataAlgo, size_t *bwdDataLimitBytes, int *convBwdFilterAlgo, size_t *bwdFilterLimitBytes)¶ convolution workspace configuration
- Parameters
input
-image descriptor
output
-image descriptor
filter
-filter descriptor
conv
-convolution descriptor
convFwdAlgo
-forward algorithm
fwdLimitBytes
-forward workspace size
convBwdDataAlgo
-backward data algorithm
bwdDataLimitBytes
-backward data workspace size
convBwdFilterAlgo
-backward filter algorithm
bwdFilterLimitBytes
-backward filter workspace size
-
void
hl_destroy_filter_descriptor
(hl_filter_descriptor filter)¶ destroy filter descriptor.
- Parameters
filter
-hppl filter descriptor.
-
void
hl_create_convolution_descriptor
(hl_convolution_descriptor *conv, hl_tensor_descriptor image, hl_filter_descriptor filter, int padding_height, int padding_width, int stride_height, int stride_width)¶ create convolution descriptor.
- Parameters
conv
-conv descriptor.
image
-input image descriptor.
filter
-filter descriptor.
padding_height
-padding height.
padding_width
-padding width.
stride_height
-stride height.
stride_width
-stride width.
-
void
hl_reset_convolution_descriptor
(hl_convolution_descriptor conv, hl_tensor_descriptor image, hl_filter_descriptor filter, int padding_height, int padding_width, int stride_height, int stride_width)¶ reset convolution descriptor.
- Parameters
conv
-conv descriptor.
image
-input image descriptor.
filter
-filter descriptor.
padding_height
-padding height.
padding_width
-padding width.
stride_height
-stride height.
stride_width
-stride width.
-
void
hl_destroy_convolution_descriptor
(hl_convolution_descriptor conv)¶ destroy convolution descriptor.
- Parameters
conv
-hppl convolution descriptor.
-
void
hl_convolution_forward
(hl_tensor_descriptor input, real *input_data, hl_tensor_descriptor output, real *output_data, hl_filter_descriptor filter, real *filter_data, hl_convolution_descriptor conv, void *gpuWorkSpace, size_t sizeInBytes, int convFwdAlgo)¶ convolution forward(calculate output image).
- Parameters
input
-input image descriptor.
input_data
-input image data.
output
-output image descriptor.
output_data
-output image data.
filter
-filter descriptor.
filter_data
-filter data.
conv
-convolution descriptor.
gpuWorkSpace
-limited gpu workspace.
sizeInBytes
-gpu workspace size (bytes).
convFwdAlgo
-forward algorithm.
-
void
hl_convolution_forward_add_bias
(hl_tensor_descriptor bias, real *bias_data, hl_tensor_descriptor output, real *output_data)¶ convolution forward add bias(calculate output add bias).
- Parameters
bias
-bias descriptor.
bias_data
-bias data.
output
-output image descriptor.
output_data
-output image data.
-
void
hl_convolution_backward_filter
(hl_tensor_descriptor input, real *input_data, hl_tensor_descriptor output, real *output_grad_data, hl_filter_descriptor filter, real *filter_grad_data, hl_convolution_descriptor conv, void *gpuWorkSpace, size_t sizeInBytes, int convBwdFilterAlgo)¶ convolution backward filter(calculate filter grad data).
- Parameters
input
-input image descriptor.
input_data
-input image data.
output
-output image descriptor.
output_grad_data
-output image grad data.
filter
-filter descriptor.
filter_grad_data
-filter grad data.
conv
-convolution descriptor.
gpuWorkSpace
-limited gpu workspace.
sizeInBytes
-gpu workspace size (bytes).
convBwdFilterAlgo
-backward filter algorithm.
-
void
hl_convolution_backward_data
(hl_tensor_descriptor input, real *input_data_grad, hl_tensor_descriptor output, real *output_grad_data, hl_filter_descriptor filter, real *filter_data, hl_convolution_descriptor conv, void *gpuWorkSpace, size_t sizeInBytes, int convBwdDataAlgo)¶ convolution backward data(calculate input image grad data).
- Parameters
input
-input image descriptor.
input_data_grad
-input image grad data.
output
-output image descriptor.
output_grad_data
-output image grad data.
filter
-filter descriptor.
filter_data
-filter data.
conv
-convolution descriptor.
gpuWorkSpace
-limited gpu workspace.
sizeInBytes
-gpu workspace size (bytes).
convBwdDataAlgo
-backward data algorithm.
-
void
hl_convolution_backward_bias
(hl_tensor_descriptor bias, real *bias_grad_data, hl_tensor_descriptor output, real *output_grad_data)¶ convolution backward bias(calculate bias grad data).
- Parameters
bias
-bias descriptor.
bias_grad_data
-bias grad data.
output
-output image descriptor.
output_grad_data
-output image grad data.
-
void
hl_softmax_forward
(real *input, real *output, int height, int width)¶ softmax forward.
- Parameters
input
-input value.
output
-output value.
height
-matrix height.
width
-matrix width.
-
void
hl_softmax_backward
(real *output_value, real *output_grad, int height, int width)¶ softmax backward.
- Parameters
output_value
-output value data.
output_grad
-output grad data.
height
-matrix height.
width
-matrix width.
-
void
hl_batch_norm_forward_training
(hl_tensor_descriptor inputDesc, real *input, hl_tensor_descriptor outputDesc, real *output, hl_tensor_descriptor bnParamDesc, real *scale, real *bias, double factor, real *runningMean, real *runningInvVar, double epsilon, real *savedMean, real *savedVar)¶ cudnn batch norm forward.
- Parameters
inputDesc
-input tensor descriptor desc.
input
-input data.
outputDesc
-output tensor descriptor desc.
output
-output data.
bnParamDesc
-tensor descriptor desc. bnScale, bnBias, running mean/var, save_mean/var.
scale
-batch normalization scale parameter (in original paper scale is referred to as gamma).
bias
-batch normalization bias parameter (in original paper scale is referred to as beta).
factor
-Factor used in the moving average computation. runningMean = newMean * factor
- runningMean * (1 - factor)
runningMean
-running mean.
runningInvVar
-running variance.
epsilon
-Epsilon value used in the batch normalization formula.
savedMean
-optional cache to save intermediate results.
savedVar
-optional cache to save intermediate results.
-
void
hl_batch_norm_forward_inference
(hl_tensor_descriptor inputDesc, real *input, hl_tensor_descriptor outputDesc, real *output, hl_tensor_descriptor bnParamDesc, real *scale, real *bias, real *estimatedMean, real *estimatedVar, double epsilon)¶ cudnn batch norm forward.
- Parameters
inputDesc
-input tensor descriptor desc.
input
-input data.
outputDesc
-output tensor descriptor desc.
output
-output data.
bnParamDesc
-tensor descriptor desc. bnScale, bnBias, running mean/var, save_mean/var.
scale
-batch normalization scale parameter (in original paper scale is referred to as gamma).
bias
-batch normalization bias parameter (in original paper scale is referred to as beta).
estimatedMean
-estimatedVar
-It is suggested that resultRunningMean, resultRunningVariance from the cudnnBatchNormalizationForwardTraining call accumulated during the training phase are passed as inputs here.
epsilon
-Epsilon value used in the batch normalization formula.
-
void
hl_batch_norm_backward
(hl_tensor_descriptor inputDesc, real *input, hl_tensor_descriptor outGradDesc, real *outGrad, hl_tensor_descriptor inGradDesc, real *inGrad, hl_tensor_descriptor dBnParamDesc, real *scale, real *scaleGrad, real *biasGrad, double epsilon, real *savedMean, real *savedInvVar)¶ cudnn batch norm forward.
- Parameters
inputDesc
-input tensor descriptor desc.
input
-input data.
outGradDesc
-output tensor descriptor desc.
outGrad
-output data.
inGradDesc
-input tensor descriptor desc.
inGrad
-input data.
dBnParamDesc
-tensor descriptor desc. bnScale, bnBias, running mean/var, save_mean/var.
scale
-batch normalization scale parameter (in original paper scale is referred to as gamma).
scaleGrad
-batch normalization scale parameter (in original paper scale is referred to as gamma) gradient.
biasGrad
-batch normalization bias parameter (in original paper scale is referred to as beta) gradient.
epsilon
-Epsilon value used in the batch normalization formula.
savedMean
-optional cache to save intermediate results.
savedInvVar
-optional cache to save intermediate results.
hl_cuda_cudnn.h¶
Defines
-
HL_CUDA_CUDNN_PH_
¶
-
GET_TENSOR_DESCRIPTOR
(image)¶
-
GET_FILTER_DESCRIPTOR
(filter)¶
-
GET_CONVOLUTION_DESCRIPTOR
(conv)¶
Typedefs
-
typedef struct _cudnn_tensor_descriptor *
cudnn_tensor_descriptor
¶
-
typedef struct _cudnn_pooling_descriptor *
cudnn_pooling_descriptor
¶
-
typedef struct _cudnn_filter_descriptor *
cudnn_filter_descriptor
¶
-
typedef struct _cudnn_convolution_descriptor *
cudnn_convolution_descriptor
¶
-
struct
_cudnn_tensor_descriptor
¶
-
struct
_cudnn_pooling_descriptor
¶
-
struct
_cudnn_filter_descriptor
¶
-
struct
_cudnn_convolution_descriptor
¶