Пример #1
0
def sliced_rows_batch_scaled_add(stream, embd_rows_indxs, nrows, ncols, alpha,
                                 dense_matrices, embd_nrows, embd_ncols,
                                 embd_matrix):
    status = gpu_matrix_kernels._slicedRowsBatchScaledAdd(
        stream, embd_rows_indxs, nrows, ncols, alpha, dense_matrices,
        embd_nrows, embd_ncols, embd_matrix)
    cudart.check_cuda_status(status)
Пример #2
0
def add_scaled_columns_slice(stream, nrows, ncols, alpha, dense_matrix,
                             embedding_column_indxs, embedding_matrix):
    status = gpu_matrix_kernels._addScaledColumnsSlice(stream, nrows, ncols,
                                                       alpha, dense_matrix,
                                                       embedding_column_indxs,
                                                       embedding_matrix)
    cudart.check_cuda_status(status)
Пример #3
0
def batch_horizontal_split(stream, n, nrows, x_ncols, y_ncols, matrices,
                           x_matrices, y_matrices):
    status = gpu_matrix_kernels._batchHorizontalSplit(stream, n, nrows,
                                                      x_ncols, y_ncols,
                                                      matrices, x_matrices,
                                                      y_matrices)
    cudart.check_cuda_status(status)
Пример #4
0
def softmax_ce_derivative(stream, batchSize, num_classes, probs,
                          target_classes, derivatives):
    status = gpu_matrix_kernels._softmaxCeDerivative(stream, batchSize,
                                                     num_classes, probs,
                                                     target_classes,
                                                     derivatives)
    cudart.check_cuda_status(status)
Пример #5
0
def slice_rows_int(stream, embedding_matrix_nrows, embedding_row_indxs,
                   embedding_matrix, nrows, ncols, dense_matrix):
    status = gpu_matrix_kernels._sliceRowsInt(stream, embedding_matrix_nrows,
                                              embedding_row_indxs,
                                              embedding_matrix, nrows, ncols,
                                              dense_matrix)
    cudart.check_cuda_status(status)
Пример #6
0
def sigmoid_der(stream, nelems, data, sigmoid_data, derivative):
    status = nonlinearities._sigmoidDer(stream, nelems, data, sigmoid_data,
                                        derivative)
    cudart.check_cuda_status(status)
Пример #7
0
def tanh_sigm_der(stream, axis, nrows, ncols, data, tanh_sigm_data, derivatve):
    status = nonlinearities._tanhSigmDer(stream, axis, nrows, ncols, data,
                                         tanh_sigm_data, derivatve)
    cudart.check_cuda_status(status)
Пример #8
0
def add_scaled_columns_slice(stream, nrows, ncols, alpha, dense_matrix, embedding_column_indxs, embedding_matrix):
    status = gpu_matrix_kernels._addScaledColumnsSlice(stream, nrows, ncols, alpha, dense_matrix, embedding_column_indxs, embedding_matrix)
    cudart.check_cuda_status(status)
Пример #9
0
def assign_sum(stream, nelems, matrices, n, s):
    status = gpu_matrix_kernels._assign_sum(stream, nelems, matrices, n, s)
    cudart.check_cuda_status(status)
Пример #10
0
def mask_column_numbers_row_wise(stream, nrows, ncols, numbers, out):
    status = gpu_matrix_kernels._maskColumnNumbersRowWise(stream, nrows, ncols, numbers, out)
    cudart.check_cuda_status(status)
Пример #11
0
def assign_masked_addition_column_broadcasted(stream, nrows, ncols, mask, a, b, out):
    status = gpu_matrix_kernels._assignMaskedAdditionColumnBroadcasted(stream, nrows, ncols, mask, a, b, out)
    cudart.check_cuda_status(status)
Пример #12
0
def assign_masked_addition(stream, nelems, mask, a, b, out):
    status = gpu_matrix_kernels._assignMaskedAddition(stream, nelems, mask, a, b, out)
    cudart.check_cuda_status(status)
Пример #13
0
def add_mask_zeros(stream, nelems, a, b, out):
    status = gpu_matrix_kernels._addMaskZeros(stream, nelems, a, b, out)
    cudart.check_cuda_status(status)
Пример #14
0
def dropout(stream, nelems, dropout_prob, data, uniform_data, out):
    status = gpu_matrix_kernels._dropout(stream, nelems, dropout_prob, data, uniform_data, out)
    cudart.check_cuda_status(status)
Пример #15
0
def add_scaled_subtraction(stream, nelems, alpha, a, b, out):
    status = gpu_matrix_kernels._addScaledSubtraction(stream, nelems, alpha, a, b, out)
    cudart.check_cuda_status(status)
Пример #16
0
def add_softmax_ce_derivative(stream, batchSize, num_classes, probs, target_classes, derivatives):
    status = gpu_matrix_kernels._addSoftmaxCeDerivative(stream, batchSize, num_classes, probs, target_classes, derivatives)
    cudart.check_cuda_status(status)
Пример #17
0
def add_mask_zeros(stream, nelems, a, b, out):
    status = gpu_matrix_kernels._addMaskZeros(stream, nelems, a, b, out)
    cudart.check_cuda_status(status)
Пример #18
0
def assign_masked_addition_column_broadcasted(stream, nrows, ncols, mask, a, b, out):
    status = gpu_matrix_kernels._assignMaskedAdditionColumnBroadcasted(stream, nrows, ncols, mask, a, b, out)
    cudart.check_cuda_status(status)
Пример #19
0
def add_hprod_one_minus_mask_column_broadcasted(stream, nrows, ncols, mask, a, out):
    status = gpu_matrix_kernels._addHprodOneMinusMaskColumnBroadcasted(stream, nrows, ncols, mask, a, out)
    cudart.check_cuda_status(status)
Пример #20
0
def add_hprod_one_minus_mask(stream, nelems, mask, a, out):
    status = gpu_matrix_kernels._addHprodOneMinusMask(stream, nelems, mask, a, out)
    cudart.check_cuda_status(status)
Пример #21
0
def repeat_along_col(stream, repeats, nrows, ncols, a, out):
    status = gpu_matrix_kernels._repeatAlongCol(stream, repeats, nrows, ncols, a, out)
    cudart.check_cuda_status(status)
Пример #22
0
def add_hprod_one_minus_mask_column_broadcasted(stream, nrows, ncols, mask, a, out):
    status = gpu_matrix_kernels._addHprodOneMinusMaskColumnBroadcasted(stream, nrows, ncols, mask, a, out)
    cudart.check_cuda_status(status)
Пример #23
0
def clip(stream, nelems, min_value, max_value, data, out):
    status = gpu_matrix_kernels._clip(stream, nelems, min_value, max_value, data, out)
    cudart.check_cuda_status(status)
Пример #24
0
def matrix_vector_column_hprod(stream, nrows, ncols, matrix, vector, out):
    status = gpu_matrix_kernels._matrixVectorColumnHprod(stream, nrows, ncols, matrix, vector, out)
    cudart.check_cuda_status(status)
Пример #25
0
 def callback(stream, status, user_data):
     cudart.check_cuda_status(status)
     args, kwargs = ct.cast(user_data, ct_py_object_p).contents.value
     function(*args, **kwargs)
     GpuContext._user_data[ct.cast(stream, ct.c_void_p).value].popleft()
Пример #26
0
def mask_column_numbers_row_wise(stream, nrows, ncols, numbers, out):
    status = gpu_matrix_kernels._maskColumnNumbersRowWise(stream, nrows, ncols, numbers, out)
    cudart.check_cuda_status(status)
Пример #27
0
def relu_der(stream, nelems, data, relu_data, derivative):
    status = nonlinearities._reluDer(stream, nelems, data, relu_data,
                                     derivative)
    cudart.check_cuda_status(status)
Пример #28
0
def assign_sum(stream, nelems, matrices, n, s):
    status = gpu_matrix_kernels._assign_sum(stream, nelems, matrices, n, s)
    cudart.check_cuda_status(status)
Пример #29
0
def tanh_der(stream, nelems, data, tanh_data, derivative):
    status = nonlinearities._tanhDer(stream, nelems, data, tanh_data,
                                     derivative)
    cudart.check_cuda_status(status)
Пример #30
0
def repeat_along_col(stream, repeats, nrows, ncols, a, out):
    status = gpu_matrix_kernels._repeatAlongCol(stream, repeats, nrows, ncols, a, out)
    cudart.check_cuda_status(status)
Пример #31
0
def add_scaled_subtraction(stream, nelems, alpha, a, b, out):
    status = gpu_matrix_kernels._addScaledSubtraction(stream, nelems, alpha, a, b, out)
    cudart.check_cuda_status(status)
Пример #32
0
def add_repeat_along_col_derivative(stream, repeats, a, nrows, ncols, derivative):
    status = gpu_matrix_kernels._addRepeatAlongColDerivative(stream, repeats, a, nrows, ncols, derivative)
    cudart.check_cuda_status(status)
Пример #33
0
def dropout(stream, nelems, dropout_prob, data, uniform_data, out):
    status = gpu_matrix_kernels._dropout(stream, nelems, dropout_prob, data, uniform_data, out)
    cudart.check_cuda_status(status)
Пример #34
0
def add_scaled_div_sqrt(stream, nelems, alpha, a, b, epsilon, c):
    status = gpu_matrix_kernels._addScaledDivSqrt(stream, nelems, alpha, a, b, epsilon, c)
    cudart.check_cuda_status(status)
Пример #35
0
def assign_masked_addition(stream, nelems, mask, a, b, out):
    status = gpu_matrix_kernels._assignMaskedAddition(stream, nelems, mask, a, b, out)
    cudart.check_cuda_status(status)
Пример #36
0
def clip(stream, nelems, min_value, max_value, data, out):
    status = gpu_matrix_kernels._clip(stream, nelems, min_value, max_value, data, out)
    cudart.check_cuda_status(status)
Пример #37
0
def add_hprod_one_minus_mask(stream, nelems, mask, a, out):
    status = gpu_matrix_kernels._addHprodOneMinusMask(stream, nelems, mask, a, out)
    cudart.check_cuda_status(status)
Пример #38
0
def transpose_int(stream, nrows, ncols, in_, out):
    status = gpu_matrix_kernels._transposeInt(stream, nrows, ncols, in_, out)
    cudart.check_cuda_status(status)
Пример #39
0
def matrix_vector_column_hprod(stream, nrows, ncols, matrix, vector, out):
    status = gpu_matrix_kernels._matrixVectorColumnHprod(stream, nrows, ncols, matrix, vector, out)
    cudart.check_cuda_status(status)
Пример #40
0
def add_hadamard_product_2(stream, nelems, a, b, alpha, c):
    status = gpu_matrix_kernels._addHadamardProduct2(stream, nelems, a, b, alpha, c)
    cudart.check_cuda_status(status)
Пример #41
0
def batch_horizontal_split(stream, n, nrows, x_ncols, y_ncols, matrices, x_matrices, y_matrices):
    status = gpu_matrix_kernels._batchHorizontalSplit(stream, n, nrows, x_ncols, y_ncols, matrices, x_matrices, y_matrices)
    cudart.check_cuda_status(status)
Пример #42
0
def assign_sequential_sum_pooling(stream, nrows, ncols, matrices, n, out):
    status = gpu_matrix_kernels._assignSequentialSumPooling(stream, nrows, ncols, matrices, n, out)
    cudart.check_cuda_status(status)
Пример #43
0
def add_repeat_along_col_derivative(stream, repeats, a, nrows, ncols, derivative):
    status = gpu_matrix_kernels._addRepeatAlongColDerivative(stream, repeats, a, nrows, ncols, derivative)
    cudart.check_cuda_status(status)
Пример #44
0
def assign_sequential_weighted_sum(stream, nrows, ncols, matrices, weights, n, out):
    status = gpu_matrix_kernels._assignSequentialWeightedSum(stream, nrows, ncols, matrices, weights, n, out)
    cudart.check_cuda_status(status)
Пример #45
0
def add_scaled_div_sqrt(stream, nelems, alpha, a, b, epsilon, c):
    status = gpu_matrix_kernels._addScaledDivSqrt(stream, nelems, alpha, a, b, epsilon, c)
    cudart.check_cuda_status(status)
Пример #46
0
def sequentially_tile(stream, nelems, a, matrices, n):
    status = gpu_matrix_kernels._sequentiallyTile(stream, nelems, a, matrices, n)
    cudart.check_cuda_status(status)
Пример #47
0
def transpose_int(stream, nrows, ncols, in_, out):
    status = gpu_matrix_kernels._transposeInt(stream, nrows, ncols, in_, out)
    cudart.check_cuda_status(status)
Пример #48
0
def assign_dL_dpre_a(stream, nrows, ncols, matrices, derivative, weights, n, out):
    status = gpu_matrix_kernels._assignDLDprea(stream, nrows, ncols, matrices, derivative, weights, n, out)
    cudart.check_cuda_status(status)
Пример #49
0
 def callback(stream, status, user_data):
     cudart.check_cuda_status(status)
     args, kwargs = ct.cast(user_data, ct_py_object_p).contents.value
     function(*args, **kwargs)
     GpuContext._user_data[ct.cast(stream, ct.c_void_p).value].popleft()
Пример #50
0
def add_attention_derivative(stream, nrows, ncols, matrices, derivative, n, out):
    status = gpu_matrix_kernels._addAttentionDerivative(stream, nrows, ncols, matrices, derivative, n, out)
    cudart.check_cuda_status(status)
Пример #51
0
def tanh_sigm(stream, axis, nrows, ncols, data, tanh_sigm_data):
    status = nonlinearities._tanhSigm(stream, axis, nrows, ncols, data,
                                      tanh_sigm_data)
    cudart.check_cuda_status(status)
Пример #52
0
def add_attention_tile(stream, nrows, ncols, derivative, a, dL_dpre_a, u, n, matrices_derivs):
    status = gpu_matrix_kernels._addAttentionTile(stream, nrows, ncols, derivative, a, dL_dpre_a, u, n, matrices_derivs)
    cudart.check_cuda_status(status)
Пример #53
0
def relu(stream, nelems, data, relu_data):
    status = nonlinearities._relu(stream, nelems, data, relu_data)
    cudart.check_cuda_status(status)
Пример #54
0
def slice_rows_batch(stream, embd_rows_indxs, nrows, ncols, embd_matrix, embd_nrows, embd_ncols, dense_matrices):
    status = gpu_matrix_kernels._sliceRowsBatch(stream, embd_rows_indxs, nrows, ncols, embd_matrix, embd_nrows, embd_ncols, dense_matrices)
    cudart.check_cuda_status(status)
Пример #55
0
def sigmoid(stream, nelems, data, sigmoid_data):
    status = nonlinearities._sigmoid(stream, nelems, data, sigmoid_data)
    cudart.check_cuda_status(status)
Пример #56
0
def sliced_rows_batch_scaled_add(stream, embd_rows_indxs, nrows, ncols, alpha, dense_matrices, embd_nrows, embd_ncols, embd_matrix):
    status = gpu_matrix_kernels._slicedRowsBatchScaledAdd(stream, embd_rows_indxs, nrows, ncols, alpha, dense_matrices, embd_nrows, embd_ncols, embd_matrix)
    cudart.check_cuda_status(status)
Пример #57
0
def tanh(stream, nelems, data, tanh_data):
    status = nonlinearities._tanh(stream, nelems, data, tanh_data)
    cudart.check_cuda_status(status)
Пример #58
0
def assign_scaled_addition(stream, nelems, alpha, a, b, out):
    status = gpu_matrix_kernels._assignScaledAddition(stream, nelems, alpha, a, b, out)
    cudart.check_cuda_status(status)
Пример #59
0
def test_dependencies(cuda_stream, node_id, blocking_nodes, blocking_nodes_num, execution_checklist, test_results):
    status = test_events._testDependencies(cuda_stream, node_id, blocking_nodes, blocking_nodes_num, execution_checklist, test_results)
    cudart.check_cuda_status(status)
Пример #60
0
def masked_fill(stream, nelems, value, mask_data, true_value, out_data):
    status = gpu_matrix_kernels.\
        _maskedFill(stream, nelems, value, mask_data, true_value, out_data)
    cudart.check_cuda_status(status)