diff --git a/paddle/operators/math/math_function.cu b/paddle/operators/math/math_function.cu index 3e2aeea1da490d524db7b04cce947c57af97d8ef..b7d2c48a5fb12b692097299ed3822d670714ec65 100644 --- a/paddle/operators/math/math_function.cu +++ b/paddle/operators/math/math_function.cu @@ -26,6 +26,7 @@ void gemm( platform::DeviceContext* context) { // Note that cublas follows fortran order, so the order is different from // the cblas convention. + /* cublasOperation_t cuTransA = (transA == CblasNoTrans) ? CUBLAS_OP_N : CUBLAS_OP_T; cublasOperation_t cuTransB = @@ -34,6 +35,8 @@ void gemm( PADDLE_ENFORCE(platform::dynload::cublasSgemm( reinterpret_cast(context)->cublas_handle(), cuTransB, cuTransA, N, M, K, &alpha, B, ldb, A, lda, &beta, C, ldc)); + */ + PADDLE_THROW("not implemented now"); } template <> @@ -44,6 +47,7 @@ void gemm( const int ldc, platform::DeviceContext* context) { // Note that cublas follows fortran order, so the order is different from // the cblas convention. + /* cublasOperation_t cuTransA = (transA == CblasNoTrans) ? CUBLAS_OP_N : CUBLAS_OP_T; cublasOperation_t cuTransB = @@ -51,6 +55,8 @@ void gemm( PADDLE_ENFORCE(platform::dynload::cublasDgemm( reinterpret_cast(context)->cublas_handle(), cuTransB, cuTransA, N, M, K, &alpha, B, ldb, A, lda, &beta, C, ldc)); + */ + PADDLE_THROW("not implemented now"); } template <> diff --git a/paddle/operators/math/math_function.h b/paddle/operators/math/math_function.h index f068f4a15eec7593c257e9a9e026413c95904398..7a214e3a5a07fd92f2100a7f2637a69c4c099648 100644 --- a/paddle/operators/math/math_function.h +++ b/paddle/operators/math/math_function.h @@ -40,36 +40,23 @@ extern "C" { #include #include "paddle/framework/tensor.h" #include "paddle/platform/device_context.h" +#include "paddle/platform/enforce.h" namespace paddle { namespace operators { namespace math { template -void gemm(const CBLAS_TRANSPOSE transA, - const CBLAS_TRANSPOSE transB, - const int M, - const int N, - const int K, - const T alpha, - const T* A, - const int lda, - const T* B, - const int ldb, - const T beta, - T* C, - const int ldc, - platform::DeviceContext* context); +void gemm(const CBLAS_TRANSPOSE transA, const CBLAS_TRANSPOSE transB, + const int M, const int N, const int K, const T alpha, const T* A, + const int lda, const T* B, const int ldb, const T beta, T* C, + const int ldc, platform::DeviceContext* context); // matrix multiply with continous memory template -void matmul(const framework::Tensor& in1, - bool in1_T, - const framework::Tensor& in2, - bool in2_T, - float alpha, - framework::Tensor* out, - float beta, +void matmul(const framework::Tensor& in1, bool in1_T, + const framework::Tensor& in2, bool in2_T, float alpha, + framework::Tensor* out, float beta, platform::DeviceContext* context); } // namespace math diff --git a/paddle/operators/mul_op.cu b/paddle/operators/mul_op.cu index 7435b74bd87a392a27c1ae7f4a8d2416526a40d0..aac5a6936ebb81d3dfb93d0cf39b47ff6f75e19c 100644 --- a/paddle/operators/mul_op.cu +++ b/paddle/operators/mul_op.cu @@ -15,4 +15,5 @@ #define EIGEN_USE_GPU #include "paddle/operators/mul_op.h" -REGISTER_OP_GPU_KERNEL(mul, ops::MulKernel); +// REGISTER_OP_GPU_KERNEL(mul, ops::MulKernel);