/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #pragma once #include #include "TensorShape.h" #include "TensorType.h" #include "paddle/math/CpuSparseMatrix.h" #include "paddle/math/Matrix.h" #include "paddle/math/SparseMatrix.h" namespace paddle { enum BufferType { TENSOR_NORMAL = 0, TENSOR_SEQUENCE_ID = 1, TENSOR_SEQUENCE_DATA = 2, TENSOR_SPARSE = 3 }; enum SparseDataType { SPARSE_NO_VALUE = 0, // do not need value pointer, all values are 1 SPARSE_FLOAT_VALUE = 1 }; enum SparseDataFormat { SPARSE_CSR_FORMAT = 0, SPARSE_CSC_FORMAT = 1 }; class BufferArg; class SequenceArg; class SparseMatrixArg; typedef std::shared_ptr BufferArgPtr; /** * \brief BufferArg used as the argument type of Function. * * The arguments of the Paddle Function have four Buffer types. * 1. BufferArg for a dense Buffer of any dimension. * 2. SequenceIdArg for a Buffer of sequence start positions. * 3. SequenceArg for a Buffer of sequence data. * 4. SparseMatrixArg for a Buffer of sparse matrix. * * There is an ArgType property for the BufferArg used as Function Output. * Whether the result of the Function calculation is assigned to the * output Buffer or added to the output Buffer is determined by the * argType_ property of the output BufferArg. */ // ArgType is only used by output BufferArg. // For input argument, argType_ is ignored. // For output argument, need to set the argType_ of the BufferArg. enum ArgType { UNSPECIFIED = 0, ASSIGN_TO = 1, ADD_TO = 2, }; class BufferArg { public: void setArgType(ArgType argType) { argType_ = argType; } ArgType getArgType() const { return argType_; } public: BufferArg(void* buf, ValueType valueType, const TensorShape& shape, ArgType argType = UNSPECIFIED) : buf_(buf), valueType_(valueType), shape_(shape), argType_(argType) {} BufferArg(void* buf, ValueType valueType) : buf_(buf), valueType_(valueType) {} BufferArg(const Matrix& matrix, ArgType argType = UNSPECIFIED) : buf_( const_cast(reinterpret_cast(matrix.getData()))), valueType_(DataType::value), shape_(2), argType_(argType) { shape_.setDim(0, matrix.getHeight()); shape_.setDim(1, matrix.getWidth()); } BufferArg(const Matrix& matrix, const TensorShape& shape, ArgType argType = UNSPECIFIED) : buf_( const_cast(reinterpret_cast(matrix.getData()))), valueType_(DataType::value), shape_(shape), argType_(argType) { CHECK_EQ(matrix.getElementCnt(), shape.getElements()); } BufferArg(const Vector& vector, ArgType argType = UNSPECIFIED) : buf_( const_cast(reinterpret_cast(vector.getData()))), valueType_(DataType::value), shape_(1), argType_(argType) { shape_.setDim(0, vector.getSize()); } BufferArg(const IVector& vector, ArgType argType = UNSPECIFIED) : buf_( const_cast(reinterpret_cast(vector.getData()))), valueType_(VALUE_TYPE_INT32), shape_(1), argType_(argType) { shape_.setDim(0, vector.getSize()); } template typename Tensor::Matrix matrix() const { CHECK(buf_); CHECK(valueType_ == DataType::value); // CHECK(deviceType_ == DType); CHECK_EQ((size_t)2, shape_.ndims()); return typename Tensor::Matrix( reinterpret_cast(buf_), shape_[0], shape_[1]); } template typename Tensor::Vector vector() const { CHECK(buf_); CHECK(valueType_ == DataType::value); // CHECK(deviceType_ == DType); CHECK_EQ((size_t)1, shape_.ndims()); return typename Tensor::Vector( shape_[0], reinterpret_cast(buf_)); } virtual ~BufferArg() {} template T* data() const { return reinterpret_cast(buf_); } void* data() const { return buf_; } ValueType valueType() const { return valueType_; } BufferType bufferType() const { return bufferType_; } const TensorShape& shape() const { return shape_; } const SequenceArg& sequence() const; const SparseMatrixArg& sparse() const; protected: void* buf_; ValueType valueType_; TensorShape shape_; BufferType bufferType_; ArgType argType_ = UNSPECIFIED; // leading dimensions. The size is dims_.size() // Dims lds_; }; // sequence start positions in a mini-batch of sequences // shape_.ndims() == 1 // valueType_ = int32 // if a < b then value_.buf_[a] < value_.buf_[b] class SequenceIdArg : public BufferArg { public: SequenceIdArg(void* buf, const TensorShape& shape, ArgType argType = UNSPECIFIED) : BufferArg(buf, VALUE_TYPE_INT32, shape, argType) { CHECK_EQ(shape_.ndims(), (size_t)1); numSeqs_ = shape_[0] - 1; } SequenceIdArg(const IVector& vector) : BufferArg(vector) { numSeqs_ = shape_[0] - 1; } ~SequenceIdArg() {} size_t numSeqs() const { return numSeqs_; } private: size_t numSeqs_; }; // sequence data class SequenceArg : public BufferArg { public: SequenceArg(void* buf, ValueType valueType, const TensorShape& shape, const SequenceIdArg& startPositions, ArgType argType = UNSPECIFIED) : BufferArg(buf, valueType, shape, argType), startPositions_(startPositions) {} SequenceArg(const Matrix& matrix, const IVector& vector, ArgType argType = UNSPECIFIED) : BufferArg(matrix, argType), startPositions_(vector) {} ~SequenceArg() {} void* getIdBuf() const { return startPositions_.data(); } size_t numSeqs() const { return startPositions_.numSeqs(); } private: SequenceIdArg startPositions_; }; // sparse matrix // valueType_ == float or double // shape_.ndims() == 2 class SparseMatrixArg : public BufferArg { public: SparseMatrixArg(void* buf, ValueType valueType, const TensorShape& shape, const BufferArg& row, const BufferArg& col, size_t nnz, SparseDataFormat format, SparseDataType type, ArgType argType = UNSPECIFIED) : BufferArg(buf, valueType, shape, argType), row_(row), col_(col), nnz_(nnz), format_(format), type_(type) { CHECK((valueType == VALUE_TYPE_FLOAT) || (valueType == VALUE_TYPE_DOUBLE)); CHECK_EQ(shape_.ndims(), (size_t)2); CHECK_EQ(row_.shape().ndims(), (size_t)1); CHECK_EQ(col_.shape().ndims(), (size_t)1); if (format == SPARSE_CSR_FORMAT) { CHECK_EQ(nnz, col.shape()[0]); } else if (format == SPARSE_CSC_FORMAT) { CHECK_EQ(nnz, row.shape()[0]); } } SparseMatrixArg(const CpuSparseMatrix& sparse, ArgType argType = UNSPECIFIED) : BufferArg(sparse, argType), row_(reinterpret_cast(sparse.getRows()), VALUE_TYPE_INT32), col_(reinterpret_cast(sparse.getCols()), VALUE_TYPE_INT32) {} SparseMatrixArg(const GpuSparseMatrix& sparse, ArgType argType = UNSPECIFIED) : BufferArg(sparse, argType), row_(reinterpret_cast(sparse.getRows()), VALUE_TYPE_INT32), col_(reinterpret_cast(sparse.getCols()), VALUE_TYPE_INT32) {} ~SparseMatrixArg() {} void* getRowBuf() const { return row_.data(); } void* getColBuf() const { return col_.data(); } size_t nnz() const { return nnz_; } SparseDataFormat dataFormat() const { return format_; } SparseDataType dataType() const { return type_; } private: BufferArg row_; BufferArg col_; size_t nnz_; SparseDataFormat format_; SparseDataType type_; }; } // namespace paddle