MKLDNNMatrix.h 4.8 KB
Newer Older
T
tensor-tang 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
/* Copyright (c) 2017 PaddlePaddle Authors. All Rights Reserve.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

#pragma once

T
tensor-tang 已提交
17 18
#include <vector>
#include "Matrix.h"
T
tensor-tang 已提交
19 20 21 22 23 24 25 26 27 28 29 30
#include "mkldnn.hpp"
#include "paddle/parameter/Parameter.h"

namespace paddle {

class MKLDNNMatrix;
typedef std::shared_ptr<MKLDNNMatrix> MKLDNNMatrixPtr;

/**
 * @brief MKLDNN Matrix.
 *
 */
T
tensor-tang 已提交
31
class MKLDNNMatrix : public CpuMatrix, public mkldnn::memory {
T
tensor-tang 已提交
32
public:
33 34 35 36
  MKLDNNMatrix(CpuMatrixPtr m, mkldnn::memory::primitive_desc pd)
      : CpuMatrix(m->getData(), m->getHeight(), m->getWidth(), false),
        mkldnn::memory(pd, m->getData()),
        m_(m) {}
T
tensor-tang 已提交
37

T
tensor-tang 已提交
38 39
  ~MKLDNNMatrix() {}

40 41 42 43 44 45 46 47
  /**
   * Create MKLDNNMatrix from a MatrixPtr and memory primitive_desc
   */
  static MKLDNNMatrixPtr create(MatrixPtr m, mkldnn::memory::primitive_desc pd);

  /**
   * Create MKLDNNMatrix from a MatrixPtr and memory details info
   */
T
tensor-tang 已提交
48
  static MKLDNNMatrixPtr create(
49
      MatrixPtr m,
T
tensor-tang 已提交
50 51 52 53 54
      mkldnn::memory::dims dims,
      mkldnn::memory::format fmt,
      mkldnn::engine& eg,
      mkldnn::memory::data_type dtype = mkldnn::memory::data_type::f32);

55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73
  /**
   * Create Memory descriptor.
   * default with any format and f32 dtype
   */
  static mkldnn::memory::desc createMemoryDesc(
      const mkldnn::memory::dims& dims,
      const mkldnn::memory::format& fmt = mkldnn::memory::format::any,
      const mkldnn::memory::data_type& dtype = mkldnn::memory::data_type::f32) {
    return mkldnn::memory::desc(dims, dtype, fmt);
  }

  /**
   * Create reorder primitive.
   */
  static std::shared_ptr<mkldnn::reorder> createReorder(
      const MKLDNNMatrixPtr& src,
      const MKLDNNMatrixPtr& dst,
      bool checkData = true);

74
public:
T
tensor-tang 已提交
75 76
  /**
   * Reorder this MKLDNNMatrix from other format.
T
refine  
tensor-tang 已提交
77 78 79
   * Support inplace reorder.
   * @note: this function would only reorder the data layout.
   *        will NOT change this original dim or format info
T
tensor-tang 已提交
80 81 82 83 84 85 86
   */
  void reorderDataFrom(const MKLDNNMatrixPtr& m,
                       memory::format srcFmt,
                       memory::dims targetDim);

  /**
   * Reorder this MKLDNNMatrix to other format.
T
refine  
tensor-tang 已提交
87 88 89
   * Support inplace reorder.
   * @note: this function would only reorder the data layout.
   *        will NOT change the dst dim or format info
T
tensor-tang 已提交
90 91 92 93 94
   */
  void reorderDataTo(const MKLDNNMatrixPtr& m,
                     memory::format dstFmt,
                     memory::dims targetDim);

95 96 97 98 99 100 101
  /**
   * Dimensionality reduction.
   * Change format "nchw --> nc" or "oihw --> oi" if the h and w are both 1
   */
  void downSpatial();

  /**
102
   * set the memory data handle.
103 104 105
   * Caution: This will not check the buffer size of the data,
   *          it should be coverd by user.
   */
106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124
  void setData(real* data) {
    set_data_handle(data);
    CpuMatrix::setData(data);
    m_.reset();
  }

  /**
   * override Matrix::getData
   * check data before return
   */
  real* getData() override {
    CHECK_EQ((void*)data_, get_data_handle());
    return data_;
  }

  const real* getData() const override {
    CHECK_EQ((void*)data_, get_data_handle());
    return data_;
  }
125

T
tensor-tang 已提交
126
  /**
T
tensor-tang 已提交
127
   * Get primitive descriptor.
T
tensor-tang 已提交
128
   */
T
refine  
tensor-tang 已提交
129 130 131
  mkldnn::memory::primitive_desc getPrimitiveDesc() {
    return this->get_primitive_desc();
  }
T
tensor-tang 已提交
132

T
tensor-tang 已提交
133
  /**
T
tensor-tang 已提交
134
   * Get memory descriptor.
T
tensor-tang 已提交
135
   */
T
refine  
tensor-tang 已提交
136
  mkldnn::memory::desc getMemoryDesc() { return getPrimitiveDesc().desc(); }
T
tensor-tang 已提交
137 138

  /**
139
   * Get dimensions.
T
tensor-tang 已提交
140
   */
T
tensor-tang 已提交
141
  mkldnn::memory::dims getDims() {
T
refine  
tensor-tang 已提交
142
    mkldnn::memory::desc md = getMemoryDesc();
143 144
    const int* src = md.data.dims;
    int ndims = md.data.ndims;
T
tensor-tang 已提交
145 146 147 148 149 150 151
    mkldnn::memory::dims dst;
    dst.resize(ndims);
    for (int i = 0; i < ndims; ++i) {
      dst[i] = src[i];
    }
    return dst;
  }
T
tensor-tang 已提交
152

T
tensor-tang 已提交
153 154 155 156
  /**
   * Get format.
   */
  mkldnn::memory::format getFormat() {
T
refine  
tensor-tang 已提交
157
    return (mkldnn::memory::format)(getMemoryDesc().data.format);
T
tensor-tang 已提交
158 159 160
  }

  /**
161
   * Get memory data type.
T
tensor-tang 已提交
162
   */
163
  mkldnn::memory::data_type getDtype() {
T
refine  
tensor-tang 已提交
164
    return (mkldnn::memory::data_type)(getMemoryDesc().data.data_type);
165 166 167 168 169
  }

  /**
   * Get engine.
   */
T
refine  
tensor-tang 已提交
170
  mkldnn::engine getEngine() { return getPrimitiveDesc().get_engine(); }
T
tensor-tang 已提交
171 172 173

protected:
  /**
T
refine  
tensor-tang 已提交
174 175
   * Do reorder once.
   * Can support inplace.
T
tensor-tang 已提交
176 177 178 179 180 181
   */
  void reorderOnce(void* srcData,
                   void* dstData,
                   memory::format srcFmt,
                   memory::format dstFmt,
                   memory::dims dm);
182 183 184 185

private:
  // save the CpuMatrixPtr in case the buffer released outside
  CpuMatrixPtr m_;
T
tensor-tang 已提交
186 187 188
};

}  // namespace paddle