tensor_util.h 4.6 KB
Newer Older
1
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
D
dzhwinter 已提交
2

L
Luo Tao 已提交
3 4 5
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
D
dzhwinter 已提交
6

L
Luo Tao 已提交
7
    http://www.apache.org/licenses/LICENSE-2.0
D
dzhwinter 已提交
8

L
Luo Tao 已提交
9 10 11 12 13
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
D
dzhwinter 已提交
14 15

#pragma once
16
#include <vector>
Y
Yi Wang 已提交
17 18 19 20 21
#include "paddle/fluid/framework/data_type.h"
#include "paddle/fluid/framework/eigen.h"
#include "paddle/fluid/framework/framework.pb.h"
#include "paddle/fluid/framework/tensor.h"
#include "paddle/fluid/platform/device_context.h"
D
dzhwinter 已提交
22 23 24 25

namespace paddle {
namespace framework {

Y
Yi Wang 已提交
26
void TensorCopy(const Tensor& src, const platform::Place& dst_place,
F
fengjiayi 已提交
27 28
                const platform::DeviceContext& ctx, Tensor* dst,
                bool sync = false);
Y
Yi Wang 已提交
29 30
void TensorCopy(const Tensor& src, const platform::Place& dst_place,
                Tensor* dst);
D
dzhwinter 已提交
31

Y
Yi Wang 已提交
32 33 34 35 36
template <typename T>
void TensorFromVector(const std::vector<T>& src,
                      const platform::DeviceContext& ctx, Tensor* dst);
template <typename T>
void TensorFromVector(const std::vector<T>& src, Tensor* dst);
D
dzhwinter 已提交
37

Y
Yi Wang 已提交
38 39 40 41 42
template <typename T>
void TensorToVector(const Tensor& src, const platform::DeviceContext& ctx,
                    std::vector<T>* dst);
template <typename T>
void TesnorToVector(const Tensor& src, std::vector<T>* dst);
D
dzhwinter 已提交
43

Y
Yi Wang 已提交
44 45
bool TensorContainsNAN(const framework::Tensor& tensor);
bool TensorContainsInf(const framework::Tensor& tensor);
D
dzhwinter 已提交
46

Y
Yi Wang 已提交
47 48 49 50
void TensorToStream(std::ostream& os, const Tensor& tensor,
                    const platform::DeviceContext& dev_ctx);
void TensorFromStream(std::istream& is, Tensor* tensor,
                      const platform::DeviceContext& dev_ctx);
D
dzhwinter 已提交
51

Y
Yi Wang 已提交
52 53 54
//
// The implementation of template functions.
//
D
dzhwinter 已提交
55

D
dzhwinter 已提交
56
template <typename T>
Y
Yi Wang 已提交
57 58
void TensorFromVector(const std::vector<T>& src,
                      const platform::DeviceContext& ctx, Tensor* dst) {
D
dzhwinter 已提交
59 60 61 62 63 64 65 66 67 68 69 70 71 72
  auto dst_place = ctx.GetPlace();
  auto src_ptr = static_cast<const void*>(src.data());
  platform::CPUPlace src_place;
  dst->Resize({static_cast<int64_t>(src.size())});
  auto dst_ptr = static_cast<void*>(dst->mutable_data<T>(dst_place));
  auto size = src.size() * sizeof(T);

  if (platform::is_cpu_place(dst_place)) {
    memory::Copy(boost::get<platform::CPUPlace>(dst_place), dst_ptr, src_place,
                 src_ptr, size);
  }
#ifdef PADDLE_WITH_CUDA
  else if (platform::is_gpu_place(dst_place)) {  // NOLINT
    memory::Copy(
D
dzhwinter 已提交
73
        boost::get<platform::CUDAPlace>(dst_place), dst_ptr, src_place, src_ptr,
D
dzhwinter 已提交
74 75 76 77 78 79
        size,
        reinterpret_cast<const platform::CUDADeviceContext&>(ctx).stream());
  }
#endif
}

D
dzhwinter 已提交
80
template <typename T>
Y
Yi Wang 已提交
81
void TensorFromVector(const std::vector<T>& src, Tensor* dst) {
D
dzhwinter 已提交
82 83 84 85 86 87 88 89 90 91
  platform::CPUPlace dst_place = platform::CPUPlace();
  auto src_ptr = static_cast<const void*>(src.data());
  platform::CPUPlace src_place;
  dst->Resize({static_cast<int64_t>(src.size())});
  auto dst_ptr = static_cast<void*>(dst->mutable_data<T>(dst_place));
  auto size = src.size() * sizeof(T);

  memory::Copy(dst_place, dst_ptr, src_place, src_ptr, size);
}

D
dzhwinter 已提交
92
template <typename T>
Y
Yi Wang 已提交
93 94
void TensorToVector(const Tensor& src, const platform::DeviceContext& ctx,
                    std::vector<T>* dst) {
D
dzhwinter 已提交
95 96 97 98 99 100 101 102
  auto src_ptr = static_cast<const void*>(src.data<T>());
  auto size = src.numel() * sizeof(T);

  platform::CPUPlace dst_place;
  dst->resize(src.numel());
  auto dst_ptr = static_cast<void*>(dst->data());

  if (platform::is_cpu_place(src.place())) {
103 104
    memory::Copy(dst_place, dst_ptr,
                 boost::get<platform::CPUPlace>(src.place()), src_ptr, size);
D
dzhwinter 已提交
105 106 107 108
  }
#ifdef PADDLE_WITH_CUDA
  else if (platform::is_gpu_place(src.place())) {  // NOLINT
    memory::Copy(
D
dzhwinter 已提交
109
        dst_place, dst_ptr, boost::get<platform::CUDAPlace>(src.place()),
110
        src_ptr, size,
D
dzhwinter 已提交
111 112 113 114 115
        reinterpret_cast<const platform::CUDADeviceContext&>(ctx).stream());
  }
#endif
}

D
dzhwinter 已提交
116
template <typename T>
Y
Yi Wang 已提交
117
void TensorToVector(const Tensor& src, std::vector<T>* dst) {
D
dzhwinter 已提交
118 119 120 121 122 123 124 125 126 127 128 129 130
  auto src_ptr = static_cast<const void*>(src.data<T>());
  auto size = src.numel() * sizeof(T);

  platform::CPUPlace dst_place;
  dst->resize(src.numel());
  auto dst_ptr = static_cast<void*>(dst->data());

  PADDLE_ENFORCE(platform::is_cpu_place(src.place()));

  memory::Copy(dst_place, dst_ptr, boost::get<platform::CPUPlace>(src.place()),
               src_ptr, size);
}

D
dzhwinter 已提交
131 132
}  // namespace framework
}  // namespace paddle