未验证 提交 71b1c397 编写于 作者: X Xin Pan 提交者: GitHub

Merge pull request #11874 from panyx0718/move_trainer

Move trainer and utils api
......@@ -18,6 +18,8 @@ learning to many products at Baidu.
Our vision is to enable deep learning for everyone via PaddlePaddle.
Please refer to our [release announcement](https://github.com/PaddlePaddle/Paddle/releases) to track the latest feature of PaddlePaddle.
### Lastest PaddlePaddle Version: [Fluid](https://github.com/PaddlePaddle/Paddle/tree/develop/paddle/fluid)
## Features
- **Flexibility**
......
......@@ -52,7 +52,7 @@ In `trainer_internal.cpp:L93 trainOneBatch`:
When doing actual network forward and backward, at the beginning of each batch, the trainer will try to download one row of data from pserver.
In `trainer/RemoteParameterUpdater.cpp`: `parameterUpdater_->getParametersRemote();`:
In `legacy/trainer/RemoteParameterUpdater.cpp`: `parameterUpdater_->getParametersRemote();`:
```c++
if (fullSize) {
......
......@@ -18,20 +18,20 @@ Figure 1. PaddlePaddle on IA
具体的完成状态可以参见[这里](https://github.com/PaddlePaddle/Paddle/projects/21)
## Contents
- [Overview](#overview)
- [Actions](#actions)
- [CMake](#cmake)
- [Matrix](#matrix)
- [Layers](#layers)
- [Activations](#activations)
- [Parameters](#parameters)
- [Gradients](#gradients)
- [Unit Tests](#unit-tests)
- [Python API](#python-api)
- [Benchmarking](#benchmarking)
- [Others](#others)
- [Design Concerns](#design-concerns)
- [Overview](#overview)
- [Actions](#actions)
- [CMake](#cmake)
- [Matrix](#matrix)
- [Layers](#layers)
- [Activations](#activations)
- [Parameters](#parameters)
- [Gradients](#gradients)
- [Unit Tests](#unit-tests)
- [Python API](#python-api)
- [Benchmarking](#benchmarking)
- [Others](#others)
- [Design Concerns](#design-concerns)
## Overview
......@@ -218,20 +218,20 @@ if use_mkldnn
我们总结出一些特别需要注意的点:
1. 使用**deviceId_**。为了尽可能少的在父类Layer中添加变量或者函数,
我们决定使用已有的`deviceId_`变量来区分layer的属性,定义`-2``MKLDNNLayer`特有的设备ID。
2. 重写父类Layer的**init**函数,修改`deviceId_``-2`,代表这个layer是用于跑在MKL-DNN的环境下。
我们决定使用已有的`deviceId_`变量来区分layer的属性,定义`-2``MKLDNNLayer`特有的设备ID。
2. 重写父类Layer的**init**函数,修改`deviceId_``-2`,代表这个layer是用于跑在MKL-DNN的环境下。
3. 创建`MKLDNNBase`,定义一些除了layer和memory相关的类和函数。
包括MKL-DNN会用到`MKLDNNStream``CPUEngine`,和未来可能还会用到`FPGAEngine`等。
包括MKL-DNN会用到`MKLDNNStream``CPUEngine`,和未来可能还会用到`FPGAEngine`等。
4. 如果MKL-DNN layer的后面接有cpu device,那么就会使`output_.value``extOutVal_`共享内存,
同时数据格式就是`NCHW`,这样下一个cpu device就能拿到正确的数据。
在有普通的CPU layer时, `extOutVal_``extOutGrad_`的格式始终是`NCHW`或者`NC`
## References
1. [MKL small library](https://github.com/01org/mkl-dnn#linking-your-application)[Intel MKL](https://software.intel.com/en-us/mkl)的一个子集。
主要包括了深度学习相关的数学原语与操作,一般由MKL-DNN在发布[新版本](https://github.com/01org/mkl-dnn/releases)时一起更新。
主要包括了深度学习相关的数学原语与操作,一般由MKL-DNN在发布[新版本](https://github.com/01org/mkl-dnn/releases)时一起更新。
2. [MKL-DNN System Requirements](https://github.com/01org/mkl-dnn#system-requirements)
目前在PaddlePaddle中,仅会在支持AVX2指令集及以上的机器才使用MKL-DNN。
3. [原来的方案](https://github.com/PaddlePaddle/Paddle/pull/3096)会引入**nextLayer**的信息。
但是在PaddlePaddle中,无论是重构前的layer还是重构后的op,都不会想要知道next layer/op的信息。
但是在PaddlePaddle中,无论是重构前的layer还是重构后的op,都不会想要知道next layer/op的信息。
4. MKL-DNN的高性能格式与PaddlePaddle原有的`NCHW`不同(PaddlePaddle中的cuDNN部分使用的也是`NCHW`,所以不存在这个问题)。
所以需要引入一个转换方法,并且只需要在必要的时候转换这种格式,才能更好的发挥MKL-DNN的性能。
所以需要引入一个转换方法,并且只需要在必要的时候转换这种格式,才能更好的发挥MKL-DNN的性能。
......@@ -339,7 +339,7 @@ If you are creating a new file for the test, such as :code:`paddle/legacy/gserve
Implement Python Wrapper
========================
Implementing Python wrapper allows us to use the added layer in configuration files. All the Python wrappers are in file :code:`python/paddle/trainer/config_parser.py`. An example of the Python wrapper for fully connected layer is listed below. It has the following steps:
Implementing Python wrapper allows us to use the added layer in configuration files. All the Python wrappers are in file :code:`python/paddle/legacy/trainer/config_parser.py`. An example of the Python wrapper for fully connected layer is listed below. It has the following steps:
- Use :code:`@config_layer('fc')` at the decorator for all the Python wrapper class. :code:`fc` is the identifier of the layer.
- Implements :code:`__init__` constructor function.
......
if(NOT WITH_FLUID_ONLY)
add_subdirectory(legacy/cuda)
add_subdirectory(legacy/function)
add_subdirectory(utils)
add_subdirectory(legacy/utils)
add_subdirectory(legacy/math)
add_subdirectory(legacy/gserver)
add_subdirectory(legacy/parameter)
......@@ -10,7 +10,7 @@ if(NOT WITH_FLUID_ONLY)
add_subdirectory(legacy/capi)
else()
add_subdirectory(legacy/pserver)
add_subdirectory(trainer)
add_subdirectory(legacy/trainer)
add_subdirectory(scripts)
if(WITH_C_API)
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#include <vector>
#include "paddle/fluid/framework/eigen.h"
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
namespace paddle {
namespace operators {
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#include "paddle/fluid/framework/lod_tensor.h"
#include "paddle/fluid/framework/tensor_util.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
#define ARITHMETIC_KERNEL(op_type, sign) \
__global__ void op_type(const half* in1, const half* in2, half* out) { \
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#include "PaddleAPI.h"
#include "PaddleAPIPrivate.h"
#include "paddle/trainer/Trainer.h"
#include "paddle/legacy/trainer/Trainer.h"
struct ParameterConfigPrivate {
paddle::ParameterPtr parameter;
......
......@@ -198,5 +198,5 @@ namespace std {
%ignore ParameterConfigPrivate;
%ignore OptimizationConfigPrivate;
%ignore ParameterTraverseCallbackPrivate;
%include "utils/GlobalConstants.h"
%include "legacy/utils/GlobalConstants.h"
%include "legacy/api/PaddleAPI.h"
......@@ -20,8 +20,8 @@ limitations under the License. */
#include <string>
#include <vector>
#include "paddle/legacy/gserver/gradientmachines/GradientMachine.h"
#include "paddle/utils/Common.h"
#include "paddle/utils/GlobalConstants.h"
#include "paddle/legacy/utils/Common.h"
#include "paddle/legacy/utils/GlobalConstants.h"
/// Import PaddlePaddle's enumeration into global namespace.
using namespace paddle::enumeration_wrapper; // NOLINT
......
......@@ -17,7 +17,7 @@ limitations under the License. */
#include "paddle/legacy/gserver/evaluators/Evaluator.h"
#include "paddle/legacy/gserver/gradientmachines/GradientMachine.h"
#include "paddle/legacy/parameter/ParameterUpdaterBase.h"
#include "paddle/trainer/TrainerConfigHelper.h"
#include "paddle/legacy/trainer/TrainerConfigHelper.h"
struct GradientMachinePrivate {
std::shared_ptr<paddle::GradientMachine> machine;
......
......@@ -16,10 +16,10 @@ limitations under the License. */
#include "PaddleAPIPrivate.h"
#ifndef PADDLE_WITHOUT_GOLANG
#include "paddle/trainer/NewRemoteParameterUpdater.h"
#include "paddle/legacy/trainer/NewRemoteParameterUpdater.h"
#endif
#include "paddle/trainer/RemoteParameterUpdater.h"
#include "paddle/trainer/ThreadParameterUpdater.h"
#include "paddle/legacy/trainer/RemoteParameterUpdater.h"
#include "paddle/legacy/trainer/ThreadParameterUpdater.h"
ParameterUpdater::ParameterUpdater() : m(new ParameterUpdaterPrivate()) {}
......
......@@ -19,7 +19,7 @@ limitations under the License. */
#include "PaddleAPI.h"
#include "paddle/legacy/gserver/gradientmachines/GradientMachine.h"
#include "paddle/legacy/parameter/Argument.h"
#include "paddle/utils/Flags.h"
#include "paddle/legacy/utils/Flags.h"
// used to represent partial sequence
struct Path {
......
......@@ -20,10 +20,10 @@ limitations under the License. */
#include <memory>
#include "paddle/legacy/gserver/gradientmachines/NeuralNetwork.h"
#include "paddle/trainer/ParamUtil.h"
#include "paddle/trainer/Trainer.h"
#include "paddle/trainer/TrainerInternal.h"
#include "paddle/utils/Flags.h"
#include "paddle/legacy/trainer/ParamUtil.h"
#include "paddle/legacy/trainer/Trainer.h"
#include "paddle/legacy/trainer/TrainerInternal.h"
#include "paddle/legacy/utils/Flags.h"
using paddle::real;
......
......@@ -15,10 +15,10 @@ limitations under the License. */
#include "PaddleAPI.h"
#include "paddle/legacy/parameter/Parameter.h"
#include "paddle/utils/Common.h"
#include "paddle/utils/Flags.h"
#include "paddle/utils/PythonUtil.h"
#include "paddle/utils/Util.h"
#include "paddle/legacy/utils/Common.h"
#include "paddle/legacy/utils/Flags.h"
#include "paddle/legacy/utils/PythonUtil.h"
#include "paddle/legacy/utils/Util.h"
#include <algorithm>
#include <iostream>
......
......@@ -18,9 +18,9 @@ limitations under the License. */
#include <vector>
#include "capi_private.h"
#include "main.h"
#include "paddle/trainer/TrainerConfigHelper.h"
#include "paddle/utils/Excepts.h"
#include "paddle/utils/PythonUtil.h"
#include "paddle/legacy/trainer/TrainerConfigHelper.h"
#include "paddle/legacy/utils/Excepts.h"
#include "paddle/legacy/utils/PythonUtil.h"
static void initPaddle(int argc, char** argv) {
paddle::initMain(argc, argv);
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#include <functional>
#include "capi.h"
#include "gtest/gtest.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/legacy/utils/ThreadLocal.h"
static std::vector<paddle_real> randomBuffer(size_t bufSize) {
auto& eng = paddle::ThreadLocalRandomEngine::get();
......
......@@ -14,12 +14,12 @@ limitations under the License. */
#include <gtest/gtest.h>
#include <paddle/legacy/gserver/gradientmachines/GradientMachine.h>
#include <paddle/trainer/TrainerConfigHelper.h>
#include <paddle/legacy/trainer/TrainerConfigHelper.h>
#include <stdlib.h>
#include <string.h>
#include <type_traits>
#include "capi.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/legacy/utils/ThreadLocal.h"
static std::vector<paddle_real> randomBuffer(size_t bufSize) {
auto& eng = paddle::ThreadLocalRandomEngine::get();
......
......@@ -208,7 +208,7 @@ typedef struct {
#include <cuda_runtime.h>
#include "paddle/legacy/cuda/include/hl_cuda.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
extern __thread bool g_sync_flag;
extern __thread cudaStream_t default_stream;
......
......@@ -18,7 +18,7 @@ limitations under the License. */
#ifdef __NVCC__
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
/*
* threads(framePerBlock, batchPerBlock)
......
......@@ -18,7 +18,7 @@ limitations under the License. */
#ifdef __NVCC__
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
#include "hl_device_functions.cuh"
/*
......
......@@ -18,7 +18,7 @@ limitations under the License. */
#define HL_GPU_MATRIX_KERNEL_CUH_
#include <algorithm>
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
#include "hl_base.h"
#ifdef __NVCC__
......
......@@ -18,7 +18,7 @@ limitations under the License. */
#include "hl_cuda.ph"
#include "hl_matrix_base.cuh"
#include "hl_thread.ph"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
/**
* @brief matrix row operator.
......
......@@ -16,8 +16,8 @@ limitations under the License. */
#include <sys/time.h>
#include "hl_cuda.h"
#include "hl_thread.ph"
#include "paddle/utils/DynamicLoader.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/DynamicLoader.h"
#include "paddle/legacy/utils/Logging.h"
namespace dynload {
......
......@@ -17,8 +17,8 @@ limitations under the License. */
#include <gflags/gflags.h>
#include "hl_cuda_cudnn.ph"
#include "hl_thread.ph"
#include "paddle/utils/DynamicLoader.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/DynamicLoader.h"
#include "paddle/legacy/utils/Logging.h"
DEFINE_int32(cudnn_conv_workspace_limit_in_mb,
4096,
......
......@@ -23,8 +23,8 @@ limitations under the License. */
#include <unistd.h>
#include "hl_cuda.ph"
#include "hl_thread.ph"
#include "paddle/utils/Logging.h"
#include "paddle/utils/DynamicLoader.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/DynamicLoader.h"
// clang-format on
namespace dynload {
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include "hl_base.h"
#include "hl_cuda_cublas.h"
#include "hl_device_functions.cuh"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
typedef hppl::Active<real>::forward t_forward;
typedef hppl::Active<real>::backward t_backward;
......
......@@ -20,7 +20,7 @@ limitations under the License. */
#include "hl_matrix_ops.cuh"
#include "hl_sequence.h"
#include "hl_sparse.ph"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
DEFINE_MATRIX_UNARY_OP(Zero, a = 0);
DEFINE_MATRIX_TERNARY_PARAMETER_OP(_add, TWO_PARAMETER, c = p1 * a + p2 * b);
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#include "hl_base.h"
#include "hl_device_functions.cuh"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
__global__ void KeMaxSequenceForward(real* input,
const int* sequence,
......
......@@ -18,7 +18,7 @@ limitations under the License. */
#include "hl_matrix_ops.cuh"
#include "hl_sparse.h"
#include "hl_sparse.ph"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
DEFINE_MATRIX_UNARY_PARAMETER_OP(mul_scalar, ONE_PARAMETER, a = a * p);
DEFINE_MATRIX_UNARY_OP(Zero, a = 0);
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#include "hl_base.h"
#include "hl_cuda.h"
#include "hl_device_functions.cuh"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
template <int blockDimX, int blockDimY, int gridDimX, bool AddRow>
__global__ void KeMatrixAddRows(real* output,
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#include "paddle/legacy/cuda/include/hl_base.h"
#include "paddle/legacy/cuda/include/hl_sparse.ph"
#include "paddle/legacy/cuda/include/hl_top_k.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
// using namespace hppl;
......
......@@ -14,8 +14,8 @@ limitations under the License. */
#include "hl_warpctc_wrap.h"
#include <mutex>
#include "paddle/utils/DynamicLoader.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/DynamicLoader.h"
#include "paddle/legacy/utils/Logging.h"
namespace dynload {
......
......@@ -18,9 +18,9 @@ limitations under the License. */
#include <vector>
#include "BufferArg.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/Any.h"
#include "paddle/utils/ClassRegistrar.h"
#include "paddle/utils/Error.h"
#include "paddle/legacy/utils/Any.h"
#include "paddle/legacy/utils/ClassRegistrar.h"
#include "paddle/legacy/utils/Error.h"
namespace paddle {
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#include "MulOp.h"
#include "GemmFunctor.h"
#include "paddle/legacy/math/SIMDFunctions.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/legacy/utils/ThreadLocal.h"
namespace {
inline void vecAddTo(real* a, const real* b, real scaleB, size_t len) {
......
......@@ -21,8 +21,8 @@ limitations under the License. */
#include <thread>
#include <type_traits>
#include "paddle/legacy/parameter/Argument.h"
#include "paddle/utils/ClassRegistrar.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/ClassRegistrar.h"
#include "paddle/legacy/utils/Logging.h"
#ifdef PADDLE_WITH_MKLDNN
#include "MKLDNNActivation.h"
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#pragma once
#include <string>
#include <vector>
#include "paddle/utils/Error.h"
#include "paddle/legacy/utils/Error.h"
namespace paddle {
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#include "MKLDNNActivation.h"
#include "mkldnn.hpp"
#include "paddle/utils/ClassRegistrar.h"
#include "paddle/legacy/utils/ClassRegistrar.h"
namespace paddle {
......
......@@ -16,10 +16,10 @@ limitations under the License. */
#include <unistd.h>
#include <algorithm>
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/utils/StringUtil.h"
#include "paddle/utils/Util.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
#include "paddle/legacy/utils/StringUtil.h"
#include "paddle/legacy/utils/Util.h"
namespace paddle {
......
......@@ -29,13 +29,13 @@ limitations under the License. */
#include "paddle/legacy/math/SparseMatrix.h"
#include "paddle/legacy/math/Vector.h"
#include "paddle/legacy/parameter/Argument.h"
#include "paddle/utils/ClassRegistrar.h"
#include "paddle/utils/Common.h"
#include "paddle/utils/Locks.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Queue.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/utils/Util.h"
#include "paddle/legacy/utils/ClassRegistrar.h"
#include "paddle/legacy/utils/Common.h"
#include "paddle/legacy/utils/Locks.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Queue.h"
#include "paddle/legacy/utils/ThreadLocal.h"
#include "paddle/legacy/utils/Util.h"
namespace paddle {
/**
......
......@@ -14,8 +14,8 @@ limitations under the License. */
#include "MultiDataProvider.h"
#include <algorithm>
#include "paddle/utils/Logging.h"
#include "paddle/utils/Util.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Util.h"
namespace paddle {
......
......@@ -13,9 +13,9 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "PyDataProvider.h"
#include "paddle/utils/Common.h"
#include "paddle/utils/PythonUtil.h"
#include "paddle/utils/Util.h"
#include "paddle/legacy/utils/Common.h"
#include "paddle/legacy/utils/PythonUtil.h"
#include "paddle/legacy/utils/Util.h"
namespace paddle {
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#pragma once
#include <paddle/utils/PythonUtil.h>
#include <paddle/legacy/utils/PythonUtil.h>
#include "DataFormat.pb.h"
#include "DataProvider.h"
......
......@@ -25,9 +25,9 @@ limitations under the License. */
#include "DataProvider.h"
#include "paddle/utils/Locks.h"
#include "paddle/utils/PythonUtil.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Locks.h"
#include "paddle/legacy/utils/PythonUtil.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#include "Evaluator.h"
#include "paddle/legacy/gserver/gradientmachines/NeuralNetwork.h"
#include "paddle/utils/StringUtil.h"
#include "paddle/legacy/utils/StringUtil.h"
namespace paddle {
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include <vector>
#include "paddle/legacy/math/Vector.h"
#include "paddle/utils/StringUtil.h"
#include "paddle/legacy/utils/StringUtil.h"
#include "Evaluator.h"
......
......@@ -14,8 +14,8 @@ limitations under the License. */
#include "paddle/legacy/gserver/evaluators/Evaluator.h"
#include "paddle/legacy/gserver/gradientmachines/NeuralNetwork.h"
#include "paddle/utils/Stat.h"
#include "paddle/utils/StringUtil.h"
#include "paddle/legacy/utils/Stat.h"
#include "paddle/legacy/utils/StringUtil.h"
DECLARE_int32(trainer_id);
......
......@@ -18,8 +18,8 @@ limitations under the License. */
#include "ModelConfig.pb.h"
#include "paddle/legacy/parameter/Argument.h"
#include "paddle/legacy/pserver/ParameterClient2.h"
#include "paddle/utils/ClassRegistrar.h"
#include "paddle/utils/Error.h"
#include "paddle/legacy/utils/ClassRegistrar.h"
#include "paddle/legacy/utils/Error.h"
namespace paddle {
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#include "GradientMachine.h"
#include <fstream>
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
#include "NeuralNetwork.h"
#include "hl_gpu.h"
......
......@@ -24,7 +24,7 @@ limitations under the License. */
#include "paddle/legacy/math/Matrix.h"
#include "paddle/legacy/parameter/Parameter.h"
#include "paddle/legacy/parameter/ParameterUpdaterBase.h"
#include "paddle/utils/Thread.h"
#include "paddle/legacy/utils/Thread.h"
#ifndef PADDLE_MOBILE_INFERENCE
#include "paddle/legacy/gserver/evaluators/Evaluator.h"
......
......@@ -14,9 +14,9 @@ limitations under the License. */
#include "MultiGradientMachine.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
#include "NeuralNetwork.h"
#include "ParallelNeuralNetwork.h"
......
......@@ -19,8 +19,8 @@ limitations under the License. */
#include "GradientMachine.h"
#include "hl_gpu.h"
#include "paddle/utils/Locks.h"
#include "paddle/utils/Queue.h"
#include "paddle/legacy/utils/Locks.h"
#include "paddle/legacy/utils/Queue.h"
namespace paddle {
......
......@@ -13,8 +13,8 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include <algorithm>
#include "paddle/utils/Stat.h"
#include "paddle/utils/Util.h"
#include "paddle/legacy/utils/Stat.h"
#include "paddle/legacy/utils/Util.h"
#include "MultiNetwork.h"
......
......@@ -17,7 +17,7 @@ limitations under the License. */
#include "GradientMachine.h"
#include "NeuralNetwork.h"
#include "paddle/utils/Locks.h"
#include "paddle/legacy/utils/Locks.h"
namespace paddle {
......
......@@ -12,13 +12,13 @@ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/utils/Util.h"
#include "paddle/legacy/utils/Util.h"
#include "NeuralNetwork.h"
#include "hl_gpu.h"
#include "paddle/utils/CustomStackTrace.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/CustomStackTrace.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
#ifdef PADDLE_WITH_MKLDNN
#include "paddle/legacy/gserver/layers/MKLDNNLayer.h"
......
......@@ -25,7 +25,7 @@ limitations under the License. */
#include "paddle/legacy/gserver/layers/DataLayer.h"
#include "paddle/legacy/gserver/layers/Layer.h"
#include "paddle/legacy/parameter/Parameter.h"
#include "paddle/utils/ClassRegistrar.h"
#include "paddle/legacy/utils/ClassRegistrar.h"
namespace paddle {
/*
......
......@@ -12,8 +12,8 @@ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/utils/Stat.h"
#include "paddle/utils/Util.h"
#include "paddle/legacy/utils/Stat.h"
#include "paddle/legacy/utils/Util.h"
#include "ParallelNeuralNetwork.h"
......
......@@ -20,9 +20,9 @@ limitations under the License. */
#include <limits>
#include "NeuralNetwork.h"
#include "paddle/legacy/gserver/layers/AgentLayer.h"
#include "paddle/utils/Flags.h"
#include "paddle/utils/Stat.h"
#include "paddle/utils/Util.h"
#include "paddle/legacy/utils/Flags.h"
#include "paddle/legacy/utils/Stat.h"
#include "paddle/legacy/utils/Util.h"
DEFINE_string(diy_beam_search_prob_so, "", "the diy beam search cost so");
......
......@@ -18,7 +18,7 @@ limitations under the License. */
#include "GradientMachine.h"
#include "NeuralNetwork.h"
#include "paddle/utils/Locks.h"
#include "paddle/legacy/utils/Locks.h"
namespace paddle {
......
......@@ -14,9 +14,9 @@ limitations under the License. */
#include "AddtoLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/legacy/utils/ThreadLocal.h"
namespace paddle {
......
......@@ -14,9 +14,9 @@ limitations under the License. */
#include "AgentLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/legacy/utils/ThreadLocal.h"
namespace paddle {
......
......@@ -14,9 +14,9 @@ limitations under the License. */
#include "AverageLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#include "BatchNormBaseLayer.h"
#include "BatchNormalizationLayer.h"
#include "Layer.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
#ifdef PADDLE_WITH_CUDA
#include "CudnnBatchNormLayer.h"
#endif
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#pragma once
#include "Layer.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -12,7 +12,7 @@ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
#ifdef PADDLE_WITH_CUDA
#include "hl_batch_transpose.h"
#endif
......
......@@ -13,8 +13,8 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "BilinearInterpLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#include "BlockExpandLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
namespace paddle {
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#include "Layer.h"
#include "Projection.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -13,7 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "ContextProjection.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -13,8 +13,8 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "Conv3DLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#include "ConvBaseLayer.h"
#include "paddle/legacy/math/MathUtils.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
namespace paddle {
bool ConvBaseLayer::init(const LayerMap& layerMap,
......
......@@ -13,7 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "ConvBaseProjection.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -13,7 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "ConvProjection.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -14,8 +14,8 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -13,7 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "ConvTransProjection.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -14,8 +14,8 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -13,8 +13,8 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "CosSimLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/legacy/utils/ThreadLocal.h"
namespace paddle {
/**
......
......@@ -14,8 +14,8 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
/**
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include <algorithm>
#include <cmath>
#include <memory>
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/math/SparseMatrix.h"
......
......@@ -13,7 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "CropLayer.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
REGISTER_LAYER(crop, CropLayer);
......
......@@ -15,7 +15,7 @@ limitations under the License. */
#include "CudnnBatchNormLayer.h"
#include "Layer.h"
#include "paddle/legacy/cuda/include/hl_batch_norm.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -17,7 +17,7 @@ limitations under the License. */
#include <cudnn.h>
#include "BatchNormBaseLayer.h"
#include "Layer.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -13,8 +13,8 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "CudnnConvBaseLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
REGISTER_LAYER(cudnn_conv, CudnnConvBaseLayer);
......
......@@ -14,8 +14,8 @@ limitations under the License. */
#include "CudnnPoolLayer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -13,8 +13,8 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "DataNormLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/legacy/utils/ThreadLocal.h"
namespace paddle {
......
......@@ -13,8 +13,8 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "DeConv3DLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -14,8 +14,8 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -13,7 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "Layer.h"
#include "paddle/utils/Logging.h"
#include "paddle/legacy/utils/Logging.h"
namespace paddle {
/**
......
......@@ -13,8 +13,8 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "ExpandConvLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
DEFINE_bool(use_nnpack,
false,
......
......@@ -13,8 +13,8 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "ExpandLayer.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -16,8 +16,8 @@ limitations under the License. */
#include <algorithm>
#include <vector>
#include "paddle/legacy/math/SparseMatrix.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/legacy/utils/ThreadLocal.h"
namespace paddle {
/**
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -13,7 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
#pragma once
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Stat.h"
#include "Projection.h"
......
......@@ -16,8 +16,8 @@ limitations under the License. */
#include <algorithm>
#include <vector>
#include "paddle/legacy/math/SparseMatrix.h"
#include "paddle/utils/Logging.h"
#include "paddle/utils/Stat.h"
#include "paddle/legacy/utils/Logging.h"
#include "paddle/legacy/utils/Stat.h"
namespace paddle {
......
......@@ -16,7 +16,7 @@ limitations under the License. */
#include "Layer.h"
#include "paddle/legacy/math/Matrix.h"
#include "paddle/utils/ThreadLocal.h"
#include "paddle/legacy/utils/ThreadLocal.h"
namespace paddle {
/**
......
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
此差异已折叠。
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册