Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
3301d440
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
3301d440
编写于
4月 16, 2018
作者:
C
chengduoZH
浏览文件
操作
浏览文件
下载
差异文件
Merge branch 'develop' of
https://github.com/PaddlePaddle/Paddle
into feature/add_reduce_op_handle
上级
3c5bbf42
18665979
变更
22
显示空白变更内容
内联
并排
Showing
22 changed file
with
407 addition
and
15 deletion
+407
-15
CMakeLists.txt
CMakeLists.txt
+6
-0
Dockerfile
Dockerfile
+8
-1
cmake/external/grpc.cmake
cmake/external/grpc.cmake
+1
-1
doc/fluid/api/layers.rst
doc/fluid/api/layers.rst
+6
-0
doc/fluid/dev/index_cn.rst
doc/fluid/dev/index_cn.rst
+1
-0
doc/fluid/dev/index_en.rst
doc/fluid/dev/index_en.rst
+1
-0
paddle/fluid/framework/details/multi_devices_graph_builder.cc
...le/fluid/framework/details/multi_devices_graph_builder.cc
+7
-7
paddle/fluid/framework/details/multi_devices_graph_builder.h
paddle/fluid/framework/details/multi_devices_graph_builder.h
+2
-2
paddle/fluid/framework/program_desc_test.cc
paddle/fluid/framework/program_desc_test.cc
+2
-2
paddle/fluid/inference/CMakeLists.txt
paddle/fluid/inference/CMakeLists.txt
+3
-0
paddle/fluid/inference/tensorrt/CMakeLists.txt
paddle/fluid/inference/tensorrt/CMakeLists.txt
+1
-0
paddle/fluid/inference/tensorrt/test_tensorrt.cc
paddle/fluid/inference/tensorrt/test_tensorrt.cc
+155
-0
paddle/fluid/platform/dynload/CMakeLists.txt
paddle/fluid/platform/dynload/CMakeLists.txt
+5
-0
paddle/fluid/platform/dynload/dynamic_loader.cc
paddle/fluid/platform/dynload/dynamic_loader.cc
+12
-0
paddle/fluid/platform/dynload/dynamic_loader.h
paddle/fluid/platform/dynload/dynamic_loader.h
+1
-0
paddle/fluid/platform/dynload/tensorrt.cc
paddle/fluid/platform/dynload/tensorrt.cc
+30
-0
paddle/fluid/platform/dynload/tensorrt.h
paddle/fluid/platform/dynload/tensorrt.h
+69
-0
paddle/utils/DynamicLoader.cpp
paddle/utils/DynamicLoader.cpp
+11
-0
paddle/utils/DynamicLoader.h
paddle/utils/DynamicLoader.h
+8
-0
python/paddle/fluid/layers/nn.py
python/paddle/fluid/layers/nn.py
+66
-0
python/paddle/fluid/metrics.py
python/paddle/fluid/metrics.py
+2
-2
python/paddle/fluid/tests/unittests/test_layers.py
python/paddle/fluid/tests/unittests/test_layers.py
+10
-0
未找到文件。
CMakeLists.txt
浏览文件 @
3301d440
...
...
@@ -39,6 +39,7 @@ option(WITH_GPU "Compile PaddlePaddle with NVIDIA GPU" ${CUDA_F
option
(
WITH_AMD_GPU
"Compile PaddlePaddle with AMD GPU"
OFF
)
option
(
WITH_AVX
"Compile PaddlePaddle with AVX intrinsics"
${
AVX_FOUND
}
)
option
(
WITH_MKL
"Compile PaddlePaddle with MKL support."
${
AVX_FOUND
}
)
option
(
WITH_TENSORRT
"Compile PaddlePaddle with TensorRT support."
OFF
)
option
(
WITH_DSO
"Compile PaddlePaddle with dynamic linked CUDA"
ON
)
option
(
WITH_TESTING
"Compile PaddlePaddle with unit testing"
OFF
)
option
(
WITH_SWIG_PY
"Compile PaddlePaddle with inference api"
ON
)
...
...
@@ -181,6 +182,11 @@ if(WITH_GPU)
include
(
cuda
)
endif
(
WITH_GPU
)
# TensorRT depends on GPU.
if
(
NOT WITH_GPU
)
set
(
WITH_TENSORRT OFF
)
endif
()
if
(
WITH_AMD_GPU
)
find_package
(
HIP
)
include
(
hip
)
...
...
Dockerfile
浏览文件 @
3301d440
...
...
@@ -45,6 +45,13 @@ ENV PATH=${PATH}:${GOROOT}/bin:${GOPATH}/bin
# install glide
RUN
curl
-s
-q
https://glide.sh/get | sh
# Install TensorRT
# The unnecessary files has been removed to make the library small.
RUN
wget
-qO-
http://paddlepaddledeps.bj.bcebos.com/TensorRT-4.0.0.3.Ubuntu-16.04.4.x86_64-gnu.cuda-8.0.cudnn7.0.tar.gz |
\
tar
-xz
-C
/usr/local
&&
\
cp
-rf
/usr/local/TensorRT/include /usr
&&
\
cp
-rf
/usr/local/TensorRT/lib /usr
# git credential to skip password typing
RUN
git config
--global
credential.helper store
...
...
@@ -57,7 +64,7 @@ RUN localedef -i en_US -f UTF-8 en_US.UTF-8
# specify sphinx version as 1.5.6 and remove -U option for [pip install -U
# sphinx-rtd-theme] since -U option will cause sphinx being updated to newest
# version(1.7.1 for now), which causes building documentation failed.
RUN
pip
install
--upgrade
pip
&&
\
RUN
pip
install
--upgrade
pip
==
9.0.3
&&
\
pip
install
-U
wheel
&&
\
pip
install
-U
docopt PyYAML
sphinx
==
1.5.6
&&
\
pip
install
sphinx-rtd-theme
==
0.1.9 recommonmark
...
...
cmake/external/grpc.cmake
浏览文件 @
3301d440
...
...
@@ -33,7 +33,7 @@ ExternalProject_Add(
extern_grpc
DEPENDS protobuf zlib
GIT_REPOSITORY
"https://github.com/grpc/grpc.git"
GIT_TAG
"v1.1
1
.x"
GIT_TAG
"v1.1
0
.x"
PREFIX
${
GRPC_SOURCES_DIR
}
UPDATE_COMMAND
""
CONFIGURE_COMMAND
""
...
...
doc/fluid/api/layers.rst
浏览文件 @
3301d440
...
...
@@ -473,6 +473,12 @@ multiplex
.. autofunction:: paddle.fluid.layers.multiplex
:noindex:
label_smooth
------------
.. autofunction:: paddle.fluid.layers.label_smooth
:noindex:
ops
===
...
...
doc/fluid/dev/index_cn.rst
浏览文件 @
3301d440
...
...
@@ -4,6 +4,7 @@
.. toctree::
:maxdepth: 1
api_doc_std_cn.md
new_op_cn.md
new_op_kernel.md
use_eigen_cn.md
...
...
doc/fluid/dev/index_en.rst
浏览文件 @
3301d440
...
...
@@ -4,6 +4,7 @@ Development
.. toctree::
:maxdepth: 1
api_doc_std_en.md
new_op_en.md
new_op_kernel.md
use_eigen_en.md
...
...
paddle/fluid/framework/details/multi_devices_graph_builder.cc
浏览文件 @
3301d440
...
...
@@ -55,21 +55,21 @@ MultiDevSSAGraphBuilder::MultiDevSSAGraphBuilder(
}
}
void
MultiDevSSAGraphBuilder
::
CreateOpHandleIOs
(
SSAGraph
*
result
,
OpDesc
*
op
,
void
MultiDevSSAGraphBuilder
::
CreateOpHandleIOs
(
SSAGraph
*
result
,
const
OpDesc
&
op
,
const
platform
::
Place
&
p
,
const
size_t
&
i
)
const
{
auto
*
op_handle
=
result
->
ops_
.
back
().
get
();
op_handle
->
dev_ctxes_
[
p
]
=
const_cast
<
platform
::
DeviceContext
*>
(
platform
::
DeviceContextPool
::
Instance
().
Get
(
p
));
op_handle
->
dev_ctxes_
[
p
]
=
platform
::
DeviceContextPool
::
Instance
().
Get
(
p
);
auto
var_names
=
op
->
InputArgumentNames
();
auto
var_names
=
op
.
InputArgumentNames
();
for
(
auto
&
each_var_name
:
var_names
)
{
VarHandle
*
var
=
CreateOrGetLatestVarHandle
(
result
,
each_var_name
,
p
,
i
);
op_handle
->
AddInput
(
var
);
}
var_names
=
op
->
OutputArgumentNames
();
var_names
=
op
.
OutputArgumentNames
();
for
(
auto
&
each_var_name
:
var_names
)
{
CreateOpOutput
(
result
,
op_handle
,
each_var_name
,
p
,
i
);
...
...
@@ -107,7 +107,7 @@ std::unique_ptr<SSAGraph> MultiDevSSAGraphBuilder::Build(
result
.
ops_
.
emplace_back
(
new
SendOpHandle
(
*
op
,
s
,
p
));
// Create inputs for output on original place and no ssa output
// is created for send op.
CreateOpHandleIOs
(
&
result
,
op
,
p
,
0
);
CreateOpHandleIOs
(
&
result
,
*
op
,
p
,
0
);
continue
;
}
...
...
@@ -117,7 +117,7 @@ std::unique_ptr<SSAGraph> MultiDevSSAGraphBuilder::Build(
result
.
ops_
.
emplace_back
(
new
ComputationOpHandle
(
*
op
,
s
,
p
));
auto
*
op_handle
=
result
.
ops_
.
back
().
get
();
CreateOpHandleIOs
(
&
result
,
op
,
p
,
i
);
CreateOpHandleIOs
(
&
result
,
*
op
,
p
,
i
);
auto
var_names
=
op
->
OutputArgumentNames
();
...
...
paddle/fluid/framework/details/multi_devices_graph_builder.h
浏览文件 @
3301d440
...
...
@@ -45,8 +45,8 @@ class MultiDevSSAGraphBuilder : public SSAGraphBuilder {
std
::
unique_ptr
<
SSAGraph
>
Build
(
const
ProgramDesc
&
program
)
const
override
;
private:
void
CreateOpHandleIOs
(
SSAGraph
*
result
,
OpDesc
*
op
,
const
platform
::
Place
&
p
,
const
size_t
&
i
)
const
;
void
CreateOpHandleIOs
(
SSAGraph
*
result
,
const
OpDesc
&
o
p
,
const
platform
::
Place
&
p
,
const
size_t
&
i
)
const
;
private:
std
::
string
loss_var_name_
;
...
...
paddle/fluid/framework/program_desc_test.cc
浏览文件 @
3301d440
...
...
@@ -66,7 +66,7 @@ TEST(ProgramDesc, copy_ctor) {
for
(
size_t
i
=
0
;
i
<
global_block
->
OpSize
();
++
i
)
{
auto
op_origin
=
global_block
->
Op
(
i
);
auto
op_copy
=
global_block
->
Op
(
i
);
auto
op_copy
=
global_block
_copy
->
Op
(
i
);
ASSERT_EQ
(
op_origin
->
Type
(),
op_copy
->
Type
());
ASSERT_EQ
(
op_origin
->
Inputs
(),
op_copy
->
Inputs
());
...
...
@@ -131,7 +131,7 @@ TEST(ProgramDescBind, serialize_and_deserialize) {
for
(
size_t
i
=
0
;
i
<
global_block
->
OpSize
();
++
i
)
{
auto
op_origin
=
global_block
->
Op
(
i
);
auto
op_restored
=
global_block
->
Op
(
i
);
auto
op_restored
=
global_block
_restored
->
Op
(
i
);
ASSERT_EQ
(
op_origin
->
Type
(),
op_restored
->
Type
());
ASSERT_EQ
(
op_origin
->
Inputs
(),
op_restored
->
Inputs
());
...
...
paddle/fluid/inference/CMakeLists.txt
浏览文件 @
3301d440
...
...
@@ -21,4 +21,7 @@ endif()
if
(
WITH_TESTING
)
add_subdirectory
(
tests/book
)
if
(
WITH_TENSORRT
)
add_subdirectory
(
tensorrt
)
endif
()
endif
()
paddle/fluid/inference/tensorrt/CMakeLists.txt
0 → 100644
浏览文件 @
3301d440
nv_test
(
test_tensorrt SRCS test_tensorrt.cc DEPS dynload_cuda device_context dynamic_loader
)
paddle/fluid/inference/tensorrt/test_tensorrt.cc
0 → 100644
浏览文件 @
3301d440
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#include <glog/logging.h>
#include <gtest/gtest.h>
#include "NvInfer.h"
#include "cuda.h"
#include "cuda_runtime_api.h"
#include "paddle/fluid/platform/dynload/tensorrt.h"
namespace
dy
=
paddle
::
platform
::
dynload
;
class
Logger
:
public
nvinfer1
::
ILogger
{
public:
void
log
(
nvinfer1
::
ILogger
::
Severity
severity
,
const
char
*
msg
)
override
{
switch
(
severity
)
{
case
Severity
::
kINFO
:
LOG
(
INFO
)
<<
msg
;
break
;
case
Severity
::
kWARNING
:
LOG
(
WARNING
)
<<
msg
;
break
;
case
Severity
::
kINTERNAL_ERROR
:
case
Severity
::
kERROR
:
LOG
(
ERROR
)
<<
msg
;
break
;
default:
break
;
}
}
};
class
ScopedWeights
{
public:
ScopedWeights
(
float
value
)
:
value_
(
value
)
{
w
.
type
=
nvinfer1
::
DataType
::
kFLOAT
;
w
.
values
=
&
value_
;
w
.
count
=
1
;
}
const
nvinfer1
::
Weights
&
get
()
{
return
w
;
}
private:
float
value_
;
nvinfer1
::
Weights
w
;
};
// The following two API are implemented in TensorRT's header file, cannot load
// from the dynamic library. So create our own implementation and directly
// trigger the method from the dynamic library.
nvinfer1
::
IBuilder
*
createInferBuilder
(
nvinfer1
::
ILogger
&
logger
)
{
return
static_cast
<
nvinfer1
::
IBuilder
*>
(
dy
::
createInferBuilder_INTERNAL
(
&
logger
,
NV_TENSORRT_VERSION
));
}
nvinfer1
::
IRuntime
*
createInferRuntime
(
nvinfer1
::
ILogger
&
logger
)
{
return
static_cast
<
nvinfer1
::
IRuntime
*>
(
dy
::
createInferRuntime_INTERNAL
(
&
logger
,
NV_TENSORRT_VERSION
));
}
const
char
*
kInputTensor
=
"input"
;
const
char
*
kOutputTensor
=
"output"
;
// Creates a network to compute y = 2x + 3
nvinfer1
::
IHostMemory
*
CreateNetwork
()
{
Logger
logger
;
// Create the engine.
nvinfer1
::
IBuilder
*
builder
=
createInferBuilder
(
logger
);
ScopedWeights
weights
(
2.
);
ScopedWeights
bias
(
3.
);
nvinfer1
::
INetworkDefinition
*
network
=
builder
->
createNetwork
();
// Add the input
auto
input
=
network
->
addInput
(
kInputTensor
,
nvinfer1
::
DataType
::
kFLOAT
,
nvinfer1
::
DimsCHW
{
1
,
1
,
1
});
EXPECT_NE
(
input
,
nullptr
);
// Add the hidden layer.
auto
layer
=
network
->
addFullyConnected
(
*
input
,
1
,
weights
.
get
(),
bias
.
get
());
EXPECT_NE
(
layer
,
nullptr
);
// Mark the output.
auto
output
=
layer
->
getOutput
(
0
);
output
->
setName
(
kOutputTensor
);
network
->
markOutput
(
*
output
);
// Build the engine.
builder
->
setMaxBatchSize
(
1
);
builder
->
setMaxWorkspaceSize
(
1
<<
10
);
auto
engine
=
builder
->
buildCudaEngine
(
*
network
);
EXPECT_NE
(
engine
,
nullptr
);
// Serialize the engine to create a model, then close.
nvinfer1
::
IHostMemory
*
model
=
engine
->
serialize
();
network
->
destroy
();
engine
->
destroy
();
builder
->
destroy
();
return
model
;
}
void
Execute
(
nvinfer1
::
IExecutionContext
&
context
,
const
float
*
input
,
float
*
output
)
{
const
nvinfer1
::
ICudaEngine
&
engine
=
context
.
getEngine
();
// Two binds, input and output
ASSERT_EQ
(
engine
.
getNbBindings
(),
2
);
const
int
input_index
=
engine
.
getBindingIndex
(
kInputTensor
);
const
int
output_index
=
engine
.
getBindingIndex
(
kOutputTensor
);
// Create GPU buffers and a stream
void
*
buffers
[
2
];
ASSERT_EQ
(
0
,
cudaMalloc
(
&
buffers
[
input_index
],
sizeof
(
float
)));
ASSERT_EQ
(
0
,
cudaMalloc
(
&
buffers
[
output_index
],
sizeof
(
float
)));
cudaStream_t
stream
;
ASSERT_EQ
(
0
,
cudaStreamCreate
(
&
stream
));
// Copy the input to the GPU, execute the network, and copy the output back.
ASSERT_EQ
(
0
,
cudaMemcpyAsync
(
buffers
[
input_index
],
input
,
sizeof
(
float
),
cudaMemcpyHostToDevice
,
stream
));
context
.
enqueue
(
1
,
buffers
,
stream
,
nullptr
);
ASSERT_EQ
(
0
,
cudaMemcpyAsync
(
output
,
buffers
[
output_index
],
sizeof
(
float
),
cudaMemcpyDeviceToHost
,
stream
));
cudaStreamSynchronize
(
stream
);
// Release the stream and the buffers
cudaStreamDestroy
(
stream
);
ASSERT_EQ
(
0
,
cudaFree
(
buffers
[
input_index
]));
ASSERT_EQ
(
0
,
cudaFree
(
buffers
[
output_index
]));
}
TEST
(
TensorrtTest
,
BasicFunction
)
{
// Create the network serialized model.
nvinfer1
::
IHostMemory
*
model
=
CreateNetwork
();
// Use the model to create an engine and an execution context.
Logger
logger
;
nvinfer1
::
IRuntime
*
runtime
=
createInferRuntime
(
logger
);
nvinfer1
::
ICudaEngine
*
engine
=
runtime
->
deserializeCudaEngine
(
model
->
data
(),
model
->
size
(),
nullptr
);
model
->
destroy
();
nvinfer1
::
IExecutionContext
*
context
=
engine
->
createExecutionContext
();
// Execute the network.
float
input
=
1234
;
float
output
;
Execute
(
*
context
,
&
input
,
&
output
);
EXPECT_EQ
(
output
,
input
*
2
+
3
);
// Destroy the engine.
context
->
destroy
();
engine
->
destroy
();
runtime
->
destroy
();
}
paddle/fluid/platform/dynload/CMakeLists.txt
浏览文件 @
3301d440
cc_library
(
dynamic_loader SRCS dynamic_loader.cc DEPS glog gflags enforce
)
list
(
APPEND CUDA_SRCS cublas.cc cudnn.cc curand.cc nccl.cc
)
if
(
WITH_TENSORRT
)
list
(
APPEND CUDA_SRCS tensorrt.cc
)
endif
()
configure_file
(
cupti_lib_path.h.in
${
CMAKE_CURRENT_BINARY_DIR
}
/cupti_lib_path.h
)
if
(
CUPTI_FOUND
)
list
(
APPEND CUDA_SRCS cupti.cc
)
...
...
paddle/fluid/platform/dynload/dynamic_loader.cc
浏览文件 @
3301d440
...
...
@@ -45,6 +45,10 @@ DEFINE_string(nccl_dir, "",
DEFINE_string
(
cupti_dir
,
""
,
"Specify path for loading cupti.so."
);
DEFINE_string
(
tensorrt_dir
,
""
,
"Specify path for loading tensorrt library, such as libnvinfer.so."
);
namespace
paddle
{
namespace
platform
{
namespace
dynload
{
...
...
@@ -194,6 +198,14 @@ void* GetNCCLDsoHandle() {
#endif
}
void
*
GetTensorRtDsoHandle
()
{
#if defined(__APPLE__) || defined(__OSX__)
return
GetDsoHandleFromSearchPath
(
FLAGS_tensorrt_dir
,
"libnvinfer.dylib"
);
#else
return
GetDsoHandleFromSearchPath
(
FLAGS_tensorrt_dir
,
"libnvinfer.so"
);
#endif
}
}
// namespace dynload
}
// namespace platform
}
// namespace paddle
paddle/fluid/platform/dynload/dynamic_loader.h
浏览文件 @
3301d440
...
...
@@ -25,6 +25,7 @@ void* GetCurandDsoHandle();
void
*
GetWarpCTCDsoHandle
();
void
*
GetLapackDsoHandle
();
void
*
GetNCCLDsoHandle
();
void
*
GetTensorRtDsoHandle
();
}
// namespace dynload
}
// namespace platform
...
...
paddle/fluid/platform/dynload/tensorrt.cc
0 → 100644
浏览文件 @
3301d440
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/fluid/platform/dynload/tensorrt.h"
namespace
paddle
{
namespace
platform
{
namespace
dynload
{
std
::
once_flag
tensorrt_dso_flag
;
void
*
tensorrt_dso_handle
;
#define DEFINE_WRAP(__name) DynLoad__##__name __name
TENSORRT_RAND_ROUTINE_EACH
(
DEFINE_WRAP
);
}
// namespace dynload
}
// namespace platform
}
// namespace paddle
paddle/fluid/platform/dynload/tensorrt.h
0 → 100644
浏览文件 @
3301d440
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#pragma once
#include <NvInfer.h>
#include <dlfcn.h>
#include <mutex> // NOLINT
#include "paddle/fluid/platform/dynload/dynamic_loader.h"
#include "paddle/fluid/platform/enforce.h"
namespace
paddle
{
namespace
platform
{
namespace
dynload
{
extern
std
::
once_flag
tensorrt_dso_flag
;
extern
void
*
tensorrt_dso_handle
;
#ifdef PADDLE_USE_DSO
#define DECLARE_DYNAMIC_LOAD_TENSORRT_WRAP(__name) \
struct DynLoad__##__name { \
template <typename... Args> \
auto operator()(Args... args) -> decltype(__name(args...)) { \
using tensorrt_func = decltype(__name(args...)) (*)(Args...); \
std::call_once(tensorrt_dso_flag, []() { \
tensorrt_dso_handle = \
paddle::platform::dynload::GetTensorRtDsoHandle(); \
PADDLE_ENFORCE(tensorrt_dso_handle, "load tensorrt so failed"); \
}); \
void* p_##__name = dlsym(tensorrt_dso_handle, #__name); \
PADDLE_ENFORCE(p_##__name, "load %s failed", #__name); \
return reinterpret_cast<tensorrt_func>(p_##__name)(args...); \
} \
}; \
extern DynLoad__##__name __name
#else
#define DECLARE_DYNAMIC_LOAD_TENSORRT_WRAP(__name) \
struct DynLoad__##__name { \
template <typename... Args> \
tensorrtResult_t operator()(Args... args) { \
return __name(args...); \
} \
}; \
extern DynLoad__##__name __name
#endif
#define TENSORRT_RAND_ROUTINE_EACH(__macro) \
__macro(createInferBuilder_INTERNAL); \
__macro(createInferRuntime_INTERNAL);
TENSORRT_RAND_ROUTINE_EACH
(
DECLARE_DYNAMIC_LOAD_TENSORRT_WRAP
)
}
// namespace dynload
}
// namespace platform
}
// namespace paddle
paddle/utils/DynamicLoader.cpp
浏览文件 @
3301d440
...
...
@@ -32,6 +32,8 @@ DEFINE_string(warpctc_dir, "", "Specify path for loading libwarpctc.so.");
DEFINE_string
(
lapack_dir
,
""
,
"Specify path for loading liblapack.so."
);
DEFINE_string
(
tensorrt_dir
,
""
,
"Specify path for loading libnvinfer.so."
);
static
inline
std
::
string
join
(
const
std
::
string
&
part1
,
const
std
::
string
&
part2
)
{
// directory separator
...
...
@@ -157,3 +159,12 @@ void GetLapackDsoHandle(void** dso_handle) {
GetDsoHandleFromSearchPath
(
FLAGS_lapack_dir
,
"liblapacke.so"
,
dso_handle
);
#endif
}
void
GetTensorRtDsoHandle
(
void
**
dso_handle
)
{
#if defined(__APPLE__) || defined(__OSX__)
GetDsoHandleFromSearchPath
(
FLAGS_tensorrt_dir
,
"libnvinfer.dylib"
,
dso_handle
);
#else
GetDsoHandleFromSearchPath
(
FLAGS_tensorrt_dir
,
"libnvinfer.so"
,
dso_handle
);
#endif
}
paddle/utils/DynamicLoader.h
浏览文件 @
3301d440
...
...
@@ -58,3 +58,11 @@ void GetWarpCTCDsoHandle(void** dso_handle);
*
*/
void
GetLapackDsoHandle
(
void
**
dso_handle
);
/**
* @brief load the DSO of tensorrt
*
* @param **dso_handle dso handler
*
*/
void
GetTensorRtDsoHandle
(
void
**
dso_handle
);
python/paddle/fluid/layers/nn.py
浏览文件 @
3301d440
...
...
@@ -77,6 +77,7 @@ __all__ = [
'lod_reset'
,
'lrn'
,
'pad'
,
'label_smooth'
,
]
...
...
@@ -3678,3 +3679,68 @@ def pad(x, paddings, pad_value=0., name=None):
attrs
=
{
'paddings'
:
paddings
,
'pad_value'
:
float
(
pad_value
)})
return
out
def
label_smooth
(
label
,
prior_dist
=
None
,
epsilon
=
0.1
,
dtype
=
"float32"
,
name
=
None
):
"""
Label smoothing is a mechanism to regularize the classifier layer and is
called label-smoothing regularization (LSR).
Label smoothing is proposed to encourage the model to be less confident,
since optimizing the log-likelihood of the correct label directly may
cause overfitting and reduce the ability of the model to adapt. Label
smoothing replaces the ground-truth label :math:`y` with the weighted sum
of itself and some fixed distribution :math:`\mu`. For class :math:`k`,
i.e.
.. math::
\\
tilde{y_k} = (1 - \epsilon) * y_k + \epsilon * \mu_k,
where :math:`1 - \epsilon` and :math:`\epsilon` are the weights
respectively, and :math:`
\\
tilde{y}_k` is the smoothed label. Usually
uniform distribution is used for :math:`\mu`.
See more details about label smoothing in https://arxiv.org/abs/1512.00567.
Args:
label(Variable): The input variable containing the label data. The
label data should use one-hot representation.
prior_dist(Variable): The prior distribution to be used to smooth
labels. If not provided, an uniform distribution
is used. The shape of :attr:`prior_dist` should
be :math:`(1, class\_num)`.
epsilon(float): The weight used to mix up the original ground-truth
distribution and the fixed distribution.
dtype(np.dtype|core.VarDesc.VarType|str): The type of data : float32,
float_64, int etc.
name(str|None): A name for this layer(optional). If set None, the layer
will be named automatically.
Returns:
Variable: The tensor variable containing the smoothed labels.
Examples:
.. code-block:: python
label = layers.data(name="label", shape=[1], dtype="float32")
one_hot_label = layers.one_hot(input=label, depth=10)
smooth_label = layers.label_smooth(
label=one_hot_label, epsilon=0.1, dtype="float32")
"""
if
epsilon
>
1.
or
epsilon
<
0.
:
raise
ValueError
(
"The value of epsilon must be between 0 and 1."
)
helper
=
LayerHelper
(
"label_smooth"
,
**
locals
())
label
.
stop_gradient
=
True
smooth_label
=
helper
.
create_tmp_variable
(
dtype
)
helper
.
append_op
(
type
=
"label_smooth"
,
inputs
=
{
"X"
:
label
,
"PriorDist"
:
prior_dist
}
if
prior_dist
else
{
"X"
:
label
},
outputs
=
{
"Out"
:
smooth_label
},
attrs
=
{
"epsilon"
:
float
(
epsilon
)})
return
smooth_label
python/paddle/fluid/metrics.py
浏览文件 @
3301d440
...
...
@@ -169,7 +169,7 @@ class Accuracy(MetricBase):
return
self
.
value
/
self
.
weight
class
ChunkEvalutor
(
MetricBase
):
class
ChunkEvalu
a
tor
(
MetricBase
):
"""
Accumulate counter numbers output by chunk_eval from mini-batches and
compute the precision recall and F1-score using the accumulated counter
...
...
@@ -177,7 +177,7 @@ class ChunkEvalutor(MetricBase):
"""
def
__init__
(
self
,
name
=
None
):
super
(
ChunkEvalutor
,
self
).
__init__
(
name
)
super
(
ChunkEvalu
a
tor
,
self
).
__init__
(
name
)
self
.
num_infer_chunks
=
0
self
.
num_label_chunks
=
0
self
.
num_correct_chunks
=
0
...
...
python/paddle/fluid/tests/unittests/test_layers.py
浏览文件 @
3301d440
...
...
@@ -340,6 +340,16 @@ class TestBook(unittest.TestCase):
print
(
layers
.
lod_reset
(
x
=
x
,
y
=
y
))
print
(
str
(
program
))
def
test_label_smooth
(
self
):
program
=
Program
()
with
program_guard
(
program
):
label
=
layers
.
data
(
name
=
"label"
,
shape
=
[
1
],
dtype
=
"float32"
)
one_hot_label
=
layers
.
one_hot
(
input
=
label
,
depth
=
10
)
smooth_label
=
layers
.
label_smooth
(
label
=
one_hot_label
,
epsilon
=
0.1
,
dtype
=
"float32"
)
self
.
assertIsNotNone
(
smooth_label
)
print
(
str
(
program
))
if
__name__
==
'__main__'
:
unittest
.
main
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录