Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
2d6bda5c
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
2d6bda5c
编写于
6月 08, 2021
作者:
P
phlrain
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
add memcpy; test=develop
上级
6ee54b49
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
64 addition
and
24 deletion
+64
-24
paddle/fluid/framework/new_exec.h
paddle/fluid/framework/new_exec.h
+64
-24
未找到文件。
paddle/fluid/framework/new_exec.h
浏览文件 @
2d6bda5c
...
...
@@ -17,12 +17,13 @@
#include "paddle/fluid/framework/variable.h"
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/framework/operator.h"
#include "paddle/fluid/framework/variable_helper.h"
#include "paddle/fluid/platform/init.h"
#include <chrono>
#include <gperftools/profiler.h>
//USE_OP(fill_constant);
//USE_OP(elementwise_add);
...
...
@@ -524,7 +525,8 @@ void build_variable_scope( const framework::ProgramDesc& pdesc, VariableScope* v
}
auto
v
=
new
Variable
();
v
->
GetMutable
<
LoDTensor
>
();
//v->GetMutable<LoDTensor>();
InitializeVariable
(
v
,
var
->
GetType
());
var_scope
->
var_list
.
push_back
(
std
::
unique_ptr
<
Variable
>
(
v
));
}
}
...
...
@@ -537,9 +539,9 @@ void build_op_func_list( const framework::ProgramDesc& pdesc, std::vector<Operat
for
(
auto
&
op
:
global_block
.
AllOps
()
)
{
//
cerr << op->Type() << endl;
cerr
<<
op
->
Type
()
<<
endl
;
//bool debug = op->Type() == "softmax_with_cross_entropy_grad";
bool
debug
=
fals
e
;
bool
debug
=
tru
e
;
//cerr << "create op" << endl;
//auto op_base_u = OpRegistry::CreateOp(*op);
...
...
@@ -634,16 +636,25 @@ void build_op_func_list( const framework::ProgramDesc& pdesc, std::vector<Operat
//cerr << "22" << endl;
// add transfer log
for
(
auto
&
var_name_item
:
ins_map
)
//cerr << "in map size " << ins_map.size() << endl;
VariableValueMap
&
ins_map_temp
=
runtime_context
.
inputs
;
cerr
<<
"ins map siz"
<<
ins_map_temp
.
size
()
<<
endl
;
for
(
auto
&
var_name_item
:
ins_map_temp
)
{
auto
&
vec_ids
=
ins_name2id
[
var_name_item
.
first
];
cerr
<<
"in name "
<<
var_name_item
.
first
<<
endl
;
//auto& vec_ids = ins_name2id[ var_name_item.first ];
for
(
size_t
i
=
0
;
i
<
var_name_item
.
second
.
size
();
++
i
)
{
auto
var
=
var_name_item
.
second
[
i
];
auto
tensor_in
=
static_cast
<
const
Tensor
*>
(
&
(
var
->
Get
<
LoDTensor
>
()));
cerr
<<
"i "
<<
i
<<
"
\t
"
<<
tensor_in
->
IsInitialized
()
<<
endl
;
auto
kernel_type_for_var
=
static_cast
<
const
framework
::
OperatorWithKernel
*>
(
op_base
)
->
GetKernelTypeForVar
(
var_name_item
.
first
,
*
tensor_in
,
expected_kernel_key
);
if
(
debug
)
{
cerr
<<
"var name "
<<
var_name_item
.
first
<<
endl
;
cerr
<<
expected_kernel_key
.
place_
<<
"
\t
"
<<
kernel_type_for_var
.
place_
<<
endl
;
}
if
(
!
platform
::
is_same_place
(
kernel_type_for_var
.
place_
,
expected_kernel_key
.
place_
)
)
{
...
...
@@ -658,7 +669,8 @@ void build_op_func_list( const framework::ProgramDesc& pdesc, std::vector<Operat
var_scope
->
var_list
.
push_back
(
std
::
unique_ptr
<
Variable
>
(
v
));
VariableNameMap
copy_in_map
;
copy_in_map
[
"X"
]
=
input_names
[
var_name_item
.
first
];
cerr
<<
"ints name is "
<<
input_names
[
var_name_item
.
first
][
i
]
<<
endl
;
copy_in_map
[
"X"
]
=
{
input_names
[
var_name_item
.
first
][
i
]
};
VariableNameMap
copy_out_map
;
copy_out_map
[
"Out"
]
=
{
new_var_name
};
AttributeMap
attr_map
;
...
...
@@ -669,25 +681,32 @@ void build_op_func_list( const framework::ProgramDesc& pdesc, std::vector<Operat
std
::
map
<
std
::
string
,
std
::
vector
<
int
>
>
copy_out_name2id
;
copy_out_name2id
[
"Out"
]
=
{
var_scope
->
name2id
[
new_var_name
]};
vec_ids
[
i
]
=
var_scope
->
name2id
[
new_var_name
];
//vec_ids[i] = var_scope->name2id[new_var_name];
// update out runtime_context
op_func_node
.
input_index
[
var_name_item
.
first
][
i
]
=
var_scope
->
name2id
[
new_var_name
];
VariableValueMap
copy_ins_value_map
;
copy_ins_value_map
[
"X"
]
=
ins_map
[
var_name_item
.
first
]
;
VariableValueMap
copy_ins_value_map
;
copy_ins_value_map
[
"X"
]
=
{
var
}
;
VariableValueMap
copy_outs_value_map
;
copy_outs_value_map
[
"Out"
]
=
{
v
};
auto
copy_op
=
info
.
Creator
()(
"memcpy"
,
copy_in_map
,
copy_out_map
,
attr_map
);
auto
&
copy_info
=
OpInfoMap
::
Instance
().
Get
(
"memcpy"
);
auto
copy_op
=
copy_info
.
Creator
()(
"memcpy"
,
copy_in_map
,
copy_out_map
,
attr_map
);
if
(
debug
)
cerr
<<
"create memcpy"
<<
endl
;
OpFuncNode
copy_op_func_node
;
copy_op_func_node
.
input_index
=
copy_ins_name2id
;
copy_op_func_node
.
output_index
=
copy_out_name2id
;
RuntimeContext
runtime_context
(
{},
{});
runtime_context
.
inputs
.
swap
(
copy_ins_value_map
);
runtime_context
.
outputs
.
swap
(
copy_outs_value_map
);
RuntimeContext
copy_
runtime_context
(
{},
{});
copy_
runtime_context
.
inputs
.
swap
(
copy_ins_value_map
);
copy_
runtime_context
.
outputs
.
swap
(
copy_outs_value_map
);
//cerr << "create runtime context" << endl;
RuntimeInferShapeContext
infer_shape_ctx
(
*
copy_op
,
runtime_context
);
static_cast
<
const
framework
::
OperatorWithKernel
*>
(
copy_op
)
->
InferShape
(
&
infer_shape_ctx
);
RuntimeInferShapeContext
copy_infer_shape_ctx
(
*
copy_op
,
copy_runtime_context
);
if
(
debug
)
cerr
<<
"before infer shape"
<<
endl
;
static_cast
<
const
framework
::
OperatorWithKernel
*>
(
copy_op
)
->
InferShape
(
&
copy_infer_shape_ctx
);
if
(
debug
)
cerr
<<
"infer shape"
<<
endl
;
//cerr << "fin infer shape" << endl;
auto
&
all_op_kernels
=
OperatorWithKernel
::
AllOpKernels
();
auto
kernels_iter
=
all_op_kernels
.
find
(
"memcpy"
);
...
...
@@ -704,20 +723,25 @@ void build_op_func_list( const framework::ProgramDesc& pdesc, std::vector<Operat
OpKernelMap
&
kernels
=
kernels_iter
->
second
;
//auto place = platform::CPUPlace();
//auto place = platform::CUDAPlace(0);
platform
::
DeviceContextPool
&
pool
=
platform
::
DeviceContextPool
::
Instance
();
auto
*
dev_ctx
=
pool
.
Get
(
place
);
Scope
scope
;
auto
exec_ctx
=
ExecutionContext
(
*
copy_op
,
scope
,
*
dev_ctx
,
runtime_context
);
auto
copy_exec_ctx
=
ExecutionContext
(
*
copy_op
,
scope
,
*
dev_ctx
,
copy_
runtime_context
);
if
(
debug
)
cerr
<<
"21"
<<
endl
;
auto
expected_kernel_key
=
dynamic_cast
<
const
framework
::
OperatorWithKernel
*>
(
copy_op
)
->
GetExpectedKernelType
(
exec_ctx
);
auto
expected_kernel_key
=
dynamic_cast
<
const
framework
::
OperatorWithKernel
*>
(
copy_op
)
->
GetExpectedKernelType
(
copy_
exec_ctx
);
if
(
debug
)
cerr
<<
"22"
<<
endl
;
//cerr << "22" << endl;
auto
kernel_iter
=
kernels
.
find
(
expected_kernel_key
);
copy_op_func_node
.
kernel_func_
=
OpKernelFunc
(
kernel_iter
->
second
);
copy_op_func_node
.
kernel_func_
(
exec_ctx
);
copy_op_func_node
.
kernel_func_
(
copy_exec_ctx
);
if
(
debug
)
cerr
<<
"run exe ctx"
<<
endl
;
op_list
.
push_back
(
copy_op
);
vec_func_list
.
push_back
(
copy_op_func_node
);
var_name_item
.
second
[
i
]
=
v
;
}
}
}
...
...
@@ -833,8 +857,10 @@ public:
paddle
::
framework
::
build_op_func_list
(
prog_
,
op_list
,
vec_func_list
,
&
global_scope
,
place_
);
is_build
=
true
;
}
paddle
::
framework
::
exec_op_func_list
(
vec_func_list
,
op_list
,
global_scope
,
place_
);
else
{
paddle
::
framework
::
exec_op_func_list
(
vec_func_list
,
op_list
,
global_scope
,
place_
);
}
for
(
size_t
i
=
0
;
i
<
vec_fetch_name
.
size
();
++
i
)
{
...
...
@@ -845,8 +871,22 @@ public:
//cerr << "out " << fetch_tensor->data<float>()[0] << endl;
if
(
platform
::
is_gpu_place
(
fetch_tensor
->
place
()
)
)
{
cerr
<<
"fetch gpu"
<<
endl
;
Tensor
out
;
platform
::
DeviceContextPool
&
pool
=
platform
::
DeviceContextPool
::
Instance
();
auto
*
dev_ctx
=
pool
.
Get
(
place_
);
dev_ctx
->
Wait
();
TensorCopySync
(
*
fetch_tensor
,
platform
::
CPUPlace
(),
&
out
);
dev_ctx
->
Wait
();
cerr
<<
"out "
<<
out
<<
endl
;
}
else
{
cerr
<<
"out "
<<
*
fetch_tensor
<<
endl
;
cerr
<<
"out "
<<
*
fetch_tensor
<<
endl
;
}
}
}
private:
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录