Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
dfc8d3c1
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
dfc8d3c1
编写于
9月 26, 2017
作者:
C
chengduoZH
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Fix (According to the review)
上级
f6e69d74
变更
9
显示空白变更内容
内联
并排
Showing
9 changed file
with
216 addition
and
545 deletion
+216
-545
paddle/operators/math/CMakeLists.txt
paddle/operators/math/CMakeLists.txt
+0
-2
paddle/operators/math/pool_test_maxPool2d.cc
paddle/operators/math/pool_test_maxPool2d.cc
+0
-154
paddle/operators/math/pool_test_maxPool3d.cc
paddle/operators/math/pool_test_maxPool3d.cc
+0
-157
paddle/operators/math/pooling.cc
paddle/operators/math/pooling.cc
+60
-69
paddle/operators/math/pooling.cu
paddle/operators/math/pooling.cu
+68
-76
paddle/operators/math/pooling.h
paddle/operators/math/pooling.h
+16
-19
paddle/operators/pool_op.cc
paddle/operators/pool_op.cc
+24
-22
paddle/operators/pool_op.cu
paddle/operators/pool_op.cu
+11
-10
paddle/operators/pool_op.h
paddle/operators/pool_op.h
+37
-36
未找到文件。
paddle/operators/math/CMakeLists.txt
浏览文件 @
dfc8d3c1
...
@@ -7,5 +7,3 @@ endif()
...
@@ -7,5 +7,3 @@ endif()
nv_test
(
math_function_test SRCS math_function_test.cc DEPS math_function tensor
)
nv_test
(
math_function_test SRCS math_function_test.cc DEPS math_function tensor
)
cc_test
(
im2col_test SRCS im2col_test.cc DEPS math_function tensor
)
cc_test
(
im2col_test SRCS im2col_test.cc DEPS math_function tensor
)
cc_test
(
pool_test_maxPool2d_test SRCS pool_test_maxPool2d.cc DEPS math_function tensor
)
cc_test
(
pool_test_maxPool3d_test SRCS pool_test_maxPool3d.cc DEPS math_function tensor
)
paddle/operators/math/pool_test_maxPool2d.cc
已删除
100644 → 0
浏览文件 @
f6e69d74
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License. */
#include <gtest/gtest.h>
#include "paddle/operators/math/pooling.h"
#include "paddle/memory/memcpy.h"
#include "paddle/platform/enforce.h"
#include <stdlib.h>
#include <time.h>
#ifndef PADDLE_ONLY_CPU
template
<
typename
PoolType
,
typename
PoolGradType
>
void
testPool2d
(
paddle
::
platform
::
DeviceContext
&
context
,
PoolType
pool_process
,
PoolGradType
poolGrad_process
,
paddle
::
framework
::
Tensor
&
input
,
paddle
::
framework
::
Tensor
&
input_grad
,
paddle
::
framework
::
Tensor
&
output
,
paddle
::
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
)
{
paddle
::
operators
::
math
::
Pool2dForwardFunctor
<
paddle
::
platform
::
GPUPlace
,
PoolType
,
float
>
pool2d_forward
;
pool2d_forward
(
context
,
input
,
output
,
ksize
,
strides
,
paddings
,
pool_process
);
int
times
=
50
;
clock_t
start
,
finish
;
double
totaltime
;
// Pool2dBackwardFunctor
start
=
clock
();
for
(
int
i
=
0
;
i
<
times
;
++
i
)
{
paddle
::
operators
::
math
::
Pool2dBackwardFunctor
<
paddle
::
platform
::
GPUPlace
,
PoolGradType
,
float
>
pool2d_backward
;
pool2d_backward
(
context
,
input
,
input_grad
,
output
,
output_grad
,
ksize
,
strides
,
paddings
,
poolGrad_process
);
PADDLE_ENFORCE
(
cudaStreamSynchronize
(
0
),
"cudaStreamSynchronize failed in pool2d_backward CopyFrom"
);
}
finish
=
clock
();
totaltime
=
(
double
)(
finish
-
start
)
/
CLOCKS_PER_SEC
;
totaltime
/=
times
;
std
::
cout
<<
"
\n
Pool3dBackwardFunctor: "
<<
totaltime
<<
"s"
<<
std
::
endl
;
// MaxPool3dBackwardFunctor
start
=
clock
();
for
(
int
j
=
0
;
j
<
times
;
++
j
)
{
paddle
::
operators
::
math
::
MaxPool2dBackwardFunctor
<
paddle
::
platform
::
GPUPlace
,
float
>
maxpool2d_backward
;
maxpool2d_backward
(
context
,
input
,
input_grad
,
output
,
output_grad
,
ksize
,
strides
,
paddings
);
PADDLE_ENFORCE
(
cudaStreamSynchronize
(
0
),
"cudaStreamSynchronize failed in maxpool2d_backward CopyFrom"
);
}
finish
=
clock
();
totaltime
=
(
double
)(
finish
-
start
)
/
CLOCKS_PER_SEC
;
totaltime
/=
times
;
std
::
cout
<<
"
\n
MaxPool3dBackwardFunctor: "
<<
totaltime
<<
"s"
<<
std
::
endl
;
}
void
test2dPool
()
{
using
paddle
::
platform
::
DeviceContext
;
using
paddle
::
platform
::
CUDADeviceContext
;
using
paddle
::
platform
::
GPUPlace
;
paddle
::
framework
::
Tensor
input_tmp
;
paddle
::
framework
::
Tensor
output_tmp
;
paddle
::
framework
::
Tensor
input
;
paddle
::
framework
::
Tensor
input_grad
;
paddle
::
framework
::
Tensor
output
;
paddle
::
framework
::
Tensor
output_grad
;
int
batch
=
32
;
int
channel
=
32
;
int
input_height
=
128
;
int
input_width
=
128
;
int
in_len
=
batch
*
channel
*
input_height
*
input_width
;
std
::
vector
<
int
>
ksize
({
3
,
3
});
std
::
vector
<
int
>
strides
({
1
,
1
});
std
::
vector
<
int
>
paddings
({
0
,
0
});
int
output_height
=
(
input_height
-
ksize
[
0
]
+
2
*
paddings
[
0
])
/
strides
[
0
]
+
1
;
int
output_width
=
(
input_width
-
ksize
[
1
]
+
2
*
paddings
[
1
])
/
strides
[
1
]
+
1
;
int
output_len
=
output_height
*
output_width
;
input_tmp
.
mutable_data
<
float
>
({
batch
,
channel
,
input_height
,
input_width
},
paddle
::
platform
::
CPUPlace
());
output_tmp
.
mutable_data
<
float
>
({
batch
,
channel
,
output_height
,
output_width
},
paddle
::
platform
::
CPUPlace
());
float
*
arr
=
new
float
[
in_len
];
auto
*
place
=
new
paddle
::
platform
::
GPUPlace
();
float
*
input_ptr
=
input_tmp
.
data
<
float
>
();
for
(
int
i
=
0
;
i
<
in_len
;
++
i
)
arr
[
i
]
=
i
;
// rand() / double(RAND_MAX/2);
memcpy
(
input_ptr
,
arr
,
in_len
*
sizeof
(
float
));
input
.
CopyFrom
<
float
>
(
input_tmp
,
*
place
);
input_ptr
=
input_tmp
.
data
<
float
>
();
for
(
int
i
=
0
;
i
<
in_len
;
++
i
)
arr
[
i
]
=
0
;
memcpy
(
input_ptr
,
arr
,
in_len
*
sizeof
(
float
));
input_grad
.
CopyFrom
<
float
>
(
input_tmp
,
*
place
);
// output
input_ptr
=
output_tmp
.
data
<
float
>
();
for
(
int
i
=
0
;
i
<
output_len
;
++
i
)
arr
[
i
]
=
0
;
// rand() / double(RAND_MAX/2);
memcpy
(
input_ptr
,
arr
,
output_len
*
sizeof
(
float
));
output
.
CopyFrom
<
float
>
(
input_tmp
,
*
place
);
// output
input_ptr
=
output_tmp
.
data
<
float
>
();
for
(
int
i
=
0
;
i
<
output_len
;
++
i
)
arr
[
i
]
=
1
;
// rand() / double(RAND_MAX/2);
memcpy
(
input_ptr
,
arr
,
output_len
*
sizeof
(
float
));
output_grad
.
CopyFrom
<
float
>
(
input_tmp
,
*
place
);
paddle
::
platform
::
DeviceContext
*
context
=
new
paddle
::
platform
::
CUDADeviceContext
(
paddle
::
platform
::
GPUPlace
());
paddle
::
operators
::
math
::
pool
::
maxPool
<
float
>
pool_process
;
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
float
>
poolGrad_process
;
testPool2d
<
paddle
::
operators
::
math
::
pool
::
maxPool
<
float
>
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
float
>>
(
*
context
,
pool_process
,
poolGrad_process
,
input
,
input_grad
,
output
,
output_grad
,
ksize
,
strides
,
paddings
);
}
int
main
()
{
// testPool3d<paddle::platform::CPUPlace>();
test2dPool
();
// testPool3d<paddle::platform::GPUPlace>();
}
#endif
\ No newline at end of file
paddle/operators/math/pool_test_maxPool3d.cc
已删除
100644 → 0
浏览文件 @
f6e69d74
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License. */
#include <gtest/gtest.h>
#include "paddle/operators/math/pooling.h"
#include "paddle/memory/memcpy.h"
#include "paddle/platform/enforce.h"
#include <stdlib.h>
#include <time.h>
#ifndef PADDLE_ONLY_CPU
template
<
typename
PoolType
,
typename
PoolGradType
>
void
testPool3d
(
paddle
::
platform
::
DeviceContext
&
context
,
PoolType
pool_process
,
PoolGradType
poolGrad_process
,
paddle
::
framework
::
Tensor
&
input
,
paddle
::
framework
::
Tensor
&
input_grad
,
paddle
::
framework
::
Tensor
&
output
,
paddle
::
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
)
{
paddle
::
operators
::
math
::
Pool3dForwardFunctor
<
paddle
::
platform
::
GPUPlace
,
PoolType
,
float
>
pool3d_forward
;
pool3d_forward
(
context
,
input
,
output
,
ksize
,
strides
,
paddings
,
pool_process
);
int
times
=
50
;
clock_t
start
,
finish
;
double
totaltime
;
// Pool3dBackwardFunctor
start
=
clock
();
for
(
int
i
=
0
;
i
<
times
;
++
i
)
{
paddle
::
operators
::
math
::
Pool3dBackwardFunctor
<
paddle
::
platform
::
GPUPlace
,
PoolGradType
,
float
>
pool3d_backward
;
pool3d_backward
(
context
,
input
,
input_grad
,
output
,
output_grad
,
ksize
,
strides
,
paddings
,
poolGrad_process
);
PADDLE_ENFORCE
(
cudaStreamSynchronize
(
0
),
"cudaStreamSynchronize failed in pool3d_backward CopyFrom"
);
}
finish
=
clock
();
totaltime
=
(
double
)(
finish
-
start
)
/
CLOCKS_PER_SEC
;
totaltime
/=
times
;
std
::
cout
<<
"
\n
Pool3dBackwardFunctor: "
<<
totaltime
<<
"s"
<<
std
::
endl
;
// MaxPool3dBackwardFunctor
start
=
clock
();
for
(
int
j
=
0
;
j
<
times
;
++
j
)
{
paddle
::
operators
::
math
::
MaxPool3dBackwardFunctor
<
paddle
::
platform
::
GPUPlace
,
float
>
maxpool3d_backward
;
maxpool3d_backward
(
context
,
input
,
input_grad
,
output
,
output_grad
,
ksize
,
strides
,
paddings
);
PADDLE_ENFORCE
(
cudaStreamSynchronize
(
0
),
"cudaStreamSynchronize failed in maxpool3d_backward CopyFrom"
);
}
finish
=
clock
();
totaltime
=
(
double
)(
finish
-
start
)
/
CLOCKS_PER_SEC
;
totaltime
/=
times
;
std
::
cout
<<
"
\n
MaxPool3dBackwardFunctor: "
<<
totaltime
<<
"s"
<<
std
::
endl
;
}
void
test3dPool
()
{
using
paddle
::
platform
::
DeviceContext
;
using
paddle
::
platform
::
CUDADeviceContext
;
using
paddle
::
platform
::
GPUPlace
;
paddle
::
framework
::
Tensor
input_tmp
;
paddle
::
framework
::
Tensor
output_tmp
;
paddle
::
framework
::
Tensor
input
;
paddle
::
framework
::
Tensor
input_grad
;
paddle
::
framework
::
Tensor
output
;
paddle
::
framework
::
Tensor
output_grad
;
int
batch
=
32
;
int
channel
=
4
;
int
input_depth
=
4
;
int
input_height
=
128
;
int
input_width
=
128
;
int
in_len
=
batch
*
channel
*
input_depth
*
input_height
*
input_width
;
std
::
vector
<
int
>
ksize
({
3
,
3
,
3
});
std
::
vector
<
int
>
strides
({
2
,
2
,
2
});
std
::
vector
<
int
>
paddings
({
1
,
1
,
1
});
int
output_depth
=
(
input_depth
-
ksize
[
0
]
+
2
*
paddings
[
0
])
/
strides
[
0
]
+
1
;
int
output_height
=
(
input_height
-
ksize
[
1
]
+
2
*
paddings
[
1
])
/
strides
[
1
]
+
1
;
int
output_width
=
(
input_width
-
ksize
[
2
]
+
2
*
paddings
[
2
])
/
strides
[
2
]
+
1
;
int
output_len
=
output_depth
*
output_height
*
output_width
;
input_tmp
.
mutable_data
<
float
>
(
{
batch
,
channel
,
input_depth
,
input_height
,
input_width
},
paddle
::
platform
::
CPUPlace
());
output_tmp
.
mutable_data
<
float
>
(
{
batch
,
channel
,
output_depth
,
output_height
,
output_width
},
paddle
::
platform
::
CPUPlace
());
float
*
arr
=
new
float
[
in_len
];
auto
*
place
=
new
paddle
::
platform
::
GPUPlace
();
// input
float
*
input_ptr
=
input_tmp
.
data
<
float
>
();
for
(
int
i
=
0
;
i
<
in_len
;
++
i
)
arr
[
i
]
=
i
;
// rand() / double(RAND_MAX/2);
memcpy
(
input_ptr
,
arr
,
in_len
*
sizeof
(
float
));
input
.
CopyFrom
<
float
>
(
input_tmp
,
*
place
);
// input_grad
input_ptr
=
input_tmp
.
data
<
float
>
();
for
(
int
i
=
0
;
i
<
in_len
;
++
i
)
arr
[
i
]
=
0
;
memcpy
(
input_ptr
,
arr
,
in_len
*
sizeof
(
float
));
input_grad
.
CopyFrom
<
float
>
(
input_tmp
,
*
place
);
// output
input_ptr
=
output_tmp
.
data
<
float
>
();
for
(
int
i
=
0
;
i
<
output_len
;
++
i
)
arr
[
i
]
=
0
;
// rand() / double(RAND_MAX/2);
memcpy
(
input_ptr
,
arr
,
output_len
*
sizeof
(
float
));
output
.
CopyFrom
<
float
>
(
input_tmp
,
*
place
);
// output_grad
input_ptr
=
output_tmp
.
data
<
float
>
();
for
(
int
i
=
0
;
i
<
output_len
;
++
i
)
arr
[
i
]
=
1
;
// rand() / double(RAND_MAX/2);
memcpy
(
input_ptr
,
arr
,
output_len
*
sizeof
(
float
));
output_grad
.
CopyFrom
<
float
>
(
input_tmp
,
*
place
);
paddle
::
platform
::
DeviceContext
*
context
=
new
paddle
::
platform
::
CUDADeviceContext
(
paddle
::
platform
::
GPUPlace
());
paddle
::
operators
::
math
::
pool
::
maxPool
<
float
>
pool_process
;
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
float
>
poolGrad_process
;
testPool3d
<
paddle
::
operators
::
math
::
pool
::
maxPool
<
float
>
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
float
>>
(
*
context
,
pool_process
,
poolGrad_process
,
input
,
input_grad
,
output
,
output_grad
,
ksize
,
strides
,
paddings
);
}
int
main
()
{
test3dPool
();
}
#endif
\ No newline at end of file
paddle/operators/math/pooling.cc
浏览文件 @
dfc8d3c1
...
@@ -19,12 +19,12 @@ namespace operators {
...
@@ -19,12 +19,12 @@ namespace operators {
namespace
math
{
namespace
math
{
template
<
typename
PoolProcess
,
typename
T
>
template
<
typename
PoolProcess
,
typename
T
>
class
Pool2dF
orwardF
unctor
<
platform
::
CPUPlace
,
PoolProcess
,
T
>
{
class
Pool2dFunctor
<
platform
::
CPUPlace
,
PoolProcess
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
)
{
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
compute
)
{
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
input_height
=
input
.
dims
()[
2
];
const
int
input_height
=
input
.
dims
()[
2
];
const
int
input_width
=
input
.
dims
()[
3
];
const
int
input_width
=
input
.
dims
()[
3
];
...
@@ -54,14 +54,14 @@ class Pool2dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
...
@@ -54,14 +54,14 @@ class Pool2dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
int
wstart
=
pw
*
stride_width
-
padding_width
;
int
wstart
=
pw
*
stride_width
-
padding_width
;
int
wend
=
std
::
min
(
wstart
+
ksize_width
,
input_width
);
int
wend
=
std
::
min
(
wstart
+
ksize_width
,
input_width
);
wstart
=
std
::
max
(
wstart
,
0
);
wstart
=
std
::
max
(
wstart
,
0
);
T
ele
=
pool_
process
.
initial
();
T
ele
=
pool_
compute
.
initial
();
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
pool_
process
.
process
(
ele
,
input_data
[
h
*
input_width
+
w
]);
pool_
compute
.
compute
(
ele
,
input_data
[
h
*
input_width
+
w
]);
}
}
}
}
int
pool_size
=
(
hend
-
hstart
)
*
(
wend
-
wstart
);
int
pool_size
=
(
hend
-
hstart
)
*
(
wend
-
wstart
);
pool_
process
.
finalize
(
ele
,
(
static_cast
<
T
>
(
pool_size
)));
pool_
compute
.
finalize
(
ele
,
(
static_cast
<
T
>
(
pool_size
)));
output_data
[
ph
*
output_width
+
pw
]
=
ele
;
output_data
[
ph
*
output_width
+
pw
]
=
ele
;
}
}
}
}
...
@@ -73,14 +73,14 @@ class Pool2dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
...
@@ -73,14 +73,14 @@ class Pool2dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
};
};
template
<
typename
PoolProcess
,
class
T
>
template
<
typename
PoolProcess
,
class
T
>
class
Pool2d
Backwar
dFunctor
<
platform
::
CPUPlace
,
PoolProcess
,
T
>
{
class
Pool2d
Gra
dFunctor
<
platform
::
CPUPlace
,
PoolProcess
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
)
{
PoolProcess
pool_
compute
)
{
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
input_height
=
input
.
dims
()[
2
];
const
int
input_height
=
input
.
dims
()[
2
];
const
int
input_width
=
input
.
dims
()[
3
];
const
int
input_width
=
input
.
dims
()[
3
];
...
@@ -115,8 +115,7 @@ class Pool2dBackwardFunctor<platform::CPUPlace, PoolProcess, T> {
...
@@ -115,8 +115,7 @@ class Pool2dBackwardFunctor<platform::CPUPlace, PoolProcess, T> {
float
scale
=
1.0
/
pool_size
;
float
scale
=
1.0
/
pool_size
;
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
pool_process
.
gradProcess
(
pool_compute
.
compute
(
input_data
[
h
*
input_width
+
w
],
input_data
[
h
*
input_width
+
w
],
output_data
[
ph
*
output_width
+
pw
],
output_data
[
ph
*
output_width
+
pw
],
output_grad_data
[
ph
*
output_width
+
pw
],
output_grad_data
[
ph
*
output_width
+
pw
],
input_grad_data
[
h
*
input_width
+
w
],
input_grad_data
[
h
*
input_width
+
w
],
...
@@ -135,7 +134,7 @@ class Pool2dBackwardFunctor<platform::CPUPlace, PoolProcess, T> {
...
@@ -135,7 +134,7 @@ class Pool2dBackwardFunctor<platform::CPUPlace, PoolProcess, T> {
};
};
template
<
class
T
>
template
<
class
T
>
class
MaxPool2d
Backwar
dFunctor
<
platform
::
CPUPlace
,
T
>
{
class
MaxPool2d
Gra
dFunctor
<
platform
::
CPUPlace
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
...
@@ -195,37 +194,33 @@ class MaxPool2dBackwardFunctor<platform::CPUPlace, T> {
...
@@ -195,37 +194,33 @@ class MaxPool2dBackwardFunctor<platform::CPUPlace, T> {
}
}
};
};
template
class
MaxPool2dBackwardFunctor
<
platform
::
CPUPlace
,
float
>;
template
class
MaxPool2dGradFunctor
<
platform
::
CPUPlace
,
float
>;
// template class MaxPool2dBackwardFunctor<platform::CPUPlace, double>;
// template class MaxPool2dGradFunctor<platform::CPUPlace, double>;
template
class
Pool2dForwardFunctor
<
template
class
Pool2dFunctor
<
platform
::
CPUPlace
,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
float
>,
float
>
;
paddle
::
operators
::
math
::
maxPool
<
float
>,
float
>
;
template
class
Pool2dForwardFunctor
<
template
class
Pool2dFunctor
<
platform
::
CPUPlace
,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
float
>,
float
>
;
paddle
::
operators
::
math
::
avgPool
<
float
>,
float
>
;
template
class
Pool2dBackwardFunctor
<
template
class
Pool2dGradFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
float
>,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
maxPoolGrad
<
float
>,
float
>
;
float
>
;
template
class
Pool2dGradFunctor
<
template
class
Pool2dBackwardFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
avgPoolGrad
<
float
>,
float
>
;
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
float
>,
template
class
Pool2dFunctor
<
platform
::
CPUPlace
,
float
>
;
paddle
::
operators
::
math
::
maxPool
<
double
>,
double
>
;
template
class
Pool2dForwardFunctor
<
template
class
Pool2dFunctor
<
platform
::
CPUPlace
,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
double
>,
double
>
;
paddle
::
operators
::
math
::
avgPool
<
double
>,
double
>
;
template
class
Pool2dForwardFunctor
<
template
class
Pool2dGradFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
double
>,
double
>
;
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
maxPoolGrad
<
double
>,
double
>
;
template
class
Pool2dBackwardFunctor
<
template
class
Pool2dGradFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
double
>,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
avgPoolGrad
<
double
>,
double
>
;
double
>
;
template
class
Pool2dBackwardFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
double
>,
double
>
;
template
<
typename
PoolProcess
,
class
T
>
template
<
typename
PoolProcess
,
class
T
>
class
Pool3dF
orwardF
unctor
<
platform
::
CPUPlace
,
PoolProcess
,
T
>
{
class
Pool3dFunctor
<
platform
::
CPUPlace
,
PoolProcess
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
)
{
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
compute
)
{
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
input_depth
=
input
.
dims
()[
2
];
const
int
input_depth
=
input
.
dims
()[
2
];
const
int
input_height
=
input
.
dims
()[
3
];
const
int
input_height
=
input
.
dims
()[
3
];
...
@@ -265,11 +260,11 @@ class Pool3dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
...
@@ -265,11 +260,11 @@ class Pool3dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
int
wend
=
std
::
min
(
wstart
+
ksize_width
,
input_width
);
int
wend
=
std
::
min
(
wstart
+
ksize_width
,
input_width
);
wstart
=
std
::
max
(
wstart
,
0
);
wstart
=
std
::
max
(
wstart
,
0
);
int
output_idx
=
(
pd
*
output_height
+
ph
)
*
output_width
+
pw
;
int
output_idx
=
(
pd
*
output_height
+
ph
)
*
output_width
+
pw
;
T
ele
=
pool_
process
.
initial
();
T
ele
=
pool_
compute
.
initial
();
for
(
int
d
=
dstart
;
d
<
dend
;
++
d
)
{
for
(
int
d
=
dstart
;
d
<
dend
;
++
d
)
{
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
pool_
process
.
process
(
pool_
compute
.
compute
(
ele
,
ele
,
input_data
[(
d
*
input_height
+
h
)
*
input_width
+
w
]);
input_data
[(
d
*
input_height
+
h
)
*
input_width
+
w
]);
}
}
...
@@ -277,7 +272,7 @@ class Pool3dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
...
@@ -277,7 +272,7 @@ class Pool3dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
}
}
int
pool_size
=
int
pool_size
=
(
dend
-
dstart
)
*
(
hend
-
hstart
)
*
(
wend
-
wstart
);
(
dend
-
dstart
)
*
(
hend
-
hstart
)
*
(
wend
-
wstart
);
pool_
process
.
finalize
(
ele
,
static_cast
<
T
>
(
pool_size
));
pool_
compute
.
finalize
(
ele
,
static_cast
<
T
>
(
pool_size
));
output_data
[
output_idx
]
=
ele
;
output_data
[
output_idx
]
=
ele
;
}
}
}
}
...
@@ -290,14 +285,14 @@ class Pool3dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
...
@@ -290,14 +285,14 @@ class Pool3dForwardFunctor<platform::CPUPlace, PoolProcess, T> {
};
};
template
<
typename
PoolProcess
,
class
T
>
template
<
typename
PoolProcess
,
class
T
>
class
Pool3d
Backwar
dFunctor
<
platform
::
CPUPlace
,
PoolProcess
,
T
>
{
class
Pool3d
Gra
dFunctor
<
platform
::
CPUPlace
,
PoolProcess
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
)
{
PoolProcess
pool_
compute
)
{
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
input_depth
=
input
.
dims
()[
2
];
const
int
input_depth
=
input
.
dims
()[
2
];
const
int
input_height
=
input
.
dims
()[
3
];
const
int
input_height
=
input
.
dims
()[
3
];
...
@@ -348,7 +343,7 @@ class Pool3dBackwardFunctor<platform::CPUPlace, PoolProcess, T> {
...
@@ -348,7 +343,7 @@ class Pool3dBackwardFunctor<platform::CPUPlace, PoolProcess, T> {
int
input_idx
=
(
d
*
input_height
+
h
)
*
input_width
+
w
;
int
input_idx
=
(
d
*
input_height
+
h
)
*
input_width
+
w
;
int
output_idx
=
int
output_idx
=
(
pd
*
output_height
+
ph
)
*
output_width
+
pw
;
(
pd
*
output_height
+
ph
)
*
output_width
+
pw
;
pool_
process
.
gradProcess
(
pool_
compute
.
compute
(
input_data
[
input_idx
],
output_data
[
output_idx
],
input_data
[
input_idx
],
output_data
[
output_idx
],
output_grad_data
[
output_idx
],
output_grad_data
[
output_idx
],
input_grad_data
[
input_idx
],
static_cast
<
T
>
(
scale
));
input_grad_data
[
input_idx
],
static_cast
<
T
>
(
scale
));
...
@@ -368,7 +363,7 @@ class Pool3dBackwardFunctor<platform::CPUPlace, PoolProcess, T> {
...
@@ -368,7 +363,7 @@ class Pool3dBackwardFunctor<platform::CPUPlace, PoolProcess, T> {
};
};
template
<
class
T
>
template
<
class
T
>
class
MaxPool3d
Backwar
dFunctor
<
platform
::
CPUPlace
,
T
>
{
class
MaxPool3d
Gra
dFunctor
<
platform
::
CPUPlace
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
...
@@ -442,29 +437,25 @@ class MaxPool3dBackwardFunctor<platform::CPUPlace, T> {
...
@@ -442,29 +437,25 @@ class MaxPool3dBackwardFunctor<platform::CPUPlace, T> {
}
}
};
};
template
class
MaxPool3dBackwardFunctor
<
platform
::
CPUPlace
,
float
>;
template
class
MaxPool3dGradFunctor
<
platform
::
CPUPlace
,
float
>;
// template class MaxPool3dBackwardFunctor<platform::CPUPlace, double>;
// template class MaxPool3dGradFunctor<platform::CPUPlace, double>;
template
class
Pool3dForwardFunctor
<
template
class
Pool3dFunctor
<
platform
::
CPUPlace
,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
float
>,
float
>
;
paddle
::
operators
::
math
::
maxPool
<
float
>,
float
>
;
template
class
Pool3dForwardFunctor
<
template
class
Pool3dFunctor
<
platform
::
CPUPlace
,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
float
>,
float
>
;
paddle
::
operators
::
math
::
avgPool
<
float
>,
float
>
;
template
class
Pool3dBackwardFunctor
<
template
class
Pool3dGradFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
float
>,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
maxPoolGrad
<
float
>,
float
>
;
float
>
;
template
class
Pool3dGradFunctor
<
template
class
Pool3dBackwardFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
avgPoolGrad
<
float
>,
float
>
;
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
float
>,
template
class
Pool3dFunctor
<
platform
::
CPUPlace
,
float
>
;
paddle
::
operators
::
math
::
maxPool
<
double
>,
double
>
;
template
class
Pool3dForwardFunctor
<
template
class
Pool3dFunctor
<
platform
::
CPUPlace
,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
double
>,
double
>
;
paddle
::
operators
::
math
::
avgPool
<
double
>,
double
>
;
template
class
Pool3dForwardFunctor
<
template
class
Pool3dGradFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
double
>,
double
>
;
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
maxPoolGrad
<
double
>,
double
>
;
template
class
Pool3dBackwardFunctor
<
template
class
Pool3dGradFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
double
>,
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
avgPoolGrad
<
double
>,
double
>
;
double
>
;
template
class
Pool3dBackwardFunctor
<
platform
::
CPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
double
>,
double
>
;
}
// namespace math
}
// namespace math
}
// namespace operators
}
// namespace operators
}
// namespace paddle
}
// namespace paddle
paddle/operators/math/pooling.cu
浏览文件 @
dfc8d3c1
...
@@ -25,7 +25,7 @@ __global__ void KernelPool2dForward(
...
@@ -25,7 +25,7 @@ __global__ void KernelPool2dForward(
const
int
input_height
,
const
int
input_width
,
const
int
output_height
,
const
int
input_height
,
const
int
input_width
,
const
int
output_height
,
const
int
output_width
,
const
int
ksize_height
,
const
int
ksize_width
,
const
int
output_width
,
const
int
ksize_height
,
const
int
ksize_width
,
const
int
stride_height
,
const
int
stride_width
,
const
int
padding_height
,
const
int
stride_height
,
const
int
stride_width
,
const
int
padding_height
,
const
int
padding_width
,
PoolProcess
pool_
process
)
{
const
int
padding_width
,
PoolProcess
pool_
compute
)
{
int
index
=
blockIdx
.
x
*
blockDim
.
x
+
threadIdx
.
x
;
int
index
=
blockIdx
.
x
*
blockDim
.
x
+
threadIdx
.
x
;
if
(
index
<
nthreads
)
{
if
(
index
<
nthreads
)
{
int
pw
=
index
%
output_width
;
int
pw
=
index
%
output_width
;
...
@@ -42,14 +42,14 @@ __global__ void KernelPool2dForward(
...
@@ -42,14 +42,14 @@ __global__ void KernelPool2dForward(
wstart
=
max
(
wstart
,
0
);
wstart
=
max
(
wstart
,
0
);
input_data
+=
(
batch_idx
*
channels
+
c
)
*
input_height
*
input_width
;
input_data
+=
(
batch_idx
*
channels
+
c
)
*
input_height
*
input_width
;
T
ele
=
pool_
process
.
initial
();
T
ele
=
pool_
compute
.
initial
();
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
pool_
process
.
process
(
ele
,
input_data
[
h
*
input_width
+
w
]);
pool_
compute
.
compute
(
ele
,
input_data
[
h
*
input_width
+
w
]);
}
}
}
}
int
pool_size
=
(
hend
-
hstart
)
*
(
wend
-
wstart
);
int
pool_size
=
(
hend
-
hstart
)
*
(
wend
-
wstart
);
pool_
process
.
finalize
(
ele
,
(
static_cast
<
T
>
(
pool_size
)));
pool_
compute
.
finalize
(
ele
,
(
static_cast
<
T
>
(
pool_size
)));
output_data
[
index
]
=
ele
;
output_data
[
index
]
=
ele
;
}
}
}
}
...
@@ -61,7 +61,7 @@ __global__ void KernelPool2dBackward(
...
@@ -61,7 +61,7 @@ __global__ void KernelPool2dBackward(
const
int
input_height
,
const
int
input_width
,
const
int
output_height
,
const
int
input_height
,
const
int
input_width
,
const
int
output_height
,
const
int
output_width
,
const
int
ksize_height
,
const
int
ksize_width
,
const
int
output_width
,
const
int
ksize_height
,
const
int
ksize_width
,
const
int
stride_height
,
const
int
stride_width
,
const
int
padding_height
,
const
int
stride_height
,
const
int
stride_width
,
const
int
padding_height
,
const
int
padding_width
,
PoolProcess
pool_
process
)
{
const
int
padding_width
,
PoolProcess
pool_
compute
)
{
int
index
=
blockIdx
.
x
*
blockDim
.
x
+
threadIdx
.
x
;
int
index
=
blockIdx
.
x
*
blockDim
.
x
+
threadIdx
.
x
;
if
(
index
<
nthreads
)
{
if
(
index
<
nthreads
)
{
int
offsetW
=
index
%
input_width
+
padding_width
;
int
offsetW
=
index
%
input_width
+
padding_width
;
...
@@ -93,7 +93,7 @@ __global__ void KernelPool2dBackward(
...
@@ -93,7 +93,7 @@ __global__ void KernelPool2dBackward(
wstart
=
max
(
wstart
,
0
);
wstart
=
max
(
wstart
,
0
);
int
pool_size
=
(
hend
-
hstart
)
*
(
wend
-
wstart
);
int
pool_size
=
(
hend
-
hstart
)
*
(
wend
-
wstart
);
int
output_sub_idx
=
ph
*
output_width
+
pw
;
int
output_sub_idx
=
ph
*
output_width
+
pw
;
pool_
process
.
gradProcess
(
input
,
output_data
[
output_sub_idx
],
pool_
compute
.
compute
(
input
,
output_data
[
output_sub_idx
],
output_grad
[
output_sub_idx
],
gradient
,
output_grad
[
output_sub_idx
],
gradient
,
static_cast
<
T
>
(
1.0
/
pool_size
));
static_cast
<
T
>
(
1.0
/
pool_size
));
}
}
...
@@ -148,12 +148,12 @@ __global__ void KernelMaxPool2dBackward(
...
@@ -148,12 +148,12 @@ __global__ void KernelMaxPool2dBackward(
}
}
template
<
typename
PoolProcess
,
typename
T
>
template
<
typename
PoolProcess
,
typename
T
>
class
Pool2dF
orwardF
unctor
<
platform
::
GPUPlace
,
PoolProcess
,
T
>
{
class
Pool2dFunctor
<
platform
::
GPUPlace
,
PoolProcess
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
)
{
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
compute
)
{
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
input_channels
=
input
.
dims
()[
1
];
const
int
input_channels
=
input
.
dims
()[
1
];
const
int
input_height
=
input
.
dims
()[
2
];
const
int
input_height
=
input
.
dims
()[
2
];
...
@@ -184,19 +184,19 @@ class Pool2dForwardFunctor<platform::GPUPlace, PoolProcess, T> {
...
@@ -184,19 +184,19 @@ class Pool2dForwardFunctor<platform::GPUPlace, PoolProcess, T> {
input_height
,
input_width
,
output_height
,
input_height
,
input_width
,
output_height
,
output_width
,
ksize_height
,
ksize_width
,
output_width
,
ksize_height
,
ksize_width
,
stride_height
,
stride_width
,
padding_height
,
stride_height
,
stride_width
,
padding_height
,
padding_width
,
pool_
process
);
padding_width
,
pool_
compute
);
}
}
};
};
template
<
typename
PoolProcess
,
typename
T
>
template
<
typename
PoolProcess
,
typename
T
>
class
Pool2d
Backwar
dFunctor
<
platform
::
GPUPlace
,
PoolProcess
,
T
>
{
class
Pool2d
Gra
dFunctor
<
platform
::
GPUPlace
,
PoolProcess
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
)
{
PoolProcess
pool_
compute
)
{
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
input_channels
=
input
.
dims
()[
1
];
const
int
input_channels
=
input
.
dims
()[
1
];
const
int
input_height
=
input
.
dims
()[
2
];
const
int
input_height
=
input
.
dims
()[
2
];
...
@@ -228,12 +228,12 @@ class Pool2dBackwardFunctor<platform::GPUPlace, PoolProcess, T> {
...
@@ -228,12 +228,12 @@ class Pool2dBackwardFunctor<platform::GPUPlace, PoolProcess, T> {
nthreads
,
input_data
,
output_data
,
output_grad_data
,
input_grad_data
,
nthreads
,
input_data
,
output_data
,
output_grad_data
,
input_grad_data
,
input_channels
,
input_height
,
input_width
,
output_height
,
output_width
,
input_channels
,
input_height
,
input_width
,
output_height
,
output_width
,
ksize_height
,
ksize_width
,
stride_height
,
stride_width
,
padding_height
,
ksize_height
,
ksize_width
,
stride_height
,
stride_width
,
padding_height
,
padding_width
,
pool_
process
);
padding_width
,
pool_
compute
);
}
}
};
};
template
<
typename
T
>
template
<
typename
T
>
class
MaxPool2d
Backwar
dFunctor
<
platform
::
GPUPlace
,
T
>
{
class
MaxPool2d
Gra
dFunctor
<
platform
::
GPUPlace
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
...
@@ -275,29 +275,25 @@ class MaxPool2dBackwardFunctor<platform::GPUPlace, T> {
...
@@ -275,29 +275,25 @@ class MaxPool2dBackwardFunctor<platform::GPUPlace, T> {
}
}
};
};
template
class
MaxPool2dBackwardFunctor
<
platform
::
GPUPlace
,
float
>;
template
class
MaxPool2dGradFunctor
<
platform
::
GPUPlace
,
float
>;
// template class MaxPool2dBackwardFunctor<platform::GPUPlace, double>;
// template class MaxPool2dGradFunctor<platform::GPUPlace, double>;
template
class
Pool2dForwardFunctor
<
template
class
Pool2dFunctor
<
platform
::
GPUPlace
,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
float
>,
float
>
;
paddle
::
operators
::
math
::
maxPool
<
float
>,
float
>
;
template
class
Pool2dForwardFunctor
<
template
class
Pool2dFunctor
<
platform
::
GPUPlace
,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
float
>,
float
>
;
paddle
::
operators
::
math
::
avgPool
<
float
>,
float
>
;
template
class
Pool2dBackwardFunctor
<
template
class
Pool2dGradFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
float
>,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
maxPoolGrad
<
float
>,
float
>
;
float
>
;
template
class
Pool2dGradFunctor
<
template
class
Pool2dBackwardFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
avgPoolGrad
<
float
>,
float
>
;
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
float
>,
template
class
Pool2dFunctor
<
platform
::
GPUPlace
,
float
>
;
paddle
::
operators
::
math
::
maxPool
<
double
>,
double
>
;
template
class
Pool2dForwardFunctor
<
template
class
Pool2dFunctor
<
platform
::
GPUPlace
,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
double
>,
double
>
;
paddle
::
operators
::
math
::
avgPool
<
double
>,
double
>
;
template
class
Pool2dForwardFunctor
<
template
class
Pool2dGradFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
double
>,
double
>
;
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
maxPoolGrad
<
double
>,
double
>
;
template
class
Pool2dBackwardFunctor
<
template
class
Pool2dGradFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
double
>,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
avgPoolGrad
<
double
>,
double
>
;
double
>
;
template
class
Pool2dBackwardFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
double
>,
double
>
;
template
<
typename
PoolProcess
,
typename
T
>
template
<
typename
PoolProcess
,
typename
T
>
__global__
void
KernelPool3DForward
(
__global__
void
KernelPool3DForward
(
...
@@ -307,7 +303,7 @@ __global__ void KernelPool3DForward(
...
@@ -307,7 +303,7 @@ __global__ void KernelPool3DForward(
const
int
ksize_depth
,
const
int
ksize_height
,
const
int
ksize_width
,
const
int
ksize_depth
,
const
int
ksize_height
,
const
int
ksize_width
,
const
int
stride_depth
,
const
int
stride_height
,
const
int
stride_width
,
const
int
stride_depth
,
const
int
stride_height
,
const
int
stride_width
,
const
int
padding_depth
,
const
int
padding_height
,
const
int
padding_width
,
const
int
padding_depth
,
const
int
padding_height
,
const
int
padding_width
,
PoolProcess
pool_
process
)
{
PoolProcess
pool_
compute
)
{
for
(
int
index
=
blockIdx
.
x
*
blockDim
.
x
+
threadIdx
.
x
;
index
<
(
nthreads
);
for
(
int
index
=
blockIdx
.
x
*
blockDim
.
x
+
threadIdx
.
x
;
index
<
(
nthreads
);
index
+=
blockDim
.
x
*
gridDim
.
x
)
{
index
+=
blockDim
.
x
*
gridDim
.
x
)
{
int
pw
=
index
%
output_width
;
int
pw
=
index
%
output_width
;
...
@@ -325,19 +321,19 @@ __global__ void KernelPool3DForward(
...
@@ -325,19 +321,19 @@ __global__ void KernelPool3DForward(
dstart
=
max
(
dstart
,
0
);
dstart
=
max
(
dstart
,
0
);
hstart
=
max
(
hstart
,
0
);
hstart
=
max
(
hstart
,
0
);
wstart
=
max
(
wstart
,
0
);
wstart
=
max
(
wstart
,
0
);
T
ele
=
pool_
process
.
initial
();
T
ele
=
pool_
compute
.
initial
();
input_data
+=
input_data
+=
(
batch_idx
*
channels
+
c
)
*
input_depth
*
input_height
*
input_width
;
(
batch_idx
*
channels
+
c
)
*
input_depth
*
input_height
*
input_width
;
for
(
int
d
=
dstart
;
d
<
dend
;
++
d
)
{
for
(
int
d
=
dstart
;
d
<
dend
;
++
d
)
{
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
h
=
hstart
;
h
<
hend
;
++
h
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
for
(
int
w
=
wstart
;
w
<
wend
;
++
w
)
{
pool_
process
.
process
(
pool_
compute
.
compute
(
ele
,
input_data
[(
d
*
input_height
+
h
)
*
input_width
+
w
]);
ele
,
input_data
[(
d
*
input_height
+
h
)
*
input_width
+
w
]);
}
}
}
}
}
}
int
pool_size
=
(
dend
-
dstart
)
*
(
hend
-
hstart
)
*
(
wend
-
wstart
);
int
pool_size
=
(
dend
-
dstart
)
*
(
hend
-
hstart
)
*
(
wend
-
wstart
);
pool_
process
.
finalize
(
ele
,
static_cast
<
T
>
(
pool_size
));
pool_
compute
.
finalize
(
ele
,
static_cast
<
T
>
(
pool_size
));
output_data
[
index
]
=
ele
;
output_data
[
index
]
=
ele
;
}
}
}
}
...
@@ -351,7 +347,7 @@ __global__ void KernelPool3DBackward(
...
@@ -351,7 +347,7 @@ __global__ void KernelPool3DBackward(
const
int
ksize_depth
,
const
int
ksize_height
,
const
int
ksize_width
,
const
int
ksize_depth
,
const
int
ksize_height
,
const
int
ksize_width
,
const
int
stride_depth
,
const
int
stride_height
,
const
int
stride_width
,
const
int
stride_depth
,
const
int
stride_height
,
const
int
stride_width
,
const
int
padding_depth
,
const
int
padding_height
,
const
int
padding_width
,
const
int
padding_depth
,
const
int
padding_height
,
const
int
padding_width
,
PoolProcess
pool_
process
)
{
PoolProcess
pool_
compute
)
{
for
(
int
index
=
blockIdx
.
x
*
blockDim
.
x
+
threadIdx
.
x
;
index
<
(
nthreads
);
for
(
int
index
=
blockIdx
.
x
*
blockDim
.
x
+
threadIdx
.
x
;
index
<
(
nthreads
);
index
+=
blockDim
.
x
*
gridDim
.
x
)
{
index
+=
blockDim
.
x
*
gridDim
.
x
)
{
int
offsetW
=
index
%
input_width
+
padding_width
;
int
offsetW
=
index
%
input_width
+
padding_width
;
...
@@ -396,7 +392,7 @@ __global__ void KernelPool3DBackward(
...
@@ -396,7 +392,7 @@ __global__ void KernelPool3DBackward(
wstart
=
max
(
wstart
,
0
);
wstart
=
max
(
wstart
,
0
);
int
pool_size
=
(
dend
-
dstart
)
*
(
hend
-
hstart
)
*
(
wend
-
wstart
);
int
pool_size
=
(
dend
-
dstart
)
*
(
hend
-
hstart
)
*
(
wend
-
wstart
);
int
output_sub_idx
=
(
pd
*
output_height
+
ph
)
*
output_width
+
pw
;
int
output_sub_idx
=
(
pd
*
output_height
+
ph
)
*
output_width
+
pw
;
pool_
process
.
gradProcess
(
input
,
output_data
[
output_sub_idx
],
pool_
compute
.
compute
(
input
,
output_data
[
output_sub_idx
],
output_grad
[
output_sub_idx
],
gradient
,
output_grad
[
output_sub_idx
],
gradient
,
static_cast
<
T
>
(
1.0
/
pool_size
));
static_cast
<
T
>
(
1.0
/
pool_size
));
}
}
...
@@ -459,12 +455,12 @@ __global__ void KernelMaxPool3DBackward(
...
@@ -459,12 +455,12 @@ __global__ void KernelMaxPool3DBackward(
}
}
template
<
typename
PoolProcess
,
class
T
>
template
<
typename
PoolProcess
,
class
T
>
class
Pool3dF
orwardF
unctor
<
platform
::
GPUPlace
,
PoolProcess
,
T
>
{
class
Pool3dFunctor
<
platform
::
GPUPlace
,
PoolProcess
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
)
{
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
compute
)
{
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
input_channels
=
input
.
dims
()[
1
];
const
int
input_channels
=
input
.
dims
()[
1
];
const
int
input_depth
=
input
.
dims
()[
2
];
const
int
input_depth
=
input
.
dims
()[
2
];
...
@@ -502,19 +498,19 @@ class Pool3dForwardFunctor<platform::GPUPlace, PoolProcess, T> {
...
@@ -502,19 +498,19 @@ class Pool3dForwardFunctor<platform::GPUPlace, PoolProcess, T> {
input_height
,
input_width
,
output_depth
,
output_height
,
output_width
,
input_height
,
input_width
,
output_depth
,
output_height
,
output_width
,
ksize_depth
,
ksize_height
,
ksize_width
,
stride_depth
,
stride_height
,
ksize_depth
,
ksize_height
,
ksize_width
,
stride_depth
,
stride_height
,
stride_width
,
padding_depth
,
padding_height
,
padding_width
,
stride_width
,
padding_depth
,
padding_height
,
padding_width
,
pool_
process
);
pool_
compute
);
}
}
};
};
template
<
typename
PoolProcess
,
class
T
>
template
<
typename
PoolProcess
,
class
T
>
class
Pool3d
Backwar
dFunctor
<
platform
::
GPUPlace
,
PoolProcess
,
T
>
{
class
Pool3d
Gra
dFunctor
<
platform
::
GPUPlace
,
PoolProcess
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
)
{
PoolProcess
pool_
compute
)
{
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
input_channels
=
input
.
dims
()[
1
];
const
int
input_channels
=
input
.
dims
()[
1
];
const
int
input_depth
=
input
.
dims
()[
2
];
const
int
input_depth
=
input
.
dims
()[
2
];
...
@@ -554,12 +550,12 @@ class Pool3dBackwardFunctor<platform::GPUPlace, PoolProcess, T> {
...
@@ -554,12 +550,12 @@ class Pool3dBackwardFunctor<platform::GPUPlace, PoolProcess, T> {
input_channels
,
input_depth
,
input_height
,
input_width
,
output_depth
,
input_channels
,
input_depth
,
input_height
,
input_width
,
output_depth
,
output_height
,
output_width
,
ksize_depth
,
ksize_height
,
ksize_width
,
output_height
,
output_width
,
ksize_depth
,
ksize_height
,
ksize_width
,
stride_depth
,
stride_height
,
stride_width
,
padding_depth
,
stride_depth
,
stride_height
,
stride_width
,
padding_depth
,
padding_height
,
padding_width
,
pool_
process
);
padding_height
,
padding_width
,
pool_
compute
);
}
}
};
};
template
<
class
T
>
template
<
class
T
>
class
MaxPool3d
Backwar
dFunctor
<
platform
::
GPUPlace
,
T
>
{
class
MaxPool3d
Gra
dFunctor
<
platform
::
GPUPlace
,
T
>
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
...
@@ -608,29 +604,25 @@ class MaxPool3dBackwardFunctor<platform::GPUPlace, T> {
...
@@ -608,29 +604,25 @@ class MaxPool3dBackwardFunctor<platform::GPUPlace, T> {
}
}
};
};
template
class
MaxPool3dBackwardFunctor
<
platform
::
GPUPlace
,
float
>;
template
class
MaxPool3dGradFunctor
<
platform
::
GPUPlace
,
float
>;
// template class MaxPool3dBackwardFunctor<platform::GPUPlace, double>;
// template class MaxPool3dGradFunctor<platform::GPUPlace, double>;
template
class
Pool3dForwardFunctor
<
template
class
Pool3dFunctor
<
platform
::
GPUPlace
,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
float
>,
float
>
;
paddle
::
operators
::
math
::
maxPool
<
float
>,
float
>
;
template
class
Pool3dForwardFunctor
<
template
class
Pool3dFunctor
<
platform
::
GPUPlace
,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
float
>,
float
>
;
paddle
::
operators
::
math
::
avgPool
<
float
>,
float
>
;
template
class
Pool3dBackwardFunctor
<
template
class
Pool3dGradFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
float
>,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
maxPoolGrad
<
float
>,
float
>
;
float
>
;
template
class
Pool3dGradFunctor
<
template
class
Pool3dBackwardFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
avgPoolGrad
<
float
>,
float
>
;
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
float
>,
template
class
Pool3dFunctor
<
platform
::
GPUPlace
,
float
>
;
paddle
::
operators
::
math
::
maxPool
<
double
>,
double
>
;
template
class
Pool3dForwardFunctor
<
template
class
Pool3dFunctor
<
platform
::
GPUPlace
,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
double
>,
double
>
;
paddle
::
operators
::
math
::
avgPool
<
double
>,
double
>
;
template
class
Pool3dForwardFunctor
<
template
class
Pool3dGradFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
double
>,
double
>
;
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
maxPoolGrad
<
double
>,
double
>
;
template
class
Pool3dBackwardFunctor
<
template
class
Pool3dGradFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
maxPoolGrad
<
double
>,
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
avgPoolGrad
<
double
>,
double
>
;
double
>
;
template
class
Pool3dBackwardFunctor
<
platform
::
GPUPlace
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
double
>,
double
>
;
}
// namespace math
}
// namespace math
}
// namespace operators
}
// namespace operators
...
...
paddle/operators/math/pooling.h
浏览文件 @
dfc8d3c1
...
@@ -21,17 +21,15 @@ limitations under the License. */
...
@@ -21,17 +21,15 @@ limitations under the License. */
namespace
paddle
{
namespace
paddle
{
namespace
operators
{
namespace
operators
{
namespace
math
{
namespace
math
{
//////////////////////
//////////////////////
#define FLT_MAX __FLT_MAX__
#define FLT_MAX __FLT_MAX__
/////////////////////
/////////////////////
namespace
pool
{
template
<
class
T
>
template
<
class
T
>
class
maxPool
{
class
maxPool
{
public:
public:
DEVICE
inline
T
initial
()
{
return
static_cast
<
T
>
(
-
FLT_MAX
);
}
DEVICE
inline
T
initial
()
{
return
static_cast
<
T
>
(
-
FLT_MAX
);
}
DEVICE
inline
void
process
(
T
&
y
,
const
T
&
x
)
{
y
=
y
>
x
?
y
:
x
;
}
DEVICE
inline
void
compute
(
T
&
y
,
const
T
&
x
)
{
y
=
y
>
x
?
y
:
x
;
}
DEVICE
inline
void
finalize
(
T
&
y
,
const
T
&
poo_size
)
{}
DEVICE
inline
void
finalize
(
T
&
y
,
const
T
&
poo_size
)
{}
};
};
...
@@ -39,13 +37,13 @@ template <class T>
...
@@ -39,13 +37,13 @@ template <class T>
class
avgPool
{
class
avgPool
{
public:
public:
DEVICE
inline
T
initial
()
{
return
static_cast
<
T
>
(
0
);
}
DEVICE
inline
T
initial
()
{
return
static_cast
<
T
>
(
0
);
}
DEVICE
inline
void
process
(
T
&
y
,
const
T
&
x
)
{
y
+=
x
;
}
DEVICE
inline
void
compute
(
T
&
y
,
const
T
&
x
)
{
y
+=
x
;
}
DEVICE
inline
void
finalize
(
T
&
y
,
const
T
&
poo_size
)
{
y
/=
poo_size
;
}
DEVICE
inline
void
finalize
(
T
&
y
,
const
T
&
poo_size
)
{
y
/=
poo_size
;
}
};
};
template
<
class
T
>
template
<
class
T
>
class
maxPoolGrad
{
class
maxPoolGrad
{
public:
public:
DEVICE
inline
void
gradProcess
(
const
T
&
x
,
const
T
&
y
,
const
T
&
dy
,
T
&
dx
,
DEVICE
inline
void
compute
(
const
T
&
x
,
const
T
&
y
,
const
T
&
dy
,
T
&
dx
,
T
scale
)
{
T
scale
)
{
dx
+=
dy
*
(
x
==
y
);
dx
+=
dy
*
(
x
==
y
);
}
}
...
@@ -54,35 +52,34 @@ class maxPoolGrad {
...
@@ -54,35 +52,34 @@ class maxPoolGrad {
template
<
class
T
>
template
<
class
T
>
class
avgPoolGrad
{
class
avgPoolGrad
{
public:
public:
DEVICE
inline
void
gradProcess
(
const
T
&
x
,
const
T
&
y
,
const
T
&
dy
,
T
&
dx
,
DEVICE
inline
void
compute
(
const
T
&
x
,
const
T
&
y
,
const
T
&
dy
,
T
&
dx
,
T
scale
)
{
T
scale
)
{
dx
+=
(
scale
*
dy
);
dx
+=
(
scale
*
dy
);
}
}
};
};
}
// namespace pool
template
<
typename
Place
,
typename
PoolProcess
,
typename
T
>
template
<
typename
Place
,
typename
PoolProcess
,
typename
T
>
class
Pool2dF
orwardF
unctor
{
class
Pool2dFunctor
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
);
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
compute
);
};
};
template
<
typename
Place
,
typename
PoolProcess
,
typename
T
>
template
<
typename
Place
,
typename
PoolProcess
,
typename
T
>
class
Pool2d
Backwar
dFunctor
{
class
Pool2d
Gra
dFunctor
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
);
PoolProcess
pool_
compute
);
};
};
template
<
typename
Place
,
class
T
>
template
<
typename
Place
,
class
T
>
class
MaxPool2d
Backwar
dFunctor
{
class
MaxPool2d
Gra
dFunctor
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
...
@@ -92,27 +89,27 @@ class MaxPool2dBackwardFunctor {
...
@@ -92,27 +89,27 @@ class MaxPool2dBackwardFunctor {
};
};
template
<
typename
Place
,
typename
PoolProcess
,
typename
T
>
template
<
typename
Place
,
typename
PoolProcess
,
typename
T
>
class
Pool3dF
orwardF
unctor
{
class
Pool3dFunctor
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
output
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
);
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
compute
);
};
};
template
<
typename
Place
,
typename
PoolProcess
,
typename
T
>
template
<
typename
Place
,
typename
PoolProcess
,
typename
T
>
class
Pool3d
Backwar
dFunctor
{
class
Pool3d
Gra
dFunctor
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
const
framework
::
Tensor
&
output_grad
,
std
::
vector
<
int
>&
ksize
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
std
::
vector
<
int
>&
strides
,
std
::
vector
<
int
>&
paddings
,
PoolProcess
pool_
process
);
PoolProcess
pool_
compute
);
};
};
template
<
typename
Place
,
class
T
>
template
<
typename
Place
,
class
T
>
class
MaxPool3d
Backwar
dFunctor
{
class
MaxPool3d
Gra
dFunctor
{
public:
public:
void
operator
()(
const
platform
::
DeviceContext
&
context
,
void
operator
()(
const
platform
::
DeviceContext
&
context
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
const
framework
::
Tensor
&
input
,
framework
::
Tensor
&
input_grad
,
...
...
paddle/operators/pool_op.cc
浏览文件 @
dfc8d3c1
...
@@ -17,7 +17,7 @@ limitations under the License. */
...
@@ -17,7 +17,7 @@ limitations under the License. */
namespace
paddle
{
namespace
paddle
{
namespace
operators
{
namespace
operators
{
int
outputSize_p
ool
(
int
input_size
,
int
filter_size
,
int
padding
,
int
stride
)
{
int
OutputSizeP
ool
(
int
input_size
,
int
filter_size
,
int
padding
,
int
stride
)
{
int
output_size
=
(
input_size
-
filter_size
+
2
*
padding
)
/
stride
+
1
;
int
output_size
=
(
input_size
-
filter_size
+
2
*
padding
)
/
stride
+
1
;
return
output_size
;
return
output_size
;
}
}
...
@@ -33,7 +33,7 @@ class PoolOp : public framework::OperatorWithKernel {
...
@@ -33,7 +33,7 @@ class PoolOp : public framework::OperatorWithKernel {
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
OutputVar
(
"Out"
),
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
OutputVar
(
"Out"
),
"Out(Output) of Pooling should not be null."
);
"Out(Output) of Pooling should not be null."
);
auto
in_
X
=
ctx
.
Input
<
Tensor
>
(
"X"
);
auto
in_
x
=
ctx
.
Input
<
Tensor
>
(
"X"
);
auto
out
=
ctx
.
Output
<
Tensor
>
(
"Out"
);
auto
out
=
ctx
.
Output
<
Tensor
>
(
"Out"
);
int
global_pooling
=
Attr
<
int
>
(
"globalPooling"
);
int
global_pooling
=
Attr
<
int
>
(
"globalPooling"
);
std
::
string
pooling_type
=
Attr
<
std
::
string
>
(
"poolingType"
);
std
::
string
pooling_type
=
Attr
<
std
::
string
>
(
"poolingType"
);
...
@@ -43,29 +43,24 @@ class PoolOp : public framework::OperatorWithKernel {
...
@@ -43,29 +43,24 @@ class PoolOp : public framework::OperatorWithKernel {
PADDLE_ENFORCE
(
pooling_type
==
"max"
||
pooling_type
==
"avg"
,
PADDLE_ENFORCE
(
pooling_type
==
"max"
||
pooling_type
==
"avg"
,
"pooling_type should be 'max' or 'avg'"
);
"pooling_type should be 'max' or 'avg'"
);
PADDLE_ENFORCE
(
in_
X
->
dims
().
size
()
==
4
||
in_X
->
dims
().
size
()
==
5
,
PADDLE_ENFORCE
(
in_
x
->
dims
().
size
()
==
4
||
in_x
->
dims
().
size
()
==
5
,
"Pooling intput should be 4-D or 5-D"
);
"Pooling intput should be 4-D or 5-D"
);
PADDLE_ENFORCE
(
ksize
.
size
()
==
2
||
ksize
.
size
()
==
3
,
"Pooling size should be 2 elements. or 3 elements."
);
PADDLE_ENFORCE_EQ
(
ksize
.
size
(),
strides
.
size
(),
"strides size and pooling size should be the same."
);
PADDLE_ENFORCE_EQ
(
ksize
.
size
(),
paddings
.
size
(),
"paddings size and pooling size should be the same."
);
if
(
global_pooling
==
1
)
{
if
(
global_pooling
==
1
)
{
ksize
.
resize
(
static_cast
<
size_t
>
(
in_
X
->
dims
().
size
())
-
2
);
ksize
.
resize
(
static_cast
<
size_t
>
(
in_
x
->
dims
().
size
())
-
2
);
for
(
size_t
i
=
0
;
i
<
ksize
.
size
();
++
i
)
for
(
size_t
i
=
0
;
i
<
ksize
.
size
();
++
i
)
ksize
[
i
]
=
static_cast
<
int
>
(
in_
X
->
dims
()[
i
+
2
]);
ksize
[
i
]
=
static_cast
<
int
>
(
in_
x
->
dims
()[
i
+
2
]);
}
}
if
(
ksize
.
size
()
==
2
)
{
std
::
vector
<
int64_t
>
output_shape
({
in_x
->
dims
()[
0
],
in_x
->
dims
()[
1
]});
PADDLE_ENFORCE_EQ
(
strides
.
size
(),
2
,
"Pool2DOp strides size should be 2 elements."
);
PADDLE_ENFORCE_EQ
(
paddings
.
size
(),
2
,
"Pool2DOp paddings size should be 2 elements"
);
}
else
{
PADDLE_ENFORCE_EQ
(
strides
.
size
(),
3
,
"Pool3DOp strides should be 3 elements."
);
PADDLE_ENFORCE_EQ
(
paddings
.
size
(),
3
,
"Pool3DOp paddings should be 3 elements."
);
}
std
::
vector
<
int64_t
>
output_shape
({
in_X
->
dims
()[
0
],
in_X
->
dims
()[
1
]});
for
(
size_t
i
=
0
;
i
<
ksize
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
ksize
.
size
();
++
i
)
{
output_shape
.
push_back
(
outputSize_pool
(
in_X
->
dims
()[
i
+
2
],
ksize
[
i
],
output_shape
.
push_back
(
OutputSizePool
(
in_x
->
dims
()[
i
+
2
],
ksize
[
i
],
paddings
[
i
],
strides
[
i
]));
paddings
[
i
],
strides
[
i
]));
}
}
out
->
Resize
(
framework
::
make_ddim
(
output_shape
));
out
->
Resize
(
framework
::
make_ddim
(
output_shape
));
...
@@ -78,9 +73,16 @@ class PoolOpGrad : public framework::OperatorWithKernel {
...
@@ -78,9 +73,16 @@ class PoolOpGrad : public framework::OperatorWithKernel {
protected:
protected:
void
InferShape
(
const
framework
::
InferShapeContext
&
ctx
)
const
override
{
void
InferShape
(
const
framework
::
InferShapeContext
&
ctx
)
const
override
{
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
InputVar
(
"X"
),
"X(Input) of Pooling should not be null."
);
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
InputVar
(
"Out"
),
"Out(Output) of Pooling should not be null."
);
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
Output
<
Tensor
>
(
framework
::
GradVarName
(
"X"
)),
"Input@Grad of Pooling should not be null."
);
auto
in
=
ctx
.
Input
<
Tensor
>
(
"X"
);
auto
in
=
ctx
.
Input
<
Tensor
>
(
"X"
);
auto
d_in
=
ctx
.
Output
<
Tensor
>
(
framework
::
GradVarName
(
"X"
));
auto
d_in
=
ctx
.
Output
<
Tensor
>
(
framework
::
GradVarName
(
"X"
));
if
(
d_in
)
d_in
->
Resize
(
in
->
dims
());
d_in
->
Resize
(
in
->
dims
());
}
}
};
};
...
@@ -92,7 +94,7 @@ class Pool2dOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -92,7 +94,7 @@ class Pool2dOpMaker : public framework::OpProtoAndCheckerMaker {
"X"
,
"X"
,
"The input tensor of pooling operator. "
"The input tensor of pooling operator. "
"The format of input tensor is NCHW. Where N is batch size, C is the "
"The format of input tensor is NCHW. Where N is batch size, C is the "
"number of channels, H and W is the height and width of
imag
e."
);
"number of channels, H and W is the height and width of
featur
e."
);
AddOutput
(
"Out"
,
AddOutput
(
"Out"
,
"The output tensor of pooling operator."
"The output tensor of pooling operator."
"The format of output tensor is also NCHW."
);
"The format of output tensor is also NCHW."
);
...
@@ -166,7 +168,7 @@ class Pool3dOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -166,7 +168,7 @@ class Pool3dOpMaker : public framework::OpProtoAndCheckerMaker {
"The format of input tensor is NCDHW. Where N is batch size, C is "
"The format of input tensor is NCDHW. Where N is batch size, C is "
"the "
"the "
"number of channels, D, H and W is the depth, height and width of "
"number of channels, D, H and W is the depth, height and width of "
"
imag
e."
);
"
featur
e."
);
AddOutput
(
"Out"
,
AddOutput
(
"Out"
,
"The output tensor of pooling operator."
"The output tensor of pooling operator."
"The format of output tensor is also NCDHW."
);
"The format of output tensor is also NCDHW."
);
...
...
paddle/operators/pool_op.cu
浏览文件 @
dfc8d3c1
/* Copyright (c) 2016 PaddlePaddle Authors All Rights Reserve.
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
Licensed under the Apache License, Version 2.0 (the "License");
You may obtain a copy of the License at
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
See the License for the specific language governing permissions and
limitations under the License. */
limitations under the License. */
#include "paddle/operators/pool_op.h"
#include "paddle/operators/pool_op.h"
...
...
paddle/operators/pool_op.h
浏览文件 @
dfc8d3c1
...
@@ -28,7 +28,7 @@ template <typename Place, typename T>
...
@@ -28,7 +28,7 @@ template <typename Place, typename T>
class
PoolKernel
:
public
framework
::
OpKernel
{
class
PoolKernel
:
public
framework
::
OpKernel
{
public:
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
const
Tensor
*
in_
X
=
context
.
Input
<
Tensor
>
(
"X"
);
const
Tensor
*
in_
x
=
context
.
Input
<
Tensor
>
(
"X"
);
Tensor
*
out
=
context
.
Output
<
Tensor
>
(
"Out"
);
Tensor
*
out
=
context
.
Output
<
Tensor
>
(
"Out"
);
int
global_pooling
=
context
.
Attr
<
int
>
(
"globalPooling"
);
int
global_pooling
=
context
.
Attr
<
int
>
(
"globalPooling"
);
...
@@ -38,43 +38,43 @@ class PoolKernel : public framework::OpKernel {
...
@@ -38,43 +38,43 @@ class PoolKernel : public framework::OpKernel {
std
::
vector
<
int
>
paddings
=
context
.
Attr
<
std
::
vector
<
int
>>
(
"paddings"
);
std
::
vector
<
int
>
paddings
=
context
.
Attr
<
std
::
vector
<
int
>>
(
"paddings"
);
if
(
global_pooling
==
1
)
{
if
(
global_pooling
==
1
)
{
for
(
size_t
i
=
0
;
i
<
ksize
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
ksize
.
size
();
++
i
)
{
ksize
[
i
]
=
in_X
->
dims
()[
i
+
2
]
;
ksize
[
i
]
=
static_cast
<
int
>
(
in_x
->
dims
()[
i
+
2
])
;
}
}
}
}
switch
(
ksize
.
size
())
{
switch
(
ksize
.
size
())
{
case
2
:
{
case
2
:
{
if
(
pooling_type
==
"max"
)
{
if
(
pooling_type
==
"max"
)
{
paddle
::
operators
::
math
::
Pool2dF
orwardF
unctor
<
paddle
::
operators
::
math
::
Pool2dFunctor
<
Place
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
T
>
,
T
>
Place
,
paddle
::
operators
::
math
::
maxPool
<
T
>
,
T
>
pool2d_forward
;
pool2d_forward
;
paddle
::
operators
::
math
::
pool
::
maxPool
<
T
>
pool_process
;
paddle
::
operators
::
math
::
maxPool
<
T
>
pool_process
;
pool2d_forward
(
context
.
device_context
(),
*
in_
X
,
*
out
,
ksize
,
strides
,
pool2d_forward
(
context
.
device_context
(),
*
in_
x
,
*
out
,
ksize
,
strides
,
paddings
,
pool_process
);
paddings
,
pool_process
);
}
else
if
(
pooling_type
==
"avg"
)
{
}
else
if
(
pooling_type
==
"avg"
)
{
paddle
::
operators
::
math
::
Pool2dF
orwardF
unctor
<
paddle
::
operators
::
math
::
Pool2dFunctor
<
Place
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
T
>
,
T
>
Place
,
paddle
::
operators
::
math
::
avgPool
<
T
>
,
T
>
pool2d_forward
;
pool2d_forward
;
paddle
::
operators
::
math
::
pool
::
avgPool
<
T
>
pool_process
;
paddle
::
operators
::
math
::
avgPool
<
T
>
pool_process
;
pool2d_forward
(
context
.
device_context
(),
*
in_
X
,
*
out
,
ksize
,
strides
,
pool2d_forward
(
context
.
device_context
(),
*
in_
x
,
*
out
,
ksize
,
strides
,
paddings
,
pool_process
);
paddings
,
pool_process
);
}
}
}
break
;
}
break
;
case
3
:
{
case
3
:
{
if
(
pooling_type
==
"max"
)
{
if
(
pooling_type
==
"max"
)
{
paddle
::
operators
::
math
::
Pool3dF
orwardF
unctor
<
paddle
::
operators
::
math
::
Pool3dFunctor
<
Place
,
paddle
::
operators
::
math
::
pool
::
maxPool
<
T
>
,
T
>
Place
,
paddle
::
operators
::
math
::
maxPool
<
T
>
,
T
>
pool3d_forward
;
pool3d_forward
;
paddle
::
operators
::
math
::
pool
::
maxPool
<
T
>
pool_process
;
paddle
::
operators
::
math
::
maxPool
<
T
>
pool_process
;
pool3d_forward
(
context
.
device_context
(),
*
in_
X
,
*
out
,
ksize
,
strides
,
pool3d_forward
(
context
.
device_context
(),
*
in_
x
,
*
out
,
ksize
,
strides
,
paddings
,
pool_process
);
paddings
,
pool_process
);
}
else
if
(
pooling_type
==
"avg"
)
{
}
else
if
(
pooling_type
==
"avg"
)
{
paddle
::
operators
::
math
::
Pool3dF
orwardF
unctor
<
paddle
::
operators
::
math
::
Pool3dFunctor
<
Place
,
paddle
::
operators
::
math
::
pool
::
avgPool
<
T
>
,
T
>
Place
,
paddle
::
operators
::
math
::
avgPool
<
T
>
,
T
>
pool3d_forward
;
pool3d_forward
;
paddle
::
operators
::
math
::
pool
::
avgPool
<
T
>
pool_process
;
paddle
::
operators
::
math
::
avgPool
<
T
>
pool_process
;
pool3d_forward
(
context
.
device_context
(),
*
in_
X
,
*
out
,
ksize
,
strides
,
pool3d_forward
(
context
.
device_context
(),
*
in_
x
,
*
out
,
ksize
,
strides
,
paddings
,
pool_process
);
paddings
,
pool_process
);
}
}
}
break
;
}
break
;
...
@@ -86,11 +86,11 @@ template <typename Place, typename T>
...
@@ -86,11 +86,11 @@ template <typename Place, typename T>
class
PoolGradKernel
:
public
framework
::
OpKernel
{
class
PoolGradKernel
:
public
framework
::
OpKernel
{
public:
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
const
Tensor
*
in_
X
=
context
.
Input
<
Tensor
>
(
"X"
);
const
Tensor
*
in_
x
=
context
.
Input
<
Tensor
>
(
"X"
);
const
Tensor
*
out
=
context
.
Input
<
Tensor
>
(
"Out"
);
const
Tensor
*
out
=
context
.
Input
<
Tensor
>
(
"Out"
);
const
Tensor
*
out_grad
=
const
Tensor
*
out_grad
=
context
.
Input
<
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
context
.
Input
<
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
Tensor
*
in_
X
_grad
=
context
.
Output
<
Tensor
>
(
framework
::
GradVarName
(
"X"
));
Tensor
*
in_
x
_grad
=
context
.
Output
<
Tensor
>
(
framework
::
GradVarName
(
"X"
));
int
global_pooling
=
context
.
Attr
<
int
>
(
"globalPooling"
);
int
global_pooling
=
context
.
Attr
<
int
>
(
"globalPooling"
);
std
::
string
pooling_type
=
context
.
Attr
<
std
::
string
>
(
"poolingType"
);
std
::
string
pooling_type
=
context
.
Attr
<
std
::
string
>
(
"poolingType"
);
...
@@ -99,43 +99,44 @@ class PoolGradKernel : public framework::OpKernel {
...
@@ -99,43 +99,44 @@ class PoolGradKernel : public framework::OpKernel {
std
::
vector
<
int
>
paddings
=
context
.
Attr
<
std
::
vector
<
int
>>
(
"paddings"
);
std
::
vector
<
int
>
paddings
=
context
.
Attr
<
std
::
vector
<
int
>>
(
"paddings"
);
if
(
global_pooling
==
1
)
{
if
(
global_pooling
==
1
)
{
for
(
size_t
i
=
0
;
i
<
ksize
.
size
();
++
i
)
ksize
[
i
]
=
in_X
->
dims
()[
i
+
2
];
for
(
size_t
i
=
0
;
i
<
ksize
.
size
();
++
i
)
ksize
[
i
]
=
static_cast
<
int
>
(
in_x
->
dims
()[
i
+
2
]);
}
}
if
(
in_
X
_grad
)
{
if
(
in_
x
_grad
)
{
in_
X
_grad
->
mutable_data
<
T
>
(
context
.
GetPlace
());
in_
x
_grad
->
mutable_data
<
T
>
(
context
.
GetPlace
());
auto
temp
=
framework
::
EigenVector
<
T
>::
Flatten
(
*
in_
X
_grad
);
auto
temp
=
framework
::
EigenVector
<
T
>::
Flatten
(
*
in_
x
_grad
);
temp
.
device
(
context
.
GetEigenDevice
<
Place
>
())
=
temp
.
device
(
context
.
GetEigenDevice
<
Place
>
())
=
temp
.
constant
(
static_cast
<
T
>
(
0
));
temp
.
constant
(
static_cast
<
T
>
(
0
));
switch
(
ksize
.
size
())
{
switch
(
ksize
.
size
())
{
case
2
:
{
case
2
:
{
if
(
pooling_type
==
"max"
)
{
if
(
pooling_type
==
"max"
)
{
paddle
::
operators
::
math
::
MaxPool2d
Backwar
dFunctor
<
Place
,
T
>
paddle
::
operators
::
math
::
MaxPool2d
Gra
dFunctor
<
Place
,
T
>
pool2d_backward
;
pool2d_backward
;
pool2d_backward
(
context
.
device_context
(),
*
in_
X
,
*
in_X
_grad
,
*
out
,
pool2d_backward
(
context
.
device_context
(),
*
in_
x
,
*
in_x
_grad
,
*
out
,
*
out_grad
,
ksize
,
strides
,
paddings
);
*
out_grad
,
ksize
,
strides
,
paddings
);
}
else
if
(
pooling_type
==
"avg"
)
{
}
else
if
(
pooling_type
==
"avg"
)
{
paddle
::
operators
::
math
::
Pool2d
Backwar
dFunctor
<
paddle
::
operators
::
math
::
Pool2d
Gra
dFunctor
<
Place
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
T
>
,
T
>
Place
,
paddle
::
operators
::
math
::
avgPoolGrad
<
T
>
,
T
>
pool2d_backward
;
pool2d_backward
;
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
T
>
pool_process
;
paddle
::
operators
::
math
::
avgPoolGrad
<
T
>
pool_process
;
pool2d_backward
(
context
.
device_context
(),
*
in_
X
,
*
in_X
_grad
,
*
out
,
pool2d_backward
(
context
.
device_context
(),
*
in_
x
,
*
in_x
_grad
,
*
out
,
*
out_grad
,
ksize
,
strides
,
paddings
,
pool_process
);
*
out_grad
,
ksize
,
strides
,
paddings
,
pool_process
);
}
}
}
break
;
}
break
;
case
3
:
{
case
3
:
{
if
(
pooling_type
==
"max"
)
{
if
(
pooling_type
==
"max"
)
{
paddle
::
operators
::
math
::
MaxPool3d
Backwar
dFunctor
<
Place
,
T
>
paddle
::
operators
::
math
::
MaxPool3d
Gra
dFunctor
<
Place
,
T
>
pool3d_backward
;
pool3d_backward
;
pool3d_backward
(
context
.
device_context
(),
*
in_
X
,
*
in_X
_grad
,
*
out
,
pool3d_backward
(
context
.
device_context
(),
*
in_
x
,
*
in_x
_grad
,
*
out
,
*
out_grad
,
ksize
,
strides
,
paddings
);
*
out_grad
,
ksize
,
strides
,
paddings
);
}
else
if
(
pooling_type
==
"avg"
)
{
}
else
if
(
pooling_type
==
"avg"
)
{
paddle
::
operators
::
math
::
Pool3d
Backwar
dFunctor
<
paddle
::
operators
::
math
::
Pool3d
Gra
dFunctor
<
Place
,
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
T
>
,
T
>
Place
,
paddle
::
operators
::
math
::
avgPoolGrad
<
T
>
,
T
>
pool3d_backward
;
pool3d_backward
;
paddle
::
operators
::
math
::
pool
::
avgPoolGrad
<
T
>
pool_process
;
paddle
::
operators
::
math
::
avgPoolGrad
<
T
>
pool_process
;
pool3d_backward
(
context
.
device_context
(),
*
in_
X
,
*
in_X
_grad
,
*
out
,
pool3d_backward
(
context
.
device_context
(),
*
in_
x
,
*
in_x
_grad
,
*
out
,
*
out_grad
,
ksize
,
strides
,
paddings
,
pool_process
);
*
out_grad
,
ksize
,
strides
,
paddings
,
pool_process
);
}
}
}
break
;
}
break
;
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录