Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
f82fa64a
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
f82fa64a
编写于
2月 12, 2018
作者:
K
kexinzhao
提交者:
GitHub
2月 12, 2018
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Move float16 into fluid folder (#8394)
* move float16 into fluid * fix include * move to platform folder
上级
432d2b5d
变更
5
隐藏空白更改
内联
并排
Showing
5 changed file
with
17 addition
and
25 deletion
+17
-25
paddle/fluid/platform/CMakeLists.txt
paddle/fluid/platform/CMakeLists.txt
+4
-1
paddle/fluid/platform/float16.h
paddle/fluid/platform/float16.h
+11
-20
paddle/fluid/platform/float16_test.cc
paddle/fluid/platform/float16_test.cc
+1
-1
paddle/fluid/platform/float16_test.cu
paddle/fluid/platform/float16_test.cu
+1
-1
paddle/math/tests/CMakeLists.txt
paddle/math/tests/CMakeLists.txt
+0
-2
未找到文件。
paddle/fluid/platform/CMakeLists.txt
浏览文件 @
f82fa64a
...
...
@@ -27,7 +27,7 @@ ELSE()
set
(
MKLDNN_CTX_DEPS
)
ENDIF
()
# memcpy de
o
ends on device_context, here add deps individually for
# memcpy de
p
ends on device_context, here add deps individually for
# avoiding cycle dependencies
cc_library
(
device_context SRCS device_context.cc DEPS memory buddy_allocator
system_allocator memory_block meta_data meta_cache place eigen3
${
GPU_CTX_DEPS
}
${
MKLDNN_CTX_DEPS
}
)
...
...
@@ -39,3 +39,6 @@ nv_test(nccl_test SRCS nccl_test.cu DEPS dynload_cuda gpu_info device_context)
cc_library
(
profiler SRCS profiler.cc DEPS device_context
)
cc_test
(
profiler_test SRCS profiler_test.cc DEPS profiler
)
nv_test
(
float16_gpu_test SRCS float16_test.cu
)
cc_test
(
float16_test SRCS float16_test.cc
)
paddle/
math
/float16.h
→
paddle/
fluid/platform
/float16.h
浏览文件 @
f82fa64a
...
...
@@ -68,7 +68,7 @@ namespace paddle {
// memory access of float16 struct and also makes float16 compatible
// with CUDA half, ARM float16_t, and Eigen::half data types.
struct
PADDLE_ALIGN
(
2
)
float16
{
public:
public:
uint16_t
x
;
// Constructors
...
...
@@ -319,7 +319,7 @@ public:
return
static_cast
<
double
>
(
float
(
*
this
));
}
private:
private:
union
Bits
{
float
f
;
int32_t
si
;
...
...
@@ -485,8 +485,7 @@ HOST inline float16 operator+(const float16& a, const float16& b) {
"st1 {v0.h}[0], [%[res_ptr]]
\n
"
:
// outputs
:
// inputs
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
res_ptr
]
"r"
(
&
(
res
.
x
))
:
// clobbers
"memory"
,
"v0"
,
"v1"
);
...
...
@@ -502,8 +501,7 @@ HOST inline float16 operator-(const float16& a, const float16& b) {
"st1 {v0.h}[0], [%[res_ptr]]
\n
"
:
// outputs
:
// inputs
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
res_ptr
]
"r"
(
&
(
res
.
x
))
:
// clobbers
"memory"
,
"v0"
,
"v1"
);
...
...
@@ -519,8 +517,7 @@ HOST inline float16 operator*(const float16& a, const float16& b) {
"st1 {v0.h}[0], [%[res_ptr]]
\n
"
:
// outputs
:
// inputs
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
res_ptr
]
"r"
(
&
(
res
.
x
))
:
// clobbers
"memory"
,
"v0"
,
"v1"
);
...
...
@@ -536,8 +533,7 @@ HOST inline float16 operator/(const float16& a, const float16& b) {
"st1 {v0.h}[0], [%[res_ptr]]
\n
"
:
// outputs
:
// inputs
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
res_ptr
]
"r"
(
&
(
res
.
x
))
:
// clobbers
"memory"
,
"v0"
,
"v1"
);
...
...
@@ -588,8 +584,7 @@ HOST inline bool operator==(const float16& a, const float16& b) {
"st1 {v0.h}[0], [%[res_ptr]]
\n
"
:
// outputs
:
// inputs
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
res_ptr
]
"r"
(
&
res
)
:
// clobbers
"memory"
,
"v0"
,
"v1"
);
...
...
@@ -609,8 +604,7 @@ HOST inline bool operator<(const float16& a, const float16& b) {
"st1 {v0.h}[0], [%[res_ptr]]
\n
"
:
// outputs
:
// inputs
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
res_ptr
]
"r"
(
&
res
)
:
// clobbers
"memory"
,
"v0"
,
"v1"
);
...
...
@@ -626,8 +620,7 @@ HOST inline bool operator<=(const float16& a, const float16& b) {
"st1 {v0.h}[0], [%[res_ptr]]
\n
"
:
// outputs
:
// inputs
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
res_ptr
]
"r"
(
&
res
)
:
// clobbers
"memory"
,
"v0"
,
"v1"
);
...
...
@@ -643,8 +636,7 @@ HOST inline bool operator>(const float16& a, const float16& b) {
"st1 {v0.h}[0], [%[res_ptr]]
\n
"
:
// outputs
:
// inputs
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
res_ptr
]
"r"
(
&
res
)
:
// clobbers
"memory"
,
"v0"
,
"v1"
);
...
...
@@ -660,8 +652,7 @@ HOST inline bool operator>=(const float16& a, const float16& b) {
"st1 {v0.h}[0], [%[res_ptr]]
\n
"
:
// outputs
:
// inputs
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
a_ptr
]
"r"
(
&
(
a
.
x
)),
[
b_ptr
]
"r"
(
&
(
b
.
x
)),
[
res_ptr
]
"r"
(
&
res
)
:
// clobbers
"memory"
,
"v0"
,
"v1"
);
...
...
paddle/
math/tests/test_float16.cpp
→
paddle/
fluid/platform/float16_test.cc
浏览文件 @
f82fa64a
...
...
@@ -9,7 +9,7 @@ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/
math
/float16.h"
#include "paddle/
fluid/platform
/float16.h"
#include <gtest/gtest.h>
...
...
paddle/
math/tests/test_float16
.cu
→
paddle/
fluid/platform/float16_test
.cu
浏览文件 @
f82fa64a
...
...
@@ -9,7 +9,7 @@ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/
math
/float16.h"
#include "paddle/
fluid/platform
/float16.h"
#include <gtest/gtest.h>
...
...
paddle/math/tests/CMakeLists.txt
浏览文件 @
f82fa64a
...
...
@@ -22,7 +22,6 @@ if(WITH_GPU)
link_paddle_test
(
test_Tensor
)
CUDA_ADD_EXECUTABLE
(
test_lazyAssign test_lazyAssign.cu
)
link_paddle_test
(
test_lazyAssign
)
nv_test
(
test_float16_gpu SRCS test_float16.cu
)
else
()
compile_cu_as_cpp
(
test_Tensor.cu
)
add_unittest
(
test_Tensor test_Tensor.cu
)
...
...
@@ -34,4 +33,3 @@ add_simple_unittest(test_FPException)
add_simple_unittest
(
test_GpuProfiler
)
add_simple_unittest
(
test_BaseMatrix
)
add_simple_unittest
(
test_Matrix
)
add_simple_unittest
(
test_float16
)
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录