Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
5e6848d9
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
5e6848d9
编写于
6月 29, 2020
作者:
W
Wilber
提交者:
GitHub
6月 29, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[Cherry-pick] Support compile for arm ft (#25241)
上级
4d8c10ae
变更
9
隐藏空白更改
内联
并排
Showing
9 changed file
with
71 addition
and
55 deletion
+71
-55
CMakeLists.txt
CMakeLists.txt
+7
-0
cmake/external/openblas.cmake
cmake/external/openblas.cmake
+3
-0
cmake/flags.cmake
cmake/flags.cmake
+1
-1
paddle/fluid/operators/match_matrix_tensor_op.cc
paddle/fluid/operators/match_matrix_tensor_op.cc
+2
-2
paddle/fluid/operators/pyramid_hash_op.cc
paddle/fluid/operators/pyramid_hash_op.cc
+4
-6
paddle/fluid/operators/search_compute.h
paddle/fluid/operators/search_compute.h
+46
-44
paddle/fluid/platform/cpu_info.cc
paddle/fluid/platform/cpu_info.cc
+2
-0
paddle/fluid/platform/cpu_info.h
paddle/fluid/platform/cpu_info.h
+2
-0
python/setup.py.in
python/setup.py.in
+4
-2
未找到文件。
CMakeLists.txt
浏览文件 @
5e6848d9
...
@@ -88,6 +88,7 @@ option(WITH_DGC "Use DGC(Deep Gradient Compression) or not" ${WITH_DISTRIBUTE}
...
@@ -88,6 +88,7 @@ option(WITH_DGC "Use DGC(Deep Gradient Compression) or not" ${WITH_DISTRIBUTE}
option
(
SANITIZER_TYPE
"Choose the type of sanitizer, options are: Address, Leak, Memory, Thread, Undefined"
OFF
)
option
(
SANITIZER_TYPE
"Choose the type of sanitizer, options are: Address, Leak, Memory, Thread, Undefined"
OFF
)
option
(
WITH_LITE
"Compile Paddle Fluid with Lite Engine"
OFF
)
option
(
WITH_LITE
"Compile Paddle Fluid with Lite Engine"
OFF
)
option
(
WITH_NCCL
"Compile PaddlePaddle with NCCL support"
ON
)
option
(
WITH_NCCL
"Compile PaddlePaddle with NCCL support"
ON
)
option
(
WITH_ARM
"Compile PaddlePaddle with arm support"
OFF
)
# PY_VERSION
# PY_VERSION
if
(
NOT PY_VERSION
)
if
(
NOT PY_VERSION
)
...
@@ -199,6 +200,12 @@ if(WITH_AMD_GPU)
...
@@ -199,6 +200,12 @@ if(WITH_AMD_GPU)
include
(
hip
)
include
(
hip
)
endif
(
WITH_AMD_GPU
)
endif
(
WITH_AMD_GPU
)
if
(
WITH_ARM
)
set
(
CMAKE_C_FLAGS
"
${
CMAKE_C_FLAGS
}
-fPIC"
)
set
(
CMAKE_CXX_FLAGS
"
${
CMAKE_CXX_FLAGS
}
-fPIC"
)
add_definitions
(
-DPADDLE_WITH_ARM
)
endif
()
set
(
PADDLE_PYTHON_BUILD_DIR
"
${
CMAKE_CURRENT_BINARY_DIR
}
/python/build"
)
set
(
PADDLE_PYTHON_BUILD_DIR
"
${
CMAKE_CURRENT_BINARY_DIR
}
/python/build"
)
set
(
CMAKE_CXX_FLAGS_RELWITHDEBINFO
"-O3 -g -DNDEBUG"
)
set
(
CMAKE_CXX_FLAGS_RELWITHDEBINFO
"-O3 -g -DNDEBUG"
)
...
...
cmake/external/openblas.cmake
浏览文件 @
5e6848d9
...
@@ -19,6 +19,9 @@ SET(CBLAS_SOURCE_DIR ${THIRD_PARTY_PATH}/openblas/src/extern_openblas)
...
@@ -19,6 +19,9 @@ SET(CBLAS_SOURCE_DIR ${THIRD_PARTY_PATH}/openblas/src/extern_openblas)
SET
(
CBLAS_INSTALL_DIR
${
THIRD_PARTY_PATH
}
/install/openblas
)
SET
(
CBLAS_INSTALL_DIR
${
THIRD_PARTY_PATH
}
/install/openblas
)
SET
(
CBLAS_REPOSITORY https://github.com/xianyi/OpenBLAS.git
)
SET
(
CBLAS_REPOSITORY https://github.com/xianyi/OpenBLAS.git
)
SET
(
CBLAS_TAG v0.3.7
)
SET
(
CBLAS_TAG v0.3.7
)
IF
(
WITH_ARM
)
SET
(
CBLAS_TAG v0.2.18
)
ENDIF
()
cache_third_party
(
extern_openblas
cache_third_party
(
extern_openblas
REPOSITORY
${
CBLAS_REPOSITORY
}
REPOSITORY
${
CBLAS_REPOSITORY
}
TAG
${
CBLAS_TAG
}
TAG
${
CBLAS_TAG
}
...
...
cmake/flags.cmake
浏览文件 @
5e6848d9
...
@@ -187,7 +187,7 @@ set(GPU_COMMON_FLAGS
...
@@ -187,7 +187,7 @@ set(GPU_COMMON_FLAGS
-Wno-error=unused-function
# Warnings in Numpy Header.
-Wno-error=unused-function
# Warnings in Numpy Header.
-Wno-error=array-bounds
# Warnings in Eigen::array
-Wno-error=array-bounds
# Warnings in Eigen::array
)
)
if
(
NOT WITH_NV_JETSON
)
if
(
NOT WITH_NV_JETSON
AND NOT WITH_ARM
)
set
(
CMAKE_CXX_FLAGS
"
${
CMAKE_CXX_FLAGS
}
-m64"
)
set
(
CMAKE_CXX_FLAGS
"
${
CMAKE_CXX_FLAGS
}
-m64"
)
endif
()
endif
()
endif
(
NOT WIN32
)
endif
(
NOT WIN32
)
...
...
paddle/fluid/operators/match_matrix_tensor_op.cc
浏览文件 @
5e6848d9
...
@@ -288,8 +288,8 @@ class CPUMatchMatrixTensorOPGradKernel : public framework::OpKernel<T> {
...
@@ -288,8 +288,8 @@ class CPUMatchMatrixTensorOPGradKernel : public framework::OpKernel<T> {
auto
*
r_data
=
bottom_r_data
+
(
offset_r
[
b
]
+
j
)
*
dim_in
;
auto
*
r_data
=
bottom_r_data
+
(
offset_r
[
b
]
+
j
)
*
dim_in
;
auto
*
r_diff
=
bottom_r_diff
+
(
offset_r
[
b
]
+
j
)
*
dim_in
;
auto
*
r_diff
=
bottom_r_diff
+
(
offset_r
[
b
]
+
j
)
*
dim_in
;
if
(
diff
!=
0.0
)
{
if
(
diff
!=
0.0
)
{
a
vx_a
xpy
(
r_data
,
l_trans_diff
,
dim_in
,
diff
);
axpy
(
r_data
,
l_trans_diff
,
dim_in
,
diff
);
a
vx_a
xpy
(
l_trans_data
,
r_diff
,
dim_in
,
diff
);
axpy
(
l_trans_data
,
r_diff
,
dim_in
,
diff
);
}
}
}
}
}
}
...
...
paddle/fluid/operators/pyramid_hash_op.cc
浏览文件 @
5e6848d9
...
@@ -385,8 +385,8 @@ class CPUPyramidHashOPKernel : public framework::OpKernel<T> {
...
@@ -385,8 +385,8 @@ class CPUPyramidHashOPKernel : public framework::OpKernel<T> {
}
}
auto
weight_type
=
_blobs_0
->
type
();
auto
weight_type
=
_blobs_0
->
type
();
if
(
_is_training
==
0
&&
weight_type
!=
framework
::
proto
::
VarType
::
INT8
)
{
if
(
_is_training
==
0
&&
weight_type
!=
framework
::
proto
::
VarType
::
INT8
)
{
a
vx_a
xpy_noadd
(
top_data
,
top_data
,
top
->
dims
()[
0
]
*
top
->
dims
()[
1
],
axpy_noadd
(
top_data
,
top_data
,
top
->
dims
()[
0
]
*
top
->
dims
()[
1
],
_drop_out_percent
);
_drop_out_percent
);
}
}
}
}
};
};
...
@@ -451,7 +451,7 @@ class CPUPyramidHashOPGradKernel : public framework::OpKernel<T> {
...
@@ -451,7 +451,7 @@ class CPUPyramidHashOPGradKernel : public framework::OpKernel<T> {
int
_space_len
)
const
{
int
_space_len
)
const
{
for
(
int
j
=
0
;
j
!=
_num_emb
;
j
+=
_rand_len
)
{
for
(
int
j
=
0
;
j
!=
_num_emb
;
j
+=
_rand_len
)
{
unsigned
int
pos
=
XXH32
(
hash_id
,
len
*
sizeof
(
T
),
j
)
%
_space_len
;
unsigned
int
pos
=
XXH32
(
hash_id
,
len
*
sizeof
(
T
),
j
)
%
_space_len
;
a
vx_a
xpy
(
top_pos
+
j
,
weights
+
pos
,
_rand_len
,
mlr
);
axpy
(
top_pos
+
j
,
weights
+
pos
,
_rand_len
,
mlr
);
}
}
}
}
...
@@ -525,9 +525,7 @@ REGISTER_OPERATOR(pyramid_hash_grad, ops::PyramidHashOpGrad);
...
@@ -525,9 +525,7 @@ REGISTER_OPERATOR(pyramid_hash_grad, ops::PyramidHashOpGrad);
REGISTER_OP_CPU_KERNEL
(
REGISTER_OP_CPU_KERNEL
(
pyramid_hash
,
ops
::
CPUPyramidHashOPKernel
<
plt
::
CPUDeviceContext
,
float
>
,
pyramid_hash
,
ops
::
CPUPyramidHashOPKernel
<
plt
::
CPUDeviceContext
,
float
>
,
ops
::
CPUPyramidHashOPKernel
<
plt
::
CPUDeviceContext
,
double
>
,
ops
::
CPUPyramidHashOPKernel
<
plt
::
CPUDeviceContext
,
int8_t
>
);
ops
::
CPUPyramidHashOPKernel
<
plt
::
CPUDeviceContext
,
int8_t
>
);
REGISTER_OP_CPU_KERNEL
(
REGISTER_OP_CPU_KERNEL
(
pyramid_hash_grad
,
pyramid_hash_grad
,
ops
::
CPUPyramidHashOPGradKernel
<
plt
::
CPUDeviceContext
,
float
>
,
ops
::
CPUPyramidHashOPGradKernel
<
plt
::
CPUDeviceContext
,
float
>
);
ops
::
CPUPyramidHashOPGradKernel
<
plt
::
CPUDeviceContext
,
double
>
);
paddle/fluid/operators/search_compute.h
浏览文件 @
5e6848d9
...
@@ -14,7 +14,9 @@ limitations under the License. */
...
@@ -14,7 +14,9 @@ limitations under the License. */
#pragma once
#pragma once
#if !defined(PADDLE_WITH_ARM)
#include <immintrin.h>
#include <immintrin.h>
#endif
#include <cfloat>
#include <cfloat>
#include <cmath>
#include <cmath>
#include <cstring>
#include <cstring>
...
@@ -72,6 +74,8 @@ void call_gemm_batched(const framework::ExecutionContext& ctx,
...
@@ -72,6 +74,8 @@ void call_gemm_batched(const framework::ExecutionContext& ctx,
}
}
}
}
#if !defined(PADDLE_WITH_ARM)
#define __m256x __m256
#define __m256x __m256
static
const
unsigned
int
AVX_STEP_SIZE
=
8
;
static
const
unsigned
int
AVX_STEP_SIZE
=
8
;
...
@@ -83,16 +87,25 @@ static const unsigned int AVX_CUT_LEN_MASK = 7U;
...
@@ -83,16 +87,25 @@ static const unsigned int AVX_CUT_LEN_MASK = 7U;
#define _mm256_store_px _mm256_storeu_ps
#define _mm256_store_px _mm256_storeu_ps
#define _mm256_broadcast_sx _mm256_broadcast_ss
#define _mm256_broadcast_sx _mm256_broadcast_ss
#define _mm256_mul_pd _mm256_mul_pd
#define __m128x __m128
#define _mm256_add_pd _mm256_add_pd
#define _mm256_load_pd _mm256_loadu_pd
static
const
unsigned
int
SSE_STEP_SIZE
=
2
;
#define _mm256_store_pd _mm256_storeu_pd
static
const
unsigned
int
SSE_CUT_LEN_MASK
=
1U
;
#define _mm256_broadcast_sd _mm256_broadcast_sd
#define _mm_add_px _mm_add_ps
#define _mm_mul_px _mm_mul_ps
#define _mm_load_px _mm_loadu_ps
#define _mm_store_px _mm_storeu_ps
#define _mm_load1_px _mm_load1_ps
#endif
inline
void
avx_axpy
(
const
float
*
x
,
float
*
y
,
size_t
len
,
const
float
alpha
)
{
template
<
typename
T
>
inline
void
axpy
(
const
T
*
x
,
T
*
y
,
size_t
len
,
const
T
alpha
)
{
unsigned
int
jjj
,
lll
;
unsigned
int
jjj
,
lll
;
jjj
=
lll
=
0
;
jjj
=
lll
=
0
;
#ifdef PADDLE_WITH_AVX
lll
=
len
&
~
AVX_CUT_LEN_MASK
;
lll
=
len
&
~
AVX_CUT_LEN_MASK
;
__m256x
mm_alpha
=
_mm256_broadcast_sx
(
&
alpha
);
__m256x
mm_alpha
=
_mm256_broadcast_sx
(
&
alpha
);
for
(
jjj
=
0
;
jjj
<
lll
;
jjj
+=
AVX_STEP_SIZE
)
{
for
(
jjj
=
0
;
jjj
<
lll
;
jjj
+=
AVX_STEP_SIZE
)
{
...
@@ -101,66 +114,55 @@ inline void avx_axpy(const float* x, float* y, size_t len, const float alpha) {
...
@@ -101,66 +114,55 @@ inline void avx_axpy(const float* x, float* y, size_t len, const float alpha) {
_mm256_add_px
(
_mm256_load_px
(
y
+
jjj
),
_mm256_add_px
(
_mm256_load_px
(
y
+
jjj
),
_mm256_mul_px
(
mm_alpha
,
_mm256_load_px
(
x
+
jjj
))));
_mm256_mul_px
(
mm_alpha
,
_mm256_load_px
(
x
+
jjj
))));
}
}
#elif defined(PADDLE_WITH_ARM)
for
(;
jjj
<
len
;
jjj
++
)
{
PADDLE_THROW
(
platform
::
errors
::
Unimplemented
(
"axpy is not supported"
));
y
[
jjj
]
+=
alpha
*
x
[
jjj
];
#else
lll
=
len
&
~
SSE_CUT_LEN_MASK
;
__m128x
mm_alpha
=
_mm_load1_px
(
&
alpha
);
for
(
jjj
=
0
;
jjj
<
lll
;
jjj
+=
SSE_STEP_SIZE
)
{
_mm_store_px
(
y
+
jjj
,
_mm_add_px
(
_mm_load_px
(
y
+
jjj
),
_mm_mul_px
(
mm_alpha
,
_mm_load_px
(
x
+
jjj
))));
}
}
}
inline
void
avx_axpy
(
const
double
*
x
,
double
*
y
,
size_t
len
,
const
float
alpha
)
{
unsigned
int
jjj
,
lll
;
jjj
=
lll
=
0
;
lll
=
len
&
~
AVX_CUT_LEN_MASK
;
#endif
double
alpha_d
=
static_cast
<
double
>
(
alpha
);
__m256d
mm_alpha
=
_mm256_broadcast_sd
(
&
alpha_d
);
for
(
jjj
=
0
;
jjj
<
lll
;
jjj
+=
AVX_STEP_SIZE
)
{
_mm256_store_pd
(
y
+
jjj
,
_mm256_add_pd
(
_mm256_load_pd
(
y
+
jjj
),
_mm256_mul_pd
(
mm_alpha
,
_mm256_load_pd
(
x
+
jjj
))));
}
for
(;
jjj
<
len
;
jjj
++
)
{
for
(;
jjj
<
len
;
jjj
++
)
{
y
[
jjj
]
+=
alpha
*
x
[
jjj
];
y
[
jjj
]
+=
alpha
*
x
[
jjj
];
}
}
}
}
inline
void
avx_axpy_noadd
(
const
double
*
x
,
double
*
y
,
size_t
len
,
const
float
alpha
)
{
unsigned
int
jjj
,
lll
;
jjj
=
lll
=
0
;
double
alpha_d
=
static_cast
<
double
>
(
alpha
);
lll
=
len
&
~
AVX_CUT_LEN_MASK
;
__m256d
mm_alpha
=
_mm256_broadcast_sd
(
&
alpha_d
);
for
(
jjj
=
0
;
jjj
<
lll
;
jjj
+=
AVX_STEP_SIZE
)
{
_mm256_store_pd
(
y
+
jjj
,
_mm256_mul_pd
(
mm_alpha
,
_mm256_load_pd
(
x
+
jjj
)));
}
for
(;
jjj
<
len
;
jjj
++
)
{
template
<
typename
T
>
y
[
jjj
]
=
alpha
*
x
[
jjj
];
inline
void
axpy_noadd
(
const
T
*
x
,
T
*
y
,
size_t
len
,
const
T
alpha
)
{
}
}
inline
void
avx_axpy_noadd
(
const
float
*
x
,
float
*
y
,
size_t
len
,
const
float
alpha
)
{
unsigned
int
jjj
,
lll
;
unsigned
int
jjj
,
lll
;
jjj
=
lll
=
0
;
jjj
=
lll
=
0
;
#ifdef PADDLE_WITH_AVX
lll
=
len
&
~
AVX_CUT_LEN_MASK
;
lll
=
len
&
~
AVX_CUT_LEN_MASK
;
__m256x
mm_alpha
=
_mm256_broadcast_sx
(
&
alpha
);
__m256x
mm_alpha
=
_mm256_broadcast_sx
(
&
alpha
);
for
(
jjj
=
0
;
jjj
<
lll
;
jjj
+=
AVX_STEP_SIZE
)
{
for
(
jjj
=
0
;
jjj
<
lll
;
jjj
+=
AVX_STEP_SIZE
)
{
_mm256_store_px
(
y
+
jjj
,
_mm256_mul_px
(
mm_alpha
,
_mm256_load_px
(
x
+
jjj
)));
_mm256_store_px
(
y
+
jjj
,
_mm256_mul_px
(
mm_alpha
,
_mm256_load_px
(
x
+
jjj
)));
}
}
#elif defined(PADDLE_WITH_ARM)
PADDLE_THROW
(
platform
::
errors
::
Unimplemented
(
"axpy_noadd is not supported"
));
#else
lll
=
len
&
~
SSE_CUT_LEN_MASK
;
__m128x
mm_alpha
=
_mm_load1_px
(
&
alpha
);
for
(
jjj
=
0
;
jjj
<
lll
;
jjj
+=
SSE_STEP_SIZE
)
{
_mm_store_px
(
y
+
jjj
,
_mm_mul_px
(
mm_alpha
,
_mm_load_px
(
x
+
jjj
)));
}
#endif
for
(;
jjj
<
len
;
jjj
++
)
{
for
(;
jjj
<
len
;
jjj
++
)
{
y
[
jjj
]
=
alpha
*
x
[
jjj
];
y
[
jjj
]
=
alpha
*
x
[
jjj
];
}
}
}
}
inline
void
avx_axpy_noadd
(
const
int8_t
*
x
,
int8_t
*
y
,
size_t
len
,
const
float
alpha
)
{
inline
void
axpy_noadd
(
const
int8_t
*
x
,
int8_t
*
y
,
size_t
len
,
const
float
alpha
)
{
PADDLE_THROW
(
platform
::
errors
::
Unimplemented
(
PADDLE_THROW
(
platform
::
errors
::
Unimplemented
(
"int8_t input of a
vx_axpy_noadd is
not supported"
));
"int8_t input of a
xpy_noadd is
not supported"
));
}
}
}
// namespace operators
}
// namespace operators
...
...
paddle/fluid/platform/cpu_info.cc
浏览文件 @
5e6848d9
...
@@ -139,6 +139,7 @@ bool MayIUse(const cpu_isa_t cpu_isa) {
...
@@ -139,6 +139,7 @@ bool MayIUse(const cpu_isa_t cpu_isa) {
if
(
cpu_isa
==
isa_any
)
{
if
(
cpu_isa
==
isa_any
)
{
return
true
;
return
true
;
}
else
{
}
else
{
#if !defined(WITH_NV_JETSON) && !defined(PADDLE_WITH_ARM)
int
reg
[
4
];
int
reg
[
4
];
cpuid
(
reg
,
0
);
cpuid
(
reg
,
0
);
int
nIds
=
reg
[
0
];
int
nIds
=
reg
[
0
];
...
@@ -168,6 +169,7 @@ bool MayIUse(const cpu_isa_t cpu_isa) {
...
@@ -168,6 +169,7 @@ bool MayIUse(const cpu_isa_t cpu_isa) {
}
}
}
}
#endif
#endif
#endif
}
// namespace platform
}
// namespace platform
}
// namespace paddle
}
// namespace paddle
paddle/fluid/platform/cpu_info.h
浏览文件 @
5e6848d9
...
@@ -40,12 +40,14 @@ limitations under the License. */
...
@@ -40,12 +40,14 @@ limitations under the License. */
#ifdef _WIN32
#ifdef _WIN32
#define cpuid(reg, x) __cpuidex(reg, x, 0)
#define cpuid(reg, x) __cpuidex(reg, x, 0)
#else
#else
#if !defined(WITH_NV_JETSON) && !defined(PADDLE_WITH_ARM)
#include <cpuid.h>
#include <cpuid.h>
inline
void
cpuid
(
int
reg
[
4
],
int
x
)
{
inline
void
cpuid
(
int
reg
[
4
],
int
x
)
{
__cpuid_count
(
x
,
0
,
reg
[
0
],
reg
[
1
],
reg
[
2
],
reg
[
3
]);
__cpuid_count
(
x
,
0
,
reg
[
0
],
reg
[
1
],
reg
[
2
],
reg
[
3
]);
}
}
#endif
#endif
#endif
#endif
#endif
namespace
paddle
{
namespace
paddle
{
namespace
platform
{
namespace
platform
{
...
...
python/setup.py.in
浏览文件 @
5e6848d9
...
@@ -6,6 +6,7 @@ import shutil
...
@@ -6,6 +6,7 @@ import shutil
import sys
import sys
import fnmatch
import fnmatch
import errno
import errno
import platform
from contextlib import contextmanager
from contextlib import contextmanager
from setuptools import Command
from setuptools import Command
...
@@ -301,8 +302,9 @@ if '${CMAKE_BUILD_TYPE}' == 'Release':
...
@@ -301,8 +302,9 @@ if '${CMAKE_BUILD_TYPE}' == 'Release':
command = "install_name_tool -id \"@loader_path/../libs/\" ${PADDLE_BINARY_DIR}/python/paddle/fluid/${FLUID_CORE_NAME}" + '.so'
command = "install_name_tool -id \"@loader_path/../libs/\" ${PADDLE_BINARY_DIR}/python/paddle/fluid/${FLUID_CORE_NAME}" + '.so'
else:
else:
command = "patchelf --set-rpath '$ORIGIN/../libs/' ${PADDLE_BINARY_DIR}/python/paddle/fluid/${FLUID_CORE_NAME}" + '.so'
command = "patchelf --set-rpath '$ORIGIN/../libs/' ${PADDLE_BINARY_DIR}/python/paddle/fluid/${FLUID_CORE_NAME}" + '.so'
if os.system(command) != 0:
if platform.machine() != 'aarch64':
raise Exception("patch ${FLUID_CORE_NAME}.%s failed, command: %s" % (ext_name, command))
if os.system(command) != 0:
raise Exception("patch ${FLUID_CORE_NAME}.%s failed, command: %s" % (ext_name, command))
ext_modules = [Extension('_foo', ['stub.cc'])]
ext_modules = [Extension('_foo', ['stub.cc'])]
if os.name == 'nt':
if os.name == 'nt':
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录