Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
0a198e0a
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
0a198e0a
编写于
6月 27, 2019
作者:
H
Huihuang Zheng
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Add CxxConfig and Related Java API and Improve CMakeLists
上级
c152fe93
变更
9
显示空白变更内容
内联
并排
Showing
9 changed file
with
242 addition
and
14 deletion
+242
-14
CMakeLists.txt
CMakeLists.txt
+1
-1
paddle/fluid/lite/api/CMakeLists.txt
paddle/fluid/lite/api/CMakeLists.txt
+2
-2
paddle/fluid/lite/api/android/CMakeLists.txt
paddle/fluid/lite/api/android/CMakeLists.txt
+5
-0
paddle/fluid/lite/api/android/jni/CMakeLists.txt
paddle/fluid/lite/api/android/jni/CMakeLists.txt
+4
-2
paddle/fluid/lite/api/android/jni/paddle_lite_jni.cc
paddle/fluid/lite/api/android/jni/paddle_lite_jni.cc
+84
-9
paddle/fluid/lite/api/android/jni/paddle_lite_jni.h
paddle/fluid/lite/api/android/jni/paddle_lite_jni.h
+20
-0
paddle/fluid/lite/api/android/jni/src/com/baidu/paddle/lite/PaddlePredictor.java
...ndroid/jni/src/com/baidu/paddle/lite/PaddlePredictor.java
+23
-0
paddle/fluid/lite/api/android/jni/src/com/baidu/paddle/lite/Place.java
...lite/api/android/jni/src/com/baidu/paddle/lite/Place.java
+102
-0
paddle/fluid/lite/api/android/jni/test/com/baidu/paddle/lite/PaddlePredictorTest.java
...d/jni/test/com/baidu/paddle/lite/PaddlePredictorTest.java
+1
-0
未找到文件。
CMakeLists.txt
浏览文件 @
0a198e0a
...
@@ -147,7 +147,7 @@ endif()
...
@@ -147,7 +147,7 @@ endif()
# for lite, both server and mobile framework.
# for lite, both server and mobile framework.
option
(
WITH_LITE
"Enable lite framework"
OFF
)
option
(
WITH_LITE
"Enable lite framework"
OFF
)
option
(
WITH_JAVA
"Compile PaddlePaddle with Java JNI lib
"
OFF
)
option
(
LITE_WITH_JAVA
"Enable Java JNI lib in lite mode
"
OFF
)
option
(
LITE_WITH_CUDA
"Enable CUDA in lite mode"
OFF
)
option
(
LITE_WITH_CUDA
"Enable CUDA in lite mode"
OFF
)
option
(
LITE_WITH_X86
"Enable X86 in lite mode"
ON
)
option
(
LITE_WITH_X86
"Enable X86 in lite mode"
ON
)
option
(
LITE_WITH_ARM
"Enable ARM in lite mode"
OFF
)
option
(
LITE_WITH_ARM
"Enable ARM in lite mode"
OFF
)
...
...
paddle/fluid/lite/api/CMakeLists.txt
浏览文件 @
0a198e0a
...
@@ -131,8 +131,8 @@ if (WITH_TESTING)
...
@@ -131,8 +131,8 @@ if (WITH_TESTING)
add_dependencies
(
test_paddle_api_lite extern_lite_download_lite_naive_model_tar_gz
)
add_dependencies
(
test_paddle_api_lite extern_lite_download_lite_naive_model_tar_gz
)
endif
()
endif
()
if
(
WITH_JAVA AND LITE_WITH_ARM
)
if
(
LITE_
WITH_JAVA AND LITE_WITH_ARM
)
add_subdirectory
(
android
/jni
)
add_subdirectory
(
android
)
endif
()
endif
()
#lite_cc_binary(cxx_api_lite_bin SRCS cxx_api_bin.cc
#lite_cc_binary(cxx_api_lite_bin SRCS cxx_api_bin.cc
...
...
paddle/fluid/lite/api/android/CMakeLists.txt
0 → 100644
浏览文件 @
0a198e0a
if
((
NOT LITE_WITH_JAVA
)
OR
(
NOT LITE_WITH_ARM
))
return
()
endif
()
add_subdirectory
(
jni
)
paddle/fluid/lite/api/android/jni/CMakeLists.txt
浏览文件 @
0a198e0a
if
((
NOT
WITH_LITE
)
OR
(
NOT
WITH_JAVA
))
if
((
NOT
LITE_WITH_ARM
)
OR
(
NOT LITE_
WITH_JAVA
))
return
()
return
()
endif
()
endif
()
...
@@ -15,7 +15,9 @@ find_package(JNI REQUIRED)
...
@@ -15,7 +15,9 @@ find_package(JNI REQUIRED)
# Generate PaddlePredictor.jar
# Generate PaddlePredictor.jar
include_directories
(
${
JNI_INCLUDE_DIRS
}
)
include_directories
(
${
JNI_INCLUDE_DIRS
}
)
add_jar
(
PaddlePredictor src/com/baidu/paddle/lite/PaddlePredictor.java
)
add_jar
(
PaddlePredictor
src/com/baidu/paddle/lite/PaddlePredictor.java
src/com/baidu/paddle/lite/Place.java
)
get_target_property
(
_jarFile PaddlePredictor JAR_FILE
)
get_target_property
(
_jarFile PaddlePredictor JAR_FILE
)
get_target_property
(
_classDir PaddlePredictor CLASSDIR
)
get_target_property
(
_classDir PaddlePredictor CLASSDIR
)
set
(
_stubDir
"
${
CMAKE_CURRENT_BINARY_DIR
}
"
)
set
(
_stubDir
"
${
CMAKE_CURRENT_BINARY_DIR
}
"
)
...
...
paddle/fluid/lite/api/android/jni/paddle_lite_jni.cc
浏览文件 @
0a198e0a
...
@@ -12,13 +12,22 @@ See the License for the specific language governing permissions and
...
@@ -12,13 +12,22 @@ See the License for the specific language governing permissions and
limitations under the License. */
limitations under the License. */
#include "paddle/fluid/lite/api/android/jni/paddle_lite_jni.h"
#include "paddle/fluid/lite/api/android/jni/paddle_lite_jni.h"
#include <memory>
#include <memory>
#include <string>
#include <string>
#include <utility>
#include <utility>
#include <vector>
#include <vector>
#include "paddle/fluid/lite/api/light_api.h"
#include "paddle/fluid/lite/api/paddle_api.h"
#include "paddle/fluid/lite/api/paddle_lite_factory_helper.h"
#include "paddle/fluid/lite/api/paddle_place.h"
#include "paddle/fluid/lite/api/paddle_use_kernels.h"
#include "paddle/fluid/lite/api/paddle_use_ops.h"
#include "paddle/fluid/lite/api/paddle_use_passes.h"
#include "paddle/fluid/lite/kernels/arm/activation_compute.h"
#include "paddle/fluid/lite/kernels/arm/activation_compute.h"
#include "paddle/fluid/lite/kernels/arm/batch_norm_compute.h"
#include "paddle/fluid/lite/kernels/arm/batch_norm_compute.h"
#include "paddle/fluid/lite/kernels/arm/calib_compute.h"
#include "paddle/fluid/lite/kernels/arm/concat_compute.h"
#include "paddle/fluid/lite/kernels/arm/concat_compute.h"
#include "paddle/fluid/lite/kernels/arm/conv_compute.h"
#include "paddle/fluid/lite/kernels/arm/conv_compute.h"
#include "paddle/fluid/lite/kernels/arm/dropout_compute.h"
#include "paddle/fluid/lite/kernels/arm/dropout_compute.h"
...
@@ -31,13 +40,6 @@ limitations under the License. */
...
@@ -31,13 +40,6 @@ limitations under the License. */
#include "paddle/fluid/lite/kernels/arm/split_compute.h"
#include "paddle/fluid/lite/kernels/arm/split_compute.h"
#include "paddle/fluid/lite/kernels/arm/transpose_compute.h"
#include "paddle/fluid/lite/kernels/arm/transpose_compute.h"
#include "paddle/fluid/lite/api/light_api.h"
#include "paddle/fluid/lite/api/paddle_api.h"
#include "paddle/fluid/lite/api/paddle_lite_factory_helper.h"
#include "paddle/fluid/lite/api/paddle_use_kernels.h"
#include "paddle/fluid/lite/api/paddle_use_ops.h"
#include "paddle/fluid/lite/api/paddle_use_passes.h"
#define ARM_KERNEL_POINTER(kernel_class_name__) \
#define ARM_KERNEL_POINTER(kernel_class_name__) \
std::unique_ptr<paddle::lite::kernels::arm::kernel_class_name__> \
std::unique_ptr<paddle::lite::kernels::arm::kernel_class_name__> \
p##kernel_class_name__( \
p##kernel_class_name__( \
...
@@ -47,8 +49,10 @@ limitations under the License. */
...
@@ -47,8 +49,10 @@ limitations under the License. */
extern
"C"
{
extern
"C"
{
#endif
#endif
using
paddle
::
lite_api
::
CxxConfig
;
using
paddle
::
lite_api
::
MobileConfig
;
using
paddle
::
lite_api
::
MobileConfig
;
using
paddle
::
lite_api
::
PaddlePredictor
;
using
paddle
::
lite_api
::
PaddlePredictor
;
using
paddle
::
lite_api
::
Place
;
using
paddle
::
lite_api
::
Tensor
;
using
paddle
::
lite_api
::
Tensor
;
static
std
::
shared_ptr
<
PaddlePredictor
>
predictor
;
static
std
::
shared_ptr
<
PaddlePredictor
>
predictor
;
...
@@ -59,6 +63,8 @@ static std::shared_ptr<PaddlePredictor> predictor;
...
@@ -59,6 +63,8 @@ static std::shared_ptr<PaddlePredictor> predictor;
*/
*/
static
void
use_arm_kernels
()
{
static
void
use_arm_kernels
()
{
ARM_KERNEL_POINTER
(
BatchNormCompute
);
ARM_KERNEL_POINTER
(
BatchNormCompute
);
ARM_KERNEL_POINTER
(
CalibComputeFp32ToInt8
);
ARM_KERNEL_POINTER
(
CalibComputeInt8ToFp32
);
ARM_KERNEL_POINTER
(
ConvCompute
);
ARM_KERNEL_POINTER
(
ConvCompute
);
ARM_KERNEL_POINTER
(
ConcatCompute
);
ARM_KERNEL_POINTER
(
ConcatCompute
);
ARM_KERNEL_POINTER
(
ElementwiseAddCompute
);
ARM_KERNEL_POINTER
(
ElementwiseAddCompute
);
...
@@ -129,6 +135,31 @@ inline std::vector<int64_t> jintarray_to_int64_vector(JNIEnv *env,
...
@@ -129,6 +135,31 @@ inline std::vector<int64_t> jintarray_to_int64_vector(JNIEnv *env,
return
dim_vec
;
return
dim_vec
;
}
}
/**
* Converts Java com.baidu.paddle.lite.Place to c++ paddle::lite_api::Place.
*/
inline
static
Place
jplace_to_cpp_place
(
JNIEnv
*
env
,
jobject
java_place
)
{
jclass
place_jclazz
=
env
->
GetObjectClass
(
java_place
);
jmethodID
target_method
=
env
->
GetMethodID
(
place_jclazz
,
"getTargetInt"
,
"()I"
);
jmethodID
precision_method
=
env
->
GetMethodID
(
place_jclazz
,
"getPrecisionInt"
,
"()I"
);
jmethodID
data_layout_method
=
env
->
GetMethodID
(
place_jclazz
,
"getDataLayoutInt"
,
"()I"
);
jmethodID
device_method
=
env
->
GetMethodID
(
place_jclazz
,
"getDevice"
,
"()I"
);
int
target
=
env
->
CallIntMethod
(
java_place
,
target_method
);
int
precision
=
env
->
CallIntMethod
(
java_place
,
precision_method
);
int
data_layout
=
env
->
CallIntMethod
(
java_place
,
data_layout_method
);
int
device
=
env
->
CallIntMethod
(
java_place
,
device_method
);
return
Place
(
static_cast
<
paddle
::
lite_api
::
TargetType
>
(
target
),
static_cast
<
paddle
::
lite_api
::
PrecisionType
>
(
precision
),
static_cast
<
paddle
::
lite_api
::
DataLayoutType
>
(
data_layout
),
device
);
}
inline
static
int64_t
product
(
const
std
::
vector
<
int64_t
>
&
vec
)
{
inline
static
int64_t
product
(
const
std
::
vector
<
int64_t
>
&
vec
)
{
if
(
vec
.
empty
())
{
if
(
vec
.
empty
())
{
return
0
;
return
0
;
...
@@ -140,6 +171,31 @@ inline static int64_t product(const std::vector<int64_t> &vec) {
...
@@ -140,6 +171,31 @@ inline static int64_t product(const std::vector<int64_t> &vec) {
return
result
;
return
result
;
}
}
JNIEXPORT
jboolean
JNICALL
Java_com_baidu_paddle_lite_PaddlePredictor_loadCxxModel
(
JNIEnv
*
env
,
jclass
thiz
,
jstring
model_path
,
jobject
preferred_place
,
jobjectArray
valid_places
)
{
if
(
predictor
!=
nullptr
)
{
return
JNI_FALSE
;
}
use_arm_kernels
();
int
valid_place_count
=
env
->
GetArrayLength
(
valid_places
);
std
::
vector
<
Place
>
cpp_valid_places
;
for
(
int
i
=
0
;
i
<
valid_place_count
;
++
i
)
{
jobject
jplace
=
env
->
GetObjectArrayElement
(
valid_places
,
i
);
cpp_valid_places
.
push_back
(
jplace_to_cpp_place
(
env
,
jplace
));
}
CxxConfig
config
;
config
.
set_model_dir
(
jstring_to_cpp_string
(
env
,
model_path
));
config
.
set_preferred_place
(
jplace_to_cpp_place
(
env
,
preferred_place
));
config
.
set_valid_places
(
cpp_valid_places
);
predictor
=
paddle
::
lite_api
::
CreatePaddlePredictor
(
config
);
return
predictor
==
nullptr
?
JNI_FALSE
:
JNI_TRUE
;
}
JNIEXPORT
jboolean
JNICALL
JNIEXPORT
jboolean
JNICALL
Java_com_baidu_paddle_lite_PaddlePredictor_loadMobileModel
(
JNIEnv
*
env
,
Java_com_baidu_paddle_lite_PaddlePredictor_loadMobileModel
(
JNIEnv
*
env
,
jclass
thiz
,
jclass
thiz
,
...
@@ -149,9 +205,19 @@ Java_com_baidu_paddle_lite_PaddlePredictor_loadMobileModel(JNIEnv *env,
...
@@ -149,9 +205,19 @@ Java_com_baidu_paddle_lite_PaddlePredictor_loadMobileModel(JNIEnv *env,
}
}
use_arm_kernels
();
use_arm_kernels
();
MobileConfig
config
;
MobileConfig
config
;
std
::
string
model_dir
=
jstring_to_cpp_string
(
env
,
model_path
);
config
.
set_model_dir
(
model_dir
);
config
.
set_model_dir
(
jstring_to_cpp_string
(
env
,
model_path
)
);
predictor
=
paddle
::
lite_api
::
CreatePaddlePredictor
(
config
);
predictor
=
paddle
::
lite_api
::
CreatePaddlePredictor
(
config
);
return
predictor
==
nullptr
?
JNI_FALSE
:
JNI_TRUE
;
}
JNIEXPORT
jboolean
JNICALL
Java_com_baidu_paddle_lite_PaddlePredictor_saveOptimizedModel
(
JNIEnv
*
env
,
jclass
thiz
,
jstring
model_path
)
{
if
(
predictor
==
nullptr
)
{
return
JNI_FALSE
;
}
predictor
->
SaveOptimizedModel
(
jstring_to_cpp_string
(
env
,
model_path
));
return
JNI_TRUE
;
return
JNI_TRUE
;
}
}
...
@@ -167,6 +233,9 @@ Java_com_baidu_paddle_lite_PaddlePredictor_clear(JNIEnv *env, jclass thiz) {
...
@@ -167,6 +233,9 @@ Java_com_baidu_paddle_lite_PaddlePredictor_clear(JNIEnv *env, jclass thiz) {
JNIEXPORT
jboolean
JNICALL
JNIEXPORT
jboolean
JNICALL
Java_com_baidu_paddle_lite_PaddlePredictor_setInput__I_3I_3F
(
Java_com_baidu_paddle_lite_PaddlePredictor_setInput__I_3I_3F
(
JNIEnv
*
env
,
jclass
thiz
,
jint
offset
,
jintArray
dims
,
jfloatArray
buf
)
{
JNIEnv
*
env
,
jclass
thiz
,
jint
offset
,
jintArray
dims
,
jfloatArray
buf
)
{
if
(
predictor
==
nullptr
)
{
return
JNI_FALSE
;
}
std
::
vector
<
int64_t
>
ddim
=
jintarray_to_int64_vector
(
env
,
dims
);
std
::
vector
<
int64_t
>
ddim
=
jintarray_to_int64_vector
(
env
,
dims
);
int
len
=
env
->
GetArrayLength
(
buf
);
int
len
=
env
->
GetArrayLength
(
buf
);
...
@@ -188,6 +257,9 @@ Java_com_baidu_paddle_lite_PaddlePredictor_setInput__I_3I_3F(
...
@@ -188,6 +257,9 @@ Java_com_baidu_paddle_lite_PaddlePredictor_setInput__I_3I_3F(
JNIEXPORT
jboolean
JNICALL
JNIEXPORT
jboolean
JNICALL
Java_com_baidu_paddle_lite_PaddlePredictor_setInput__I_3I_3B
(
Java_com_baidu_paddle_lite_PaddlePredictor_setInput__I_3I_3B
(
JNIEnv
*
env
,
jclass
thiz
,
jint
offset
,
jintArray
dims
,
jbyteArray
buf
)
{
JNIEnv
*
env
,
jclass
thiz
,
jint
offset
,
jintArray
dims
,
jbyteArray
buf
)
{
if
(
predictor
==
nullptr
)
{
return
JNI_FALSE
;
}
std
::
vector
<
int64_t
>
ddim
=
jintarray_to_int64_vector
(
env
,
dims
);
std
::
vector
<
int64_t
>
ddim
=
jintarray_to_int64_vector
(
env
,
dims
);
int
len
=
env
->
GetArrayLength
(
buf
);
int
len
=
env
->
GetArrayLength
(
buf
);
...
@@ -209,6 +281,9 @@ Java_com_baidu_paddle_lite_PaddlePredictor_setInput__I_3I_3B(
...
@@ -209,6 +281,9 @@ Java_com_baidu_paddle_lite_PaddlePredictor_setInput__I_3I_3B(
JNIEXPORT
jboolean
JNICALL
JNIEXPORT
jboolean
JNICALL
Java_com_baidu_paddle_lite_PaddlePredictor_run
(
JNIEnv
*
,
jclass
)
{
Java_com_baidu_paddle_lite_PaddlePredictor_run
(
JNIEnv
*
,
jclass
)
{
if
(
predictor
==
nullptr
)
{
return
JNI_FALSE
;
}
predictor
->
Run
();
predictor
->
Run
();
return
JNI_TRUE
;
return
JNI_TRUE
;
}
}
...
...
paddle/fluid/lite/api/android/jni/paddle_lite_jni.h
浏览文件 @
0a198e0a
...
@@ -21,6 +21,17 @@
...
@@ -21,6 +21,17 @@
#ifdef __cplusplus
#ifdef __cplusplus
extern
"C"
{
extern
"C"
{
#endif
#endif
/*
* Class: com_baidu_paddle_lite_PaddlePredictor
* Method: loadCxxModel
* Signature:
* (Ljava/lang/String;Lcom/baidu/paddle/lite/Place;[Lcom/baidu/paddle/lite/Place;)Z
*/
JNIEXPORT
jboolean
JNICALL
Java_com_baidu_paddle_lite_PaddlePredictor_loadCxxModel
(
JNIEnv
*
,
jclass
,
jstring
,
jobject
,
jobjectArray
);
/*
/*
* Class: com_baidu_paddle_lite_PaddlePredictor
* Class: com_baidu_paddle_lite_PaddlePredictor
* Method: loadMobileModel
* Method: loadMobileModel
...
@@ -30,6 +41,15 @@ JNIEXPORT jboolean JNICALL
...
@@ -30,6 +41,15 @@ JNIEXPORT jboolean JNICALL
Java_com_baidu_paddle_lite_PaddlePredictor_loadMobileModel
(
JNIEnv
*
,
jclass
,
Java_com_baidu_paddle_lite_PaddlePredictor_loadMobileModel
(
JNIEnv
*
,
jclass
,
jstring
);
jstring
);
/*
* Class: com_baidu_paddle_lite_PaddlePredictor
* Method: saveOptimizedModel
* Signature: (Ljava/lang/String;)Z
*/
JNIEXPORT
jboolean
JNICALL
Java_com_baidu_paddle_lite_PaddlePredictor_saveOptimizedModel
(
JNIEnv
*
,
jclass
,
jstring
);
/*
/*
* Class: com_baidu_paddle_lite_PaddlePredictor
* Class: com_baidu_paddle_lite_PaddlePredictor
* Method: clear
* Method: clear
...
...
paddle/fluid/lite/api/android/jni/src/com/baidu/paddle/lite/PaddlePredictor.java
浏览文件 @
0a198e0a
...
@@ -24,6 +24,20 @@ public class PaddlePredictor {
...
@@ -24,6 +24,20 @@ public class PaddlePredictor {
System
.
loadLibrary
(
JNI_LIB_NAME
);
System
.
loadLibrary
(
JNI_LIB_NAME
);
}
}
/**
* Loads mobile cxx model, which is the model before optimizing passes. The cxx
* model allow users to manage hardware place resources. Caller uses a place at
* Java to control Target, DataLayout, Precision, and Device ID. More details
* about the four fields see our Paddle-Mobile document.
*
*
* @param modelPath modelPath model file path
* @param preferredPlace preferred place to run Cxx Model
* @param validPlaces n * 4 int array, valid places to run Cxx Model
* @return true if load successfully
*/
public
static
native
boolean
loadCxxModel
(
String
modelPath
,
Place
preferredPlace
,
Place
[]
validPlaces
);
/**
/**
* Loads mobile lite model, which is the model after optimizing passes.
* Loads mobile lite model, which is the model after optimizing passes.
*
*
...
@@ -32,6 +46,15 @@ public class PaddlePredictor {
...
@@ -32,6 +46,15 @@ public class PaddlePredictor {
*/
*/
public
static
native
boolean
loadMobileModel
(
String
modelPath
);
public
static
native
boolean
loadMobileModel
(
String
modelPath
);
/**
* Saves optimized model, which is the model can be used by
* {@link loadMobileModel}
*
* @param modelPath model file path
* @return true if save successfully
*/
public
static
native
boolean
saveOptimizedModel
(
String
modelPath
);
/**
/**
* Clears the current loaded model.
* Clears the current loaded model.
*
*
...
...
paddle/fluid/lite/api/android/jni/src/com/baidu/paddle/lite/Place.java
0 → 100644
浏览文件 @
0a198e0a
/* Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
package
com.baidu.paddle.lite
;
/**
* Place specifies the execution context of a Kernel or input/output for a
* kernel. It is used to make the analysis of the MIR more clear and accurate.
*/
public
class
Place
{
public
enum
TargetType
{
UNKNOWN
(
0
),
HOST
(
1
),
X86
(
2
),
CUDA
(
3
),
ARM
(
4
),
OPEN_CL
(
5
),
ANY
(
6
);
public
final
int
value
;
private
TargetType
(
int
value
)
{
this
.
value
=
value
;
}
}
public
enum
PrecisionType
{
UNKNOWN
(
0
),
FLOAT
(
1
),
INT8
(
2
),
INT32
(
3
),
ANY
(
4
);
public
final
int
value
;
private
PrecisionType
(
int
value
)
{
this
.
value
=
value
;
}
}
public
enum
DataLayoutType
{
UNKNOWN
(
0
),
NCHW
(
1
),
ANY
(
2
);
public
final
int
value
;
private
DataLayoutType
(
int
value
)
{
this
.
value
=
value
;
}
}
public
TargetType
target
;
public
PrecisionType
precision
;
public
DataLayoutType
layout
;
public
int
device
;
public
Place
()
{
target
=
TargetType
.
UNKNOWN
;
precision
=
PrecisionType
.
UNKNOWN
;
layout
=
DataLayoutType
.
UNKNOWN
;
device
=
0
;
}
public
Place
(
TargetType
target
)
{
this
(
target
,
PrecisionType
.
FLOAT
);
}
public
Place
(
TargetType
target
,
PrecisionType
precision
)
{
this
(
target
,
precision
,
DataLayoutType
.
NCHW
);
}
public
Place
(
TargetType
target
,
PrecisionType
precision
,
DataLayoutType
layout
)
{
this
(
target
,
precision
,
layout
,
0
);
}
public
Place
(
TargetType
target
,
PrecisionType
precision
,
DataLayoutType
layout
,
int
device
)
{
this
.
target
=
target
;
this
.
precision
=
precision
;
this
.
layout
=
layout
;
this
.
device
=
device
;
}
public
boolean
isValid
()
{
return
target
!=
TargetType
.
UNKNOWN
&&
precision
!=
PrecisionType
.
UNKNOWN
&&
layout
!=
DataLayoutType
.
UNKNOWN
;
}
public
int
getTargetInt
()
{
return
target
.
value
;
}
public
int
getPrecisionInt
()
{
return
precision
.
value
;
}
public
int
getDataLayoutInt
()
{
return
layout
.
value
;
}
public
int
getDevice
()
{
return
device
;
}
}
paddle/fluid/lite/api/android/jni/test/com/baidu/paddle/lite/PaddlePredictorTest.java
浏览文件 @
0a198e0a
...
@@ -33,6 +33,7 @@ class PaddlePredictorTest {
...
@@ -33,6 +33,7 @@ class PaddlePredictorTest {
PaddlePredictor
.
run
();
PaddlePredictor
.
run
();
float
[]
output
=
PaddlePredictor
.
getFloatOutput
(
0
);
float
[]
output
=
PaddlePredictor
.
getFloatOutput
(
0
);
assertEquals
(
output
.
length
,
50000
);
assertEquals
(
output
[
0
],
50.2132f
,
1
e
-
3
f
);
assertEquals
(
output
[
0
],
50.2132f
,
1
e
-
3
f
);
assertEquals
(
output
[
1
],
-
28.8729f
,
1
e
-
3
f
);
assertEquals
(
output
[
1
],
-
28.8729f
,
1
e
-
3
f
);
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录