Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
OAID
Tengine
提交
cc225b4e
T
Tengine
项目概览
OAID
/
Tengine
大约 1 年 前同步成功
通知
53
Star
4429
Fork
1032
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
DevOps
流水线
流水线任务
计划
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
T
Tengine
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
DevOps
DevOps
流水线
流水线任务
计划
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
流水线任务
提交
Issue看板
体验新版 GitCode,发现更多精彩内容 >>
未验证
提交
cc225b4e
编写于
11月 05, 2021
作者:
T
teng
提交者:
GitHub
11月 05, 2021
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fix typo (#1202)
* fix typo
上级
64d90a12
变更
16
隐藏空白更改
内联
并排
Showing
16 changed file
with
22 addition
and
22 deletion
+22
-22
source/device/cpu/cpu_device.c
source/device/cpu/cpu_device.c
+1
-1
source/device/cpu/op/concat/concat_ref.c
source/device/cpu/op/concat/concat_ref.c
+1
-1
source/device/cpu/op/deconv/cortex_a/deconv_dw_hcl_arm.c
source/device/cpu/op/deconv/cortex_a/deconv_dw_hcl_arm.c
+1
-1
source/device/cpu/op/eltwise/eltwise_ref.c
source/device/cpu/op/eltwise/eltwise_ref.c
+1
-1
source/device/cpu/op/scatter/scatter_ref.c
source/device/cpu/op/scatter/scatter_ref.c
+1
-1
source/device/cuda/op/nvgpu_clip.cu
source/device/cuda/op/nvgpu_clip.cu
+1
-1
source/device/cuda/op/nvgpu_eltwise.cu
source/device/cuda/op/nvgpu_eltwise.cu
+1
-1
source/device/opencl/op/ocl_slice.cc
source/device/opencl/op/ocl_slice.cc
+1
-1
source/device/opendla/op/odla_deconv.cc
source/device/opendla/op/odla_deconv.cc
+2
-2
source/device/opendla/op/odla_fc.cc
source/device/opendla/op/odla_fc.cc
+3
-3
source/device/tensorrt/op/trt_crop.cc
source/device/tensorrt/op/trt_crop.cc
+1
-1
source/device/tensorrt/op/trt_deconvolution.cc
source/device/tensorrt/op/trt_deconvolution.cc
+2
-2
source/device/tensorrt/op/trt_reduction.cc
source/device/tensorrt/op/trt_reduction.cc
+2
-2
source/device/tensorrt/op/trt_resize.cc
source/device/tensorrt/op/trt_resize.cc
+1
-1
source/device/tensorrt/op/trt_upsample.cc
source/device/tensorrt/op/trt_upsample.cc
+2
-2
source/device/tensorrt/trt_executor.cc
source/device/tensorrt/trt_executor.cc
+1
-1
未找到文件。
source/device/cpu/cpu_device.c
浏览文件 @
cc225b4e
...
...
@@ -288,7 +288,7 @@ static int cpu_describe(struct device* device, struct vector* allowed_ops, struc
}
if
(
NULL
==
blocked_ops
)
{
TLOG_ERR
(
"Error:
Allow
ed op list pointer is NULL
\n
"
);
TLOG_ERR
(
"Error:
Block
ed op list pointer is NULL
\n
"
);
}
for
(
int
i
=
OP_GENERIC
+
1
;
i
<
OP_BUILTIN_LAST
-
1
;
i
++
)
...
...
source/device/cpu/op/concat/concat_ref.c
浏览文件 @
cc225b4e
...
...
@@ -69,7 +69,7 @@ static int run(struct node_ops* node_ops, struct exec_node* exec_node, struct ex
ret
=
ref_concat_int8
(
ir_graph
,
ir_node
,
concat_param
->
axis
);
}
else
TLOG_ERR
(
"
In
put data type %d not to be supported.
\n
"
,
output_tensor
->
data_type
);
TLOG_ERR
(
"
Out
put data type %d not to be supported.
\n
"
,
output_tensor
->
data_type
);
return
ret
;
}
...
...
source/device/cpu/op/deconv/cortex_a/deconv_dw_hcl_arm.c
浏览文件 @
cc225b4e
...
...
@@ -57,7 +57,7 @@ static int run(struct node_ops* node_ops, struct exec_node* exec_node, struct ex
if
(
deconv_dw_run
(
input_tensor
,
weight_tensor
,
bias_tensor
,
output_tensor
,
deconv_param
,
num_thread
,
cpu_affinity
)
<
0
)
{
TLOG_ERR
(
"hcl conv run failed
\n
"
);
TLOG_ERR
(
"hcl
de
conv run failed
\n
"
);
// set_tengine_errno(EFAULT);
return
-
1
;
}
...
...
source/device/cpu/op/eltwise/eltwise_ref.c
浏览文件 @
cc225b4e
...
...
@@ -967,7 +967,7 @@ static int run(struct node_ops* node_ops, struct exec_node* exec_node, struct ex
input_chan_0
,
input_hw_0
,
input1_count4
,
exec_graph
->
num_thread
,
input_hw_1
,
eltwise_param
);
else
{
TLOG_ERR
(
"
In
put data type %d not to be supported.
\n
"
,
output_tensor
->
data_type
);
TLOG_ERR
(
"
Out
put data type %d not to be supported.
\n
"
,
output_tensor
->
data_type
);
return
-
1
;
}
...
...
source/device/cpu/op/scatter/scatter_ref.c
浏览文件 @
cc225b4e
...
...
@@ -98,7 +98,7 @@ static int ref_scatter_fp32(float* input, float* output, int* indices, float* up
{
int
axis
=
op_param
->
axis
;
bool
is_onnx
=
op_param
->
is_onnx
;
TLOG_ERR
(
"
indic
es %f %f
\n
"
,
updates
[
0
],
updates
[
1
]);
TLOG_ERR
(
"
updat
es %f %f
\n
"
,
updates
[
0
],
updates
[
1
]);
TLOG_ERR
(
"indices %d %d
\n
"
,
indices
[
0
],
indices
[
1
]);
int
outSize
=
1
;
for
(
int
i
=
0
;
i
<
op_param
->
dim_size
;
i
++
)
...
...
source/device/cuda/op/nvgpu_clip.cu
浏览文件 @
cc225b4e
...
...
@@ -63,7 +63,7 @@ void relu6_gpu_kernel(struct graph* ir_graph, struct node* ir_node, dict_uint2vo
void
CUDAEngine
::
AddClipNode
(
struct
graph
*
ir_graph
,
struct
node
*
ir_node
)
{
TLOG_INFO
(
"Tengine GPU: Support OP(%d) OP_
RELU
.
\n
"
,
ir_node
->
index
);
TLOG_INFO
(
"Tengine GPU: Support OP(%d) OP_
CLIP
.
\n
"
,
ir_node
->
index
);
relu6_gpu_kernel
(
ir_graph
,
ir_node
,
this
->
gpu_addr_map
);
this
->
ops
.
push_back
(
std
::
bind
(
&
relu6_gpu_kernel
,
ir_graph
,
ir_node
,
this
->
gpu_addr_map
));
}
source/device/cuda/op/nvgpu_eltwise.cu
浏览文件 @
cc225b4e
...
...
@@ -68,7 +68,7 @@ void eltwisesum_gpu_kernel(struct graph* ir_graph, struct node* ir_node, dict_ui
void
CUDAEngine
::
AddEltwiseNode
(
struct
graph
*
ir_graph
,
struct
node
*
ir_node
)
{
TLOG_INFO
(
"Tengine GPU: Support OP(%d) OP_
RELU
.
\n
"
,
ir_node
->
index
);
TLOG_INFO
(
"Tengine GPU: Support OP(%d) OP_
ELTWISE
.
\n
"
,
ir_node
->
index
);
eltwisesum_gpu_kernel
(
ir_graph
,
ir_node
,
this
->
gpu_addr_map
);
this
->
ops
.
push_back
(
std
::
bind
(
&
eltwisesum_gpu_kernel
,
ir_graph
,
ir_node
,
this
->
gpu_addr_map
));
}
source/device/opencl/op/ocl_slice.cc
浏览文件 @
cc225b4e
...
...
@@ -10,7 +10,7 @@ extern "C"
bool
OCLEngine
::
AddSliceNode
(
struct
node
*
ir_node
)
{
TLOG_INFO
(
"Tengine OpenCL: Support OP_
FLATTEN
(%d).
\n
"
,
ir_node
->
index
);
TLOG_INFO
(
"Tengine OpenCL: Support OP_
SLICE
(%d).
\n
"
,
ir_node
->
index
);
struct
graph
*
ir_graph
=
ir_node
->
graph
;
...
...
source/device/opendla/op/odla_deconv.cc
浏览文件 @
cc225b4e
...
...
@@ -96,7 +96,7 @@ nvdla::priv::canonical_ast::Node * ODLAEngine::AddDeconvlutionNode(struct node*
break
;
}
default:
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of conv(id: %d, name: %s).
\n
"
,
conv_weight
->
data_type
,
ir_node
->
index
,
ir_node
->
name
);
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of
de
conv(id: %d, name: %s).
\n
"
,
conv_weight
->
data_type
,
ir_node
->
index
,
ir_node
->
name
);
return
nullptr
;
}
...
...
@@ -138,7 +138,7 @@ nvdla::priv::canonical_ast::Node * ODLAEngine::AddDeconvlutionNode(struct node*
break
;
}
default:
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of conv(id: %d, name: %s).
\n
"
,
conv_bias
->
data_type
,
ir_node
->
index
,
ir_node
->
name
);
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of
de
conv(id: %d, name: %s).
\n
"
,
conv_bias
->
data_type
,
ir_node
->
index
,
ir_node
->
name
);
return
nullptr
;
}
...
...
source/device/opendla/op/odla_fc.cc
浏览文件 @
cc225b4e
...
...
@@ -66,7 +66,7 @@ nvdla::priv::canonical_ast::Node * ODLAEngine::AddFullyConnectionNode(struct nod
{
if
(
weight_tensor
->
quant_param_num
!=
weight_tensor
->
dims
[
0
])
{
fprintf
(
stderr
,
"Tengine: Unsupported weight quant channel of
conv
(id: %d, name: %s).
\n
"
,
ir_node
->
index
,
ir_node
->
name
);
fprintf
(
stderr
,
"Tengine: Unsupported weight quant channel of
fc
(id: %d, name: %s).
\n
"
,
ir_node
->
index
,
ir_node
->
name
);
return
nullptr
;
}
...
...
@@ -104,7 +104,7 @@ nvdla::priv::canonical_ast::Node * ODLAEngine::AddFullyConnectionNode(struct nod
break
;
}
default:
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of
conv
(id: %d, name: %s).
\n
"
,
weight_tensor
->
data_type
,
ir_node
->
index
,
ir_node
->
name
);
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of
fc
(id: %d, name: %s).
\n
"
,
weight_tensor
->
data_type
,
ir_node
->
index
,
ir_node
->
name
);
return
nullptr
;
}
...
...
@@ -147,7 +147,7 @@ nvdla::priv::canonical_ast::Node * ODLAEngine::AddFullyConnectionNode(struct nod
break
;
}
default:
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of
conv
(id: %d, name: %s).
\n
"
,
bias_tensor
->
data_type
,
ir_node
->
index
,
ir_node
->
name
);
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of
fc
(id: %d, name: %s).
\n
"
,
bias_tensor
->
data_type
,
ir_node
->
index
,
ir_node
->
name
);
return
nullptr
;
}
...
...
source/device/tensorrt/op/trt_crop.cc
浏览文件 @
cc225b4e
...
...
@@ -83,7 +83,7 @@ bool TensorRTEngine::AddCropNode(struct graph* ir_graph, struct node* node)
nvinfer1
::
IPaddingLayer
*
layer
=
this
->
network
->
addPadding
(
*
crop_input_tensor
,
prePadding
,
postPadding
);
if
(
nullptr
==
layer
)
{
fprintf
(
stderr
,
"Tengine: Add
Flatten
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
fprintf
(
stderr
,
"Tengine: Add
Crop
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
return
false
;
}
...
...
source/device/tensorrt/op/trt_deconvolution.cc
浏览文件 @
cc225b4e
...
...
@@ -103,7 +103,7 @@ bool TensorRTEngine::AddDeConvolutionNode(struct graph* ir_graph, struct node *n
break
;
}
default:
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of conv(id: %d, name: %s).
\n
"
,
deconv_weight
->
data_type
,
node
->
index
,
node
->
name
);
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of
de
conv(id: %d, name: %s).
\n
"
,
deconv_weight
->
data_type
,
node
->
index
,
node
->
name
);
return
false
;
}
...
...
@@ -143,7 +143,7 @@ bool TensorRTEngine::AddDeConvolutionNode(struct graph* ir_graph, struct node *n
break
;
}
default:
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of conv(id: %d, name: %s).
\n
"
,
fprintf
(
stderr
,
"Tengine: Unsupported weight quant data type(%d) of
de
conv(id: %d, name: %s).
\n
"
,
deconv_bias
->
data_type
,
node
->
index
,
node
->
name
);
return
false
;
}
...
...
source/device/tensorrt/op/trt_reduction.cc
浏览文件 @
cc225b4e
...
...
@@ -37,13 +37,13 @@ bool TensorRTEngine::AddReductionNode(struct graph* ir_graph, struct node* node)
struct
tensor
*
output_tensor
=
get_ir_graph_tensor
(
ir_graph
,
node
->
output_tensors
[
0
]);
if
(
nullptr
==
input_tensor
||
nullptr
==
output_tensor
)
{
fprintf
(
stderr
,
"Tengine: Get input & output for Re
shape
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
fprintf
(
stderr
,
"Tengine: Get input & output for Re
duction
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
return
false
;
}
if
(
!
check_if_input_in_map
(
input_tensor
->
index
,
this
->
tensor_swap_map
))
{
fprintf
(
stderr
,
"Tengine: Query input for Re
shape
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
fprintf
(
stderr
,
"Tengine: Query input for Re
duction
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
return
false
;
}
...
...
source/device/tensorrt/op/trt_resize.cc
浏览文件 @
cc225b4e
...
...
@@ -64,7 +64,7 @@ bool TensorRTEngine::AddResizeNode(struct graph* ir_graph, struct node* node)
nvinfer1
::
IResizeLayer
*
layer
=
this
->
network
->
addResize
(
*
interp_input_tensor
);
if
(
nullptr
==
layer
)
{
fprintf
(
stderr
,
"Tengine: Add
Interp
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
fprintf
(
stderr
,
"Tengine: Add
Resize
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
return
false
;
}
...
...
source/device/tensorrt/op/trt_upsample.cc
浏览文件 @
cc225b4e
...
...
@@ -37,13 +37,13 @@ bool TensorRTEngine::AddUpSampleNode(struct graph* ir_graph, struct node* node)
struct
tensor
*
upsample_output
=
get_ir_graph_tensor
(
ir_graph
,
node
->
output_tensors
[
0
]);
if
(
nullptr
==
upsample_input
||
nullptr
==
upsample_output
)
{
fprintf
(
stderr
,
"Tengine: Get input & output for
Interp
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
fprintf
(
stderr
,
"Tengine: Get input & output for
Upsample
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
return
false
;
}
if
(
!
check_if_input_in_map
(
upsample_input
->
index
,
this
->
tensor_swap_map
))
{
fprintf
(
stderr
,
"Tengine: Query input for
Interp
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
fprintf
(
stderr
,
"Tengine: Query input for
Upsample
(id: %d, name: %s) layer failed.
\n
"
,
node
->
index
,
node
->
name
);
return
false
;
}
...
...
source/device/tensorrt/trt_executor.cc
浏览文件 @
cc225b4e
...
...
@@ -260,7 +260,7 @@ int TensorRTEngine::Build(struct subgraph* subgraph)
{
if
(
!
AddInterpNode
(
ir_graph
,
ir_node
))
{
TLOG_ERR
(
"Tengine: Cannot add
FullyConnected
op(%d).
\n
"
,
ir_node
->
index
);
TLOG_ERR
(
"Tengine: Cannot add
Interp
op(%d).
\n
"
,
ir_node
->
index
);
return
-
6
;
}
break
;
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录