提交 39c67eb2 编写于 作者: H hjchen2

Remove unused op_symbols.h and redundant code in framework.pb c/h file

上级 4d001f09
......@@ -20,20 +20,6 @@ size_t paddle_mobile__framework__proto__version__get_packed_size(
return protobuf_c_message_get_packed_size(
(const ProtobufCMessage *)(message));
}
size_t paddle_mobile__framework__proto__version__pack(
const PaddleMobile__Framework__Proto__Version *message, uint8_t *out) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__version__descriptor);
return protobuf_c_message_pack((const ProtobufCMessage *)message, out);
}
size_t paddle_mobile__framework__proto__version__pack_to_buffer(
const PaddleMobile__Framework__Proto__Version *message,
ProtobufCBuffer *buffer) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__version__descriptor);
return protobuf_c_message_pack_to_buffer((const ProtobufCMessage *)message,
buffer);
}
PaddleMobile__Framework__Proto__Version *
paddle_mobile__framework__proto__version__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -75,20 +61,6 @@ size_t paddle_mobile__framework__proto__op_desc__get_packed_size(
return protobuf_c_message_get_packed_size(
(const ProtobufCMessage *)(message));
}
size_t paddle_mobile__framework__proto__op_desc__pack(
const PaddleMobile__Framework__Proto__OpDesc *message, uint8_t *out) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__op_desc__descriptor);
return protobuf_c_message_pack((const ProtobufCMessage *)message, out);
}
size_t paddle_mobile__framework__proto__op_desc__pack_to_buffer(
const PaddleMobile__Framework__Proto__OpDesc *message,
ProtobufCBuffer *buffer) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__op_desc__descriptor);
return protobuf_c_message_pack_to_buffer((const ProtobufCMessage *)message,
buffer);
}
PaddleMobile__Framework__Proto__OpDesc *
paddle_mobile__framework__proto__op_desc__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -130,20 +102,6 @@ size_t paddle_mobile__framework__proto__op_proto__get_packed_size(
return protobuf_c_message_get_packed_size(
(const ProtobufCMessage *)(message));
}
size_t paddle_mobile__framework__proto__op_proto__pack(
const PaddleMobile__Framework__Proto__OpProto *message, uint8_t *out) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__op_proto__descriptor);
return protobuf_c_message_pack((const ProtobufCMessage *)message, out);
}
size_t paddle_mobile__framework__proto__op_proto__pack_to_buffer(
const PaddleMobile__Framework__Proto__OpProto *message,
ProtobufCBuffer *buffer) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__op_proto__descriptor);
return protobuf_c_message_pack_to_buffer((const ProtobufCMessage *)message,
buffer);
}
PaddleMobile__Framework__Proto__OpProto *
paddle_mobile__framework__proto__op_proto__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -211,20 +169,6 @@ size_t paddle_mobile__framework__proto__var_type__get_packed_size(
return protobuf_c_message_get_packed_size(
(const ProtobufCMessage *)(message));
}
size_t paddle_mobile__framework__proto__var_type__pack(
const PaddleMobile__Framework__Proto__VarType *message, uint8_t *out) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__var_type__descriptor);
return protobuf_c_message_pack((const ProtobufCMessage *)message, out);
}
size_t paddle_mobile__framework__proto__var_type__pack_to_buffer(
const PaddleMobile__Framework__Proto__VarType *message,
ProtobufCBuffer *buffer) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__var_type__descriptor);
return protobuf_c_message_pack_to_buffer((const ProtobufCMessage *)message,
buffer);
}
PaddleMobile__Framework__Proto__VarType *
paddle_mobile__framework__proto__var_type__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -254,20 +198,6 @@ size_t paddle_mobile__framework__proto__var_desc__get_packed_size(
return protobuf_c_message_get_packed_size(
(const ProtobufCMessage *)(message));
}
size_t paddle_mobile__framework__proto__var_desc__pack(
const PaddleMobile__Framework__Proto__VarDesc *message, uint8_t *out) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__var_desc__descriptor);
return protobuf_c_message_pack((const ProtobufCMessage *)message, out);
}
size_t paddle_mobile__framework__proto__var_desc__pack_to_buffer(
const PaddleMobile__Framework__Proto__VarDesc *message,
ProtobufCBuffer *buffer) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__var_desc__descriptor);
return protobuf_c_message_pack_to_buffer((const ProtobufCMessage *)message,
buffer);
}
PaddleMobile__Framework__Proto__VarDesc *
paddle_mobile__framework__proto__var_desc__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -297,20 +227,6 @@ size_t paddle_mobile__framework__proto__block_desc__get_packed_size(
return protobuf_c_message_get_packed_size(
(const ProtobufCMessage *)(message));
}
size_t paddle_mobile__framework__proto__block_desc__pack(
const PaddleMobile__Framework__Proto__BlockDesc *message, uint8_t *out) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__block_desc__descriptor);
return protobuf_c_message_pack((const ProtobufCMessage *)message, out);
}
size_t paddle_mobile__framework__proto__block_desc__pack_to_buffer(
const PaddleMobile__Framework__Proto__BlockDesc *message,
ProtobufCBuffer *buffer) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__block_desc__descriptor);
return protobuf_c_message_pack_to_buffer((const ProtobufCMessage *)message,
buffer);
}
PaddleMobile__Framework__Proto__BlockDesc *
paddle_mobile__framework__proto__block_desc__unpack(
ProtobufCAllocator *allocator, size_t len, const uint8_t *data) {
......@@ -339,20 +255,6 @@ size_t paddle_mobile__framework__proto__program_desc__get_packed_size(
return protobuf_c_message_get_packed_size(
(const ProtobufCMessage *)(message));
}
size_t paddle_mobile__framework__proto__program_desc__pack(
const PaddleMobile__Framework__Proto__ProgramDesc *message, uint8_t *out) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__program_desc__descriptor);
return protobuf_c_message_pack((const ProtobufCMessage *)message, out);
}
size_t paddle_mobile__framework__proto__program_desc__pack_to_buffer(
const PaddleMobile__Framework__Proto__ProgramDesc *message,
ProtobufCBuffer *buffer) {
assert(message->base.descriptor ==
&paddle_mobile__framework__proto__program_desc__descriptor);
return protobuf_c_message_pack_to_buffer((const ProtobufCMessage *)message,
buffer);
}
PaddleMobile__Framework__Proto__ProgramDesc *
paddle_mobile__framework__proto__program_desc__unpack(
ProtobufCAllocator *allocator, size_t len, const uint8_t *data) {
......
......@@ -419,11 +419,6 @@ void paddle_mobile__framework__proto__version__init(
PaddleMobile__Framework__Proto__Version *message);
size_t paddle_mobile__framework__proto__version__get_packed_size(
const PaddleMobile__Framework__Proto__Version *message);
size_t paddle_mobile__framework__proto__version__pack(
const PaddleMobile__Framework__Proto__Version *message, uint8_t *out);
size_t paddle_mobile__framework__proto__version__pack_to_buffer(
const PaddleMobile__Framework__Proto__Version *message,
ProtobufCBuffer *buffer);
PaddleMobile__Framework__Proto__Version *
paddle_mobile__framework__proto__version__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -442,11 +437,6 @@ void paddle_mobile__framework__proto__op_desc__init(
PaddleMobile__Framework__Proto__OpDesc *message);
size_t paddle_mobile__framework__proto__op_desc__get_packed_size(
const PaddleMobile__Framework__Proto__OpDesc *message);
size_t paddle_mobile__framework__proto__op_desc__pack(
const PaddleMobile__Framework__Proto__OpDesc *message, uint8_t *out);
size_t paddle_mobile__framework__proto__op_desc__pack_to_buffer(
const PaddleMobile__Framework__Proto__OpDesc *message,
ProtobufCBuffer *buffer);
PaddleMobile__Framework__Proto__OpDesc *
paddle_mobile__framework__proto__op_desc__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -465,11 +455,6 @@ void paddle_mobile__framework__proto__op_proto__init(
PaddleMobile__Framework__Proto__OpProto *message);
size_t paddle_mobile__framework__proto__op_proto__get_packed_size(
const PaddleMobile__Framework__Proto__OpProto *message);
size_t paddle_mobile__framework__proto__op_proto__pack(
const PaddleMobile__Framework__Proto__OpProto *message, uint8_t *out);
size_t paddle_mobile__framework__proto__op_proto__pack_to_buffer(
const PaddleMobile__Framework__Proto__OpProto *message,
ProtobufCBuffer *buffer);
PaddleMobile__Framework__Proto__OpProto *
paddle_mobile__framework__proto__op_proto__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -500,11 +485,6 @@ void paddle_mobile__framework__proto__var_type__init(
PaddleMobile__Framework__Proto__VarType *message);
size_t paddle_mobile__framework__proto__var_type__get_packed_size(
const PaddleMobile__Framework__Proto__VarType *message);
size_t paddle_mobile__framework__proto__var_type__pack(
const PaddleMobile__Framework__Proto__VarType *message, uint8_t *out);
size_t paddle_mobile__framework__proto__var_type__pack_to_buffer(
const PaddleMobile__Framework__Proto__VarType *message,
ProtobufCBuffer *buffer);
PaddleMobile__Framework__Proto__VarType *
paddle_mobile__framework__proto__var_type__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -517,11 +497,6 @@ void paddle_mobile__framework__proto__var_desc__init(
PaddleMobile__Framework__Proto__VarDesc *message);
size_t paddle_mobile__framework__proto__var_desc__get_packed_size(
const PaddleMobile__Framework__Proto__VarDesc *message);
size_t paddle_mobile__framework__proto__var_desc__pack(
const PaddleMobile__Framework__Proto__VarDesc *message, uint8_t *out);
size_t paddle_mobile__framework__proto__var_desc__pack_to_buffer(
const PaddleMobile__Framework__Proto__VarDesc *message,
ProtobufCBuffer *buffer);
PaddleMobile__Framework__Proto__VarDesc *
paddle_mobile__framework__proto__var_desc__unpack(ProtobufCAllocator *allocator,
size_t len,
......@@ -534,11 +509,6 @@ void paddle_mobile__framework__proto__block_desc__init(
PaddleMobile__Framework__Proto__BlockDesc *message);
size_t paddle_mobile__framework__proto__block_desc__get_packed_size(
const PaddleMobile__Framework__Proto__BlockDesc *message);
size_t paddle_mobile__framework__proto__block_desc__pack(
const PaddleMobile__Framework__Proto__BlockDesc *message, uint8_t *out);
size_t paddle_mobile__framework__proto__block_desc__pack_to_buffer(
const PaddleMobile__Framework__Proto__BlockDesc *message,
ProtobufCBuffer *buffer);
PaddleMobile__Framework__Proto__BlockDesc *
paddle_mobile__framework__proto__block_desc__unpack(
ProtobufCAllocator *allocator, size_t len, const uint8_t *data);
......@@ -550,11 +520,6 @@ void paddle_mobile__framework__proto__program_desc__init(
PaddleMobile__Framework__Proto__ProgramDesc *message);
size_t paddle_mobile__framework__proto__program_desc__get_packed_size(
const PaddleMobile__Framework__Proto__ProgramDesc *message);
size_t paddle_mobile__framework__proto__program_desc__pack(
const PaddleMobile__Framework__Proto__ProgramDesc *message, uint8_t *out);
size_t paddle_mobile__framework__proto__program_desc__pack_to_buffer(
const PaddleMobile__Framework__Proto__ProgramDesc *message,
ProtobufCBuffer *buffer);
PaddleMobile__Framework__Proto__ProgramDesc *
paddle_mobile__framework__proto__program_desc__unpack(
ProtobufCAllocator *allocator, size_t len, const uint8_t *data);
......
......@@ -65,146 +65,146 @@ limitations under the License. */
LOAD_OP(feed)
LOAD_OP(fetch)
#ifdef BATCHNORM_OP
LOAD_OP2(batch_norm, CPU, MALI_GPU) // NOLINT
LOAD_OP2(batch_norm, CPU, MALI_GPU);
#endif
#ifdef BILINEAR_INTERP_OP
LOAD_OP1(bilinear_interp, CPU) // NOLINT
LOAD_OP1(bilinear_interp, CPU);
#endif
#ifdef BOXCODER_OP
LOAD_OP1(box_coder, CPU) // NOLINT
LOAD_OP1(box_coder, CPU);
#endif
#ifdef CONCAT_OP
LOAD_OP3(concat, CPU, MALI_GPU, FPGA) // NOLINT
LOAD_OP3(concat, CPU, MALI_GPU, FPGA);
#endif
#ifdef CONV_OP
LOAD_OP3(conv2d, CPU, MALI_GPU, FPGA) // NOLINT
LOAD_OP3(conv2d, CPU, MALI_GPU, FPGA);
#endif
#ifdef LRN_OP
LOAD_OP2(lrn, CPU, MALI_GPU) // NOLINT
LOAD_OP2(lrn, CPU, MALI_GPU);
#endif
#ifdef SIGMOID_OP
LOAD_OP1(sigmoid, CPU) // NOLINT
LOAD_OP1(sigmoid, CPU);
#endif
#ifdef FUSION_FC_RELU_OP
LOAD_OP3(fusion_fc_relu, CPU, MALI_GPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_fc_relu) // NOLINT
LOAD_OP3(fusion_fc_relu, CPU, MALI_GPU, FPGA);
LOAD_FUSION_MATCHER(fusion_fc_relu);
#endif
#ifdef FUSION_ELEMENTWISEADDRELU_OP
LOAD_OP3(fusion_elementwise_add_relu, CPU, MALI_GPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_elementwise_add_relu) // NOLINT
LOAD_OP3(fusion_elementwise_add_relu, CPU, MALI_GPU, FPGA);
LOAD_FUSION_MATCHER(fusion_elementwise_add_relu);
#endif
#ifdef SPLIT_OP
LOAD_OP1(split, CPU) // NOLINT
LOAD_OP1(split, CPU);
#endif
#ifdef RESIZE_OP
LOAD_OP2(resize, CPU, MALI_GPU) // NOLINT
LOAD_OP2(resize, CPU, MALI_GPU);
#endif
#ifdef FUSION_CONVADDBNRELU_OP
LOAD_OP2(fusion_conv_add_bn_relu, CPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_conv_add_bn_relu) // NOLINT
LOAD_OP2(fusion_conv_add_bn_relu, CPU, FPGA);
LOAD_FUSION_MATCHER(fusion_conv_add_bn_relu);
#endif
#ifdef RESHAPE_OP
LOAD_OP2(reshape, CPU, MALI_GPU) // NOLINT
LOAD_OP2(reshape, CPU, MALI_GPU);
#endif
#ifdef TRANSPOSE_OP
LOAD_OP1(transpose, CPU) // NOLINT
LOAD_OP1(transpose, CPU);
#endif
#ifdef PRIORBOX_OP
LOAD_OP1(prior_box, CPU) // NOLINT
LOAD_OP1(prior_box, CPU);
#endif
#ifdef FUSION_CONVADDRELU_OP
LOAD_OP2(fusion_conv_add_relu, CPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_conv_add_relu) // NOLINT
LOAD_OP2(fusion_conv_add_relu, CPU, FPGA);
LOAD_FUSION_MATCHER(fusion_conv_add_relu);
#endif
#ifdef FUSION_CONVADDADDPRELU_OP
LOAD_OP2(fusion_conv_add_add_prelu, CPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_conv_add_add_prelu) // NOLINT
LOAD_OP2(fusion_conv_add_add_prelu, CPU, FPGA);
LOAD_FUSION_MATCHER(fusion_conv_add_add_prelu);
#endif
#ifdef FUSION_CONVADD_OP
LOAD_OP2(fusion_conv_add, CPU, MALI_GPU) // NOLINT
LOAD_FUSION_MATCHER(fusion_conv_add) // NOLINT
LOAD_OP2(fusion_conv_add, CPU, MALI_GPU);
LOAD_FUSION_MATCHER(fusion_conv_add);
#endif
#ifdef SOFTMAX_OP
LOAD_OP2(softmax, CPU, MALI_GPU) // NOLINT
LOAD_OP2(softmax, CPU, MALI_GPU);
#endif
#ifdef SHAPE_OP
LOAD_OP1(shape, CPU) // NOLINT
LOAD_OP1(shape, CPU);
#endif
#ifdef DEPTHWISECONV_OP
LOAD_OP1(depthwise_conv2d, CPU) // NOLINT
LOAD_OP1(depthwise_conv2d, CPU);
#endif
#ifdef CONV_TRANSPOSE_OP
LOAD_OP1(conv2d_transpose, CPU) // NOLINT
LOAD_OP1(conv2d_transpose, CPU);
#endif
#ifdef SCALE_OP
LOAD_OP2(scale, CPU, MALI_GPU) // NOLINT
LOAD_OP2(scale, CPU, MALI_GPU);
#endif
#ifdef ELEMENTWISEADD_OP
LOAD_OP2(elementwise_add, CPU, MALI_GPU) // NOLINT
LOAD_OP2(elementwise_add, CPU, MALI_GPU);
#endif
#ifdef PRELU_OP
LOAD_OP2(prelu, CPU, MALI_GPU) // NOLINT
LOAD_OP2(prelu, CPU, MALI_GPU);
#endif
#ifdef FLATTEN_OP
LOAD_OP1(flatten, CPU) // NOLINT
LOAD_OP1(flatten, CPU);
#endif
#ifdef FUSION_CONVBNADDRELU_OP
LOAD_OP2(fusion_conv_bn_add_relu, CPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_conv_bn_add_relu) // NOLINT
LOAD_OP2(fusion_conv_bn_add_relu, CPU, FPGA);
LOAD_FUSION_MATCHER(fusion_conv_bn_add_relu);
#endif
#ifdef FUSION_CONVBNRELU_OP
LOAD_OP2(fusion_conv_bn_relu, CPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_conv_bn_relu) // NOLINT
LOAD_OP2(fusion_conv_bn_relu, CPU, FPGA);
LOAD_FUSION_MATCHER(fusion_conv_bn_relu);
#endif
#ifdef GRU_OP
LOAD_OP1(gru, CPU) // NOLINT
LOAD_OP1(gru, CPU);
#endif
#ifdef FUSION_CONVADDBN_OP
LOAD_OP2(fusion_conv_add_bn, CPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_conv_add_bn) // NOLINT
LOAD_OP2(fusion_conv_add_bn, CPU, FPGA);
LOAD_FUSION_MATCHER(fusion_conv_add_bn);
#endif
#ifdef DROPOUT_OP
LOAD_OP2(dropout, CPU, FPGA) // NOLINT
LOAD_OP2(dropout, CPU, FPGA);
#endif
#ifdef FUSION_CONVADDPRELU_OP
LOAD_OP2(fusion_conv_add_prelu, CPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_conv_add_prelu) // NOLINT
LOAD_OP2(fusion_conv_add_prelu, CPU, FPGA);
LOAD_FUSION_MATCHER(fusion_conv_add_prelu);
#endif
#ifdef FUSION_DWCONVBNRELU_OP
LOAD_OP1(fusion_dwconv_bn_relu, CPU) // NOLINT
LOAD_FUSION_MATCHER(fusion_dwconv_bn_relu) // NOLINT
LOAD_OP1(fusion_dwconv_bn_relu, CPU);
LOAD_FUSION_MATCHER(fusion_dwconv_bn_relu);
#endif
#ifdef CRF_OP
LOAD_OP1(crf_decoding, CPU) // NOLINT
LOAD_OP1(crf_decoding, CPU);
#endif
#ifdef MUL_OP
LOAD_OP2(mul, CPU, MALI_GPU) // NOLINT
LOAD_OP2(mul, CPU, MALI_GPU);
#endif
#ifdef RELU_OP
LOAD_OP2(relu, CPU, MALI_GPU) // NOLINT
LOAD_OP2(relu, CPU, MALI_GPU);
#endif
#ifdef IM2SEQUENCE_OP
LOAD_OP1(im2sequence, CPU) // NOLINT
LOAD_OP1(im2sequence, CPU);
#endif
#ifdef LOOKUP_OP
LOAD_OP1(lookup_table, CPU) // NOLINT
LOAD_OP1(lookup_table, CPU);
#endif
#ifdef FUSION_FC_OP
LOAD_OP3(fusion_fc, CPU, MALI_GPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_fc) // NOLINT
LOAD_OP3(fusion_fc, CPU, MALI_GPU, FPGA);
LOAD_FUSION_MATCHER(fusion_fc);
#endif
#ifdef POOL_OP
LOAD_OP3(pool2d, CPU, MALI_GPU, FPGA) // NOLINT
LOAD_OP3(pool2d, CPU, MALI_GPU, FPGA);
#endif
#ifdef MULTICLASSNMS_OP
LOAD_OP1(multiclass_nms, CPU) // NOLINT
LOAD_OP1(multiclass_nms, CPU);
#endif
#ifdef SLICE_OP
LOAD_OP2(slice, CPU, MALI_GPU) // NOLINT
LOAD_OP2(slice, CPU, MALI_GPU);
#endif
#ifdef FUSION_CONVBN_OP
LOAD_OP2(fusion_conv_bn, CPU, FPGA) // NOLINT
LOAD_FUSION_MATCHER(fusion_conv_bn) // NOLINT
LOAD_OP2(fusion_conv_bn, CPU, FPGA);
LOAD_FUSION_MATCHER(fusion_conv_bn);
#endif
LOAD_OP1(quantize, CPU) // NOLINT
LOAD_OP1(dequantize, CPU) // NOLINT
LOAD_OP1(quantize, CPU);
LOAD_OP1(dequantize, CPU);
......@@ -13,7 +13,6 @@
limitations under the License. */
#import "PaddleMobileCPU.h"
#import "op_symbols.h"
#import "framework/load_ops.h"
#import "framework/tensor.h"
#import "io/paddle_mobile.h"
......
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#pragma once
#include "operators/batchnorm_op.h"
#include "operators/bilinear_interp_op.h"
#include "operators/box_coder_op.h"
#include "operators/concat_op.h"
#include "operators/conv_op.h"
#include "operators/conv_transpose_op.h"
#include "operators/crf_op.h"
#include "operators/depthwise_conv_op.h"
#include "operators/dequantize_op.h"
#include "operators/dropout_op.h"
#include "operators/elementwise_add_op.h"
#include "operators/feed_op.h"
#include "operators/fetch_op.h"
#include "operators/flatten_op.h"
#include "operators/fusion_conv_add_add_prelu_op.h"
#include "operators/fusion_conv_add_bn_op.h"
#include "operators/fusion_conv_add_bn_relu_op.h"
#include "operators/fusion_conv_add_op.h"
#include "operators/fusion_conv_add_prelu_op.h"
#include "operators/fusion_conv_add_relu_op.h"
#include "operators/fusion_conv_bn_add_relu_op.h"
#include "operators/fusion_conv_bn_relu_op.h"
#include "operators/fusion_dwconv_bn_relu_op.h"
#include "operators/fusion_elementwise_add_relu_op.h"
#include "operators/fusion_fc_op.h"
#include "operators/fusion_fc_relu_op.h"
#include "operators/gru_op.h"
#include "operators/im2sequence_op.h"
#include "operators/lookup_op.h"
#include "operators/lrn_op.h"
#include "operators/mul_op.h"
#include "operators/multiclass_nms_op.h"
#include "operators/pool_op.h"
#include "operators/prelu_op.h"
#include "operators/prior_box_op.h"
#include "operators/quantize_op.h"
#include "operators/relu_op.h"
#include "operators/reshape_op.h"
#include "operators/resize_op.h"
#include "operators/scale_op.h"
#include "operators/shape_op.h"
#include "operators/sigmoid_op.h"
#include "operators/slice_op.h"
#include "operators/softmax_op.h"
#include "operators/split_op.h"
#include "operators/transpose_op.h"
......@@ -13,6 +13,11 @@ See the License for the specific language governing permissions and
limitations under the License. */
#ifdef FLATTEN_OP
#ifndef RESHAPE_OP
#define RESHAPE_OP
#endif
#pragma once
#include <operators/kernel/reshape_kernel.h>
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册