export_model.py 3.5 KB
Newer Older
Q
qingqing01 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17
# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
M
Manuel Garcia 已提交
18 19 20 21

import os
import sys

Q
qingqing01 已提交
22 23 24 25 26
# add python path of PadleDetection to sys.path
parent_path = os.path.abspath(os.path.join(__file__, *(['..'] * 2)))
if parent_path not in sys.path:
    sys.path.append(parent_path)

G
Guanghua Yu 已提交
27 28 29 30
# ignore warning log
import warnings
warnings.filterwarnings('ignore')

Q
qingqing01 已提交
31 32
import paddle

K
Kaipeng Deng 已提交
33
from ppdet.core.workspace import load_config, merge_config
Q
qingqing01 已提交
34 35
from ppdet.utils.check import check_gpu, check_version, check_config
from ppdet.utils.cli import ArgsParser
36
from ppdet.engine import Trainer
37
from ppdet.slim import build_slim_model
Q
qingqing01 已提交
38 39

from ppdet.utils.logger import setup_logger
K
Kaipeng Deng 已提交
40
logger = setup_logger('export_model')
Q
qingqing01 已提交
41 42 43 44 45 46 47 48 49


def parse_args():
    parser = ArgsParser()
    parser.add_argument(
        "--output_dir",
        type=str,
        default="output_inference",
        help="Directory for storing the output model files.")
C
cnn 已提交
50 51 52 53 54
    parser.add_argument(
        "--export_serving_model",
        type=bool,
        default=False,
        help="Whether to export serving model or not.")
55 56 57 58 59
    parser.add_argument(
        "--slim_config",
        default=None,
        type=str,
        help="Configuration file of slim method.")
Q
qingqing01 已提交
60 61 62 63 64
    args = parser.parse_args()
    return args


def run(FLAGS, cfg):
K
Kaipeng Deng 已提交
65
    # build detector
66
    trainer = Trainer(cfg, mode='test')
Q
qingqing01 已提交
67

K
Kaipeng Deng 已提交
68
    # load weights
69 70 71 72 73 74 75
    if cfg.architecture in ['DeepSORT']:
        if cfg.det_weights != 'None':
            trainer.load_weights_sde(cfg.det_weights, cfg.reid_weights)
        else:
            trainer.load_weights_sde(None, cfg.reid_weights)
    else:
        trainer.load_weights(cfg.weights)
Q
qingqing01 已提交
76

K
Kaipeng Deng 已提交
77
    # export model
78
    trainer.export(FLAGS.output_dir)
Q
qingqing01 已提交
79

C
cnn 已提交
80 81 82 83 84 85 86 87 88 89 90 91 92
    if FLAGS.export_serving_model:
        from paddle_serving_client.io import inference_model_to_serving
        model_name = os.path.splitext(os.path.split(cfg.filename)[-1])[0]

        inference_model_to_serving(
            dirname="{}/{}".format(FLAGS.output_dir, model_name),
            serving_server="{}/{}/serving_server".format(FLAGS.output_dir,
                                                         model_name),
            serving_client="{}/{}/serving_client".format(FLAGS.output_dir,
                                                         model_name),
            model_filename="model.pdmodel",
            params_filename="model.pdiparams")

Q
qingqing01 已提交
93 94 95 96 97 98

def main():
    paddle.set_device("cpu")
    FLAGS = parse_args()
    cfg = load_config(FLAGS.config)
    # TODO: to be refined in the future
99
    if 'norm_type' in cfg and cfg['norm_type'] == 'sync_bn':
Q
qingqing01 已提交
100 101
        FLAGS.opt['norm_type'] = 'bn'
    merge_config(FLAGS.opt)
102

103
    if FLAGS.slim_config:
104 105
        cfg = build_slim_model(cfg, FLAGS.slim_config, mode='test')

S
shangliang Xu 已提交
106 107
    # FIXME: Temporarily solve the priority problem of FLAGS.opt
    merge_config(FLAGS.opt)
Q
qingqing01 已提交
108 109 110 111 112 113 114 115 116
    check_config(cfg)
    check_gpu(cfg.use_gpu)
    check_version()

    run(FLAGS, cfg)


if __name__ == '__main__':
    main()