未验证 提交 b1186bc8 编写于 作者: G Guanghua Yu 提交者: GitHub

update picodet model zoo (#4117)

上级 339fe716
...@@ -28,8 +28,8 @@ Optimizing method of we use: ...@@ -28,8 +28,8 @@ Optimizing method of we use:
| Backbone | Input size | lr schedule | Box AP(0.5:0.95) | Box AP(0.5) | FLOPS | Model Size | Inference Time | download | config | | Backbone | Input size | lr schedule | Box AP(0.5:0.95) | Box AP(0.5) | FLOPS | Model Size | Inference Time | download | config |
| :------------------------ | :-------: | :-------: | :------: | :---: | :---: | :---: | :------------: | :-------------------------------------------------: | :-----: | | :------------------------ | :-------: | :-------: | :------: | :---: | :---: | :---: | :------------: | :-------------------------------------------------: | :-----: |
| ShuffleNetv2-1x | 320*320 | 280e | 22.3 | 36.8 | -- | 3.8M | -- | [model](https://paddledet.bj.bcebos.com/models/picodet_s_shufflenetv2_320_coco.pdparams) | [log](https://paddledet.bj.bcebos.com/logs/train_picodet_s_shufflenetv2_320_coco.log) | [config](https://github.com/PaddlePaddle/PaddleDetection/tree/develop/configs/picodet/picodet_s_shufflenetv2_320_coco.yml) | | ShuffleNetv2-1x | 320*320 | 280e | 22.8 | 37.7 | -- | 3.8M | -- | [model](https://paddledet.bj.bcebos.com/models/picodet_s_shufflenetv2_320_coco.pdparams) | [log](https://paddledet.bj.bcebos.com/logs/train_picodet_s_shufflenetv2_320_coco.log) | [config](https://github.com/PaddlePaddle/PaddleDetection/tree/develop/configs/picodet/picodet_s_shufflenetv2_320_coco.yml) |
| ShuffleNetv2-1x | 416*416 | 280e | 24.6 | 44.3 | -- | 3.8M | -- | [model](https://paddledet.bj.bcebos.com/models/picodet_s_shufflenetv2_416_coco.pdparams) | [log](https://paddledet.bj.bcebos.com/logs/train_picodet_s_shufflenetv2_416_coco.log) | [config](https://github.com/PaddlePaddle/PaddleDetection/tree/develop/configs/picodet/picodet_s_shufflenetv2_416_coco.yml) | | ShuffleNetv2-1x | 416*416 | 280e | 25.3 | 41.1 | -- | 3.8M | -- | [model](https://paddledet.bj.bcebos.com/models/picodet_s_shufflenetv2_416_coco.pdparams) | [log](https://paddledet.bj.bcebos.com/logs/train_picodet_s_shufflenetv2_416_coco.log) | [config](https://github.com/PaddlePaddle/PaddleDetection/tree/develop/configs/picodet/picodet_s_shufflenetv2_416_coco.yml) |
### PicoDet-M ### PicoDet-M
......
epoch: 280 epoch: 300
LearningRate: LearningRate:
base_lr: 0.4 base_lr: 0.4
schedulers: schedulers:
- !CosineDecay - !CosineDecay
max_epochs: 280 max_epochs: 300
- !LinearWarmup - !LinearWarmup
start_factor: 0.1 start_factor: 0.1
steps: 300 steps: 300
...@@ -14,5 +14,5 @@ OptimizerBuilder: ...@@ -14,5 +14,5 @@ OptimizerBuilder:
momentum: 0.9 momentum: 0.9
type: Momentum type: Momentum
regularizer: regularizer:
factor: 0.0001 factor: 0.00004
type: L2 type: L2
worker_num: 6 worker_num: 8
TrainReader: TrainReader:
sample_transforms: sample_transforms:
- Decode: {} - Decode: {}
- RandomCrop: {} - RandomCrop: {}
- RandomFlip: {prob: 0.5} - RandomFlip: {prob: 0.5}
- Resize: {target_size: [320, 320], keep_ratio: False, interp: 1}
- NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]}
- RandomDistort: {} - RandomDistort: {}
- Permute: {}
batch_transforms: batch_transforms:
- PadBatch: {pad_to_stride: 32} - BatchRandomResize: {target_size: [256, 288, 320, 352, 384], random_size: True, random_interp: True, keep_ratio: False}
- NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]}
- Permute: {}
- Gt2GFLTarget: - Gt2GFLTarget:
downsample_ratios: [8, 16, 32] downsample_ratios: [8, 16, 32]
grid_cell_scale: 5 grid_cell_scale: 5
...@@ -22,7 +21,7 @@ TrainReader: ...@@ -22,7 +21,7 @@ TrainReader:
EvalReader: EvalReader:
sample_transforms: sample_transforms:
- Decode: {} - Decode: {}
- Resize: {interp: 1, target_size: [320, 320], keep_ratio: False} - Resize: {interp: 2, target_size: [320, 320], keep_ratio: False}
- NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]} - NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]}
- Permute: {} - Permute: {}
batch_transforms: batch_transforms:
...@@ -36,7 +35,7 @@ TestReader: ...@@ -36,7 +35,7 @@ TestReader:
image_shape: [3, 320, 320] image_shape: [3, 320, 320]
sample_transforms: sample_transforms:
- Decode: {} - Decode: {}
- Resize: {interp: 1, target_size: [320, 320], keep_ratio: False} - Resize: {interp: 2, target_size: [320, 320], keep_ratio: False}
- NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]} - NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]}
- Permute: {} - Permute: {}
batch_transforms: batch_transforms:
......
...@@ -4,17 +4,16 @@ TrainReader: ...@@ -4,17 +4,16 @@ TrainReader:
- Decode: {} - Decode: {}
- RandomCrop: {} - RandomCrop: {}
- RandomFlip: {prob: 0.5} - RandomFlip: {prob: 0.5}
- Resize: {target_size: [416, 416], keep_ratio: False, interp: 1}
- NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]}
- RandomDistort: {} - RandomDistort: {}
- Permute: {}
batch_transforms: batch_transforms:
- PadBatch: {pad_to_stride: 32} - BatchRandomResize: {target_size: [352, 384, 416, 448, 480], random_size: True, random_interp: True, keep_ratio: False}
- NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]}
- Permute: {}
- Gt2GFLTarget: - Gt2GFLTarget:
downsample_ratios: [8, 16, 32] downsample_ratios: [8, 16, 32]
grid_cell_scale: 5 grid_cell_scale: 5
cell_offset: 0.5 cell_offset: 0.5
batch_size: 96 batch_size: 80
shuffle: true shuffle: true
drop_last: true drop_last: true
...@@ -22,7 +21,7 @@ TrainReader: ...@@ -22,7 +21,7 @@ TrainReader:
EvalReader: EvalReader:
sample_transforms: sample_transforms:
- Decode: {} - Decode: {}
- Resize: {interp: 1, target_size: [416, 416], keep_ratio: False} - Resize: {interp: 2, target_size: [416, 416], keep_ratio: False}
- NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]} - NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]}
- Permute: {} - Permute: {}
batch_transforms: batch_transforms:
...@@ -36,7 +35,7 @@ TestReader: ...@@ -36,7 +35,7 @@ TestReader:
image_shape: [3, 416, 416] image_shape: [3, 416, 416]
sample_transforms: sample_transforms:
- Decode: {} - Decode: {}
- Resize: {interp: 1, target_size: [416, 416], keep_ratio: False} - Resize: {interp: 2, target_size: [416, 416], keep_ratio: False}
- NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]} - NormalizeImage: {is_scale: true, mean: [0.485,0.456,0.406], std: [0.229, 0.224,0.225]}
- Permute: {} - Permute: {}
batch_transforms: batch_transforms:
......
...@@ -2,11 +2,11 @@ _BASE_: [ ...@@ -2,11 +2,11 @@ _BASE_: [
'../datasets/coco_detection.yml', '../datasets/coco_detection.yml',
'../runtime.yml', '../runtime.yml',
'_base_/picodet_mbv3_0_5x.yml', '_base_/picodet_mbv3_0_5x.yml',
'_base_/optimizer_280e.yml', '_base_/optimizer_300e.yml',
'_base_/picodet_320_reader.yml', '_base_/picodet_320_reader.yml',
] ]
weights: output/picodet_m_r18_320_coco/model_final weights: output/picodet_l_r18_320_coco/model_final
pretrain_weights: https://paddledet.bj.bcebos.com/models/pretrained/ResNet18_vd_pretrained.pdparams pretrain_weights: https://paddledet.bj.bcebos.com/models/pretrained/ResNet18_vd_pretrained.pdparams
find_unused_parameters: True find_unused_parameters: True
use_ema: true use_ema: true
......
...@@ -2,7 +2,7 @@ _BASE_: [ ...@@ -2,7 +2,7 @@ _BASE_: [
'../datasets/coco_detection.yml', '../datasets/coco_detection.yml',
'../runtime.yml', '../runtime.yml',
'_base_/picodet_mobilenetv3.yml', '_base_/picodet_mobilenetv3.yml',
'_base_/optimizer_280e.yml', '_base_/optimizer_300e.yml',
'_base_/picodet_320_reader.yml', '_base_/picodet_320_reader.yml',
] ]
......
...@@ -2,7 +2,7 @@ _BASE_: [ ...@@ -2,7 +2,7 @@ _BASE_: [
'../datasets/coco_detection.yml', '../datasets/coco_detection.yml',
'../runtime.yml', '../runtime.yml',
'_base_/picodet_mobilenetv3.yml', '_base_/picodet_mobilenetv3.yml',
'_base_/optimizer_280e.yml', '_base_/optimizer_300e.yml',
'_base_/picodet_416_reader.yml', '_base_/picodet_416_reader.yml',
] ]
......
...@@ -2,7 +2,7 @@ _BASE_: [ ...@@ -2,7 +2,7 @@ _BASE_: [
'../datasets/coco_detection.yml', '../datasets/coco_detection.yml',
'../runtime.yml', '../runtime.yml',
'_base_/picodet_shufflenetv2_1x.yml', '_base_/picodet_shufflenetv2_1x.yml',
'_base_/optimizer_280e.yml', '_base_/optimizer_300e.yml',
'_base_/picodet_320_reader.yml', '_base_/picodet_320_reader.yml',
] ]
......
...@@ -2,7 +2,7 @@ _BASE_: [ ...@@ -2,7 +2,7 @@ _BASE_: [
'../datasets/coco_detection.yml', '../datasets/coco_detection.yml',
'../runtime.yml', '../runtime.yml',
'_base_/picodet_shufflenetv2_1x.yml', '_base_/picodet_shufflenetv2_1x.yml',
'_base_/optimizer_280e.yml', '_base_/optimizer_300e.yml',
'_base_/picodet_416_reader.yml', '_base_/picodet_416_reader.yml',
] ]
......
_BASE_: [
'../datasets/coco_detection.yml',
'../runtime.yml',
'_base_/picodet_shufflenetv2_1x.yml',
'_base_/optimizer_300e.yml',
'_base_/picodet_320_reader.yml',
]
pretrain_weights: https://paddledet.bj.bcebos.com/models/pretrained/LCNet_x1_0_pretrained.pdparams
weights: output/picodet_s_lcnet_320_coco/model_final
find_unused_parameters: True
use_ema: true
cycle_epoch: 40
snapshot_epoch: 10
PicoDet:
backbone: LCNet
neck: PAN
head: PicoHead
LCNet:
scale: 1.0
feature_maps: [3, 4, 5]
_BASE_: [
'../datasets/coco_detection.yml',
'../runtime.yml',
'_base_/picodet_shufflenetv2_1x.yml',
'_base_/optimizer_300e.yml',
'_base_/picodet_416_reader.yml',
]
pretrain_weights: https://paddledet.bj.bcebos.com/models/pretrained/LCNet_x1_0_pretrained.pdparams
weights: output/picodet_s_lcnet_416_coco/model_final
find_unused_parameters: True
use_ema: true
cycle_epoch: 40
snapshot_epoch: 10
PicoDet:
backbone: LCNet
neck: PAN
head: PicoHead
LCNet:
scale: 1.0
feature_maps: [3, 4, 5]
...@@ -2,7 +2,7 @@ _BASE_: [ ...@@ -2,7 +2,7 @@ _BASE_: [
'../datasets/coco_detection.yml', '../datasets/coco_detection.yml',
'../runtime.yml', '../runtime.yml',
'_base_/picodet_shufflenetv2_1x.yml', '_base_/picodet_shufflenetv2_1x.yml',
'_base_/optimizer_280e.yml', '_base_/optimizer_300e.yml',
'_base_/picodet_320_reader.yml', '_base_/picodet_320_reader.yml',
] ]
......
...@@ -2,7 +2,7 @@ _BASE_: [ ...@@ -2,7 +2,7 @@ _BASE_: [
'../datasets/coco_detection.yml', '../datasets/coco_detection.yml',
'../runtime.yml', '../runtime.yml',
'_base_/picodet_shufflenetv2_1x.yml', '_base_/picodet_shufflenetv2_1x.yml',
'_base_/optimizer_280e.yml', '_base_/optimizer_300e.yml',
'_base_/picodet_416_reader.yml', '_base_/picodet_416_reader.yml',
] ]
......
_BASE_: [
'../datasets/coco_detection.yml',
'../runtime.yml',
'_base_/picodet_shufflenetv2_1x.yml',
'_base_/optimizer_280e.yml',
'_base_/picodet_320_reader.yml',
]
pretrain_weights: https://paddledet.bj.bcebos.com/models/pretrained/LCNet_x0_25_pretrained.pdparams
weights: output/picodet_s_shufflenetv2_320_coco/model_final
find_unused_parameters: True
use_ema: true
cycle_epoch: 40
snapshot_epoch: 10
PicoDet:
backbone: LCNet
neck: PAN
head: PicoHead
LCNet:
scale: 0.25
feature_maps: [3, 4, 5]
...@@ -178,8 +178,6 @@ class ATSSAssigner(object): ...@@ -178,8 +178,6 @@ class ATSSAssigner(object):
""" """
bboxes = bboxes[:, :4] bboxes = bboxes[:, :4]
num_gt, num_bboxes = gt_bboxes.shape[0], bboxes.shape[0] num_gt, num_bboxes = gt_bboxes.shape[0], bboxes.shape[0]
# compute iou between all bbox and gt
overlaps = bbox_overlaps(bboxes, gt_bboxes)
# assign 0 by default # assign 0 by default
assigned_gt_inds = np.zeros((num_bboxes, ), dtype=np.int64) assigned_gt_inds = np.zeros((num_bboxes, ), dtype=np.int64)
...@@ -194,8 +192,10 @@ class ATSSAssigner(object): ...@@ -194,8 +192,10 @@ class ATSSAssigner(object):
assigned_labels = None assigned_labels = None
else: else:
assigned_labels = -np.ones((num_bboxes, ), dtype=np.int64) assigned_labels = -np.ones((num_bboxes, ), dtype=np.int64)
return assigned_gt_inds, max_overlaps, assigned_labels return assigned_gt_inds, max_overlaps
# compute iou between all bbox and gt
overlaps = bbox_overlaps(bboxes, gt_bboxes)
# compute center distance between all bbox and gt # compute center distance between all bbox and gt
gt_cx = (gt_bboxes[:, 0] + gt_bboxes[:, 2]) / 2.0 gt_cx = (gt_bboxes[:, 0] + gt_bboxes[:, 2]) / 2.0
gt_cy = (gt_bboxes[:, 1] + gt_bboxes[:, 3]) / 2.0 gt_cy = (gt_bboxes[:, 1] + gt_bboxes[:, 3]) / 2.0
......
...@@ -26,6 +26,7 @@ from . import res2net ...@@ -26,6 +26,7 @@ from . import res2net
from . import dla from . import dla
from . import shufflenet_v2 from . import shufflenet_v2
from . import swin_transformer from . import swin_transformer
from . import lcnet
from .vgg import * from .vgg import *
from .resnet import * from .resnet import *
...@@ -41,3 +42,4 @@ from .res2net import * ...@@ -41,3 +42,4 @@ from .res2net import *
from .dla import * from .dla import *
from .shufflenet_v2 import * from .shufflenet_v2 import *
from .swin_transformer import * from .swin_transformer import *
from .lcnet import *
# copyright (c) 2021 PaddlePaddle Authors. All Rights Reserve.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import paddle
import paddle.nn as nn
from paddle import ParamAttr
from paddle.nn import AdaptiveAvgPool2D, BatchNorm, Conv2D, Dropout, Linear
from paddle.regularizer import L2Decay
from paddle.nn.initializer import KaimingNormal
from ppdet.core.workspace import register, serializable
from numbers import Integral
from ..shape_spec import ShapeSpec
__all__ = ['LCNet']
NET_CONFIG = {
"blocks2":
#k, in_c, out_c, s, use_se
[[3, 16, 32, 1, False], ],
"blocks3": [
[3, 32, 64, 2, False],
[3, 64, 64, 1, False],
],
"blocks4": [
[3, 64, 128, 2, False],
[3, 128, 128, 1, False],
],
"blocks5": [
[3, 128, 256, 2, False],
[5, 256, 256, 1, False],
[5, 256, 256, 1, False],
[5, 256, 256, 1, False],
[5, 256, 256, 1, False],
[5, 256, 256, 1, False],
],
"blocks6": [[5, 256, 512, 2, True], [5, 512, 512, 1, True]]
}
def make_divisible(v, divisor=8, min_value=None):
if min_value is None:
min_value = divisor
new_v = max(min_value, int(v + divisor / 2) // divisor * divisor)
if new_v < 0.9 * v:
new_v += divisor
return new_v
class ConvBNLayer(nn.Layer):
def __init__(self,
num_channels,
filter_size,
num_filters,
stride,
num_groups=1):
super().__init__()
self.conv = Conv2D(
in_channels=num_channels,
out_channels=num_filters,
kernel_size=filter_size,
stride=stride,
padding=(filter_size - 1) // 2,
groups=num_groups,
weight_attr=ParamAttr(initializer=KaimingNormal()),
bias_attr=False)
self.bn = BatchNorm(
num_filters,
param_attr=ParamAttr(regularizer=L2Decay(0.0)),
bias_attr=ParamAttr(regularizer=L2Decay(0.0)))
self.hardswish = nn.Hardswish()
def forward(self, x):
x = self.conv(x)
x = self.bn(x)
x = self.hardswish(x)
return x
class DepthwiseSeparable(nn.Layer):
def __init__(self,
num_channels,
num_filters,
stride,
dw_size=3,
use_se=False):
super().__init__()
self.use_se = use_se
self.dw_conv = ConvBNLayer(
num_channels=num_channels,
num_filters=num_channels,
filter_size=dw_size,
stride=stride,
num_groups=num_channels)
if use_se:
self.se = SEModule(num_channels)
self.pw_conv = ConvBNLayer(
num_channels=num_channels,
filter_size=1,
num_filters=num_filters,
stride=1)
def forward(self, x):
x = self.dw_conv(x)
if self.use_se:
x = self.se(x)
x = self.pw_conv(x)
return x
class SEModule(nn.Layer):
def __init__(self, channel, reduction=4):
super().__init__()
self.avg_pool = AdaptiveAvgPool2D(1)
self.conv1 = Conv2D(
in_channels=channel,
out_channels=channel // reduction,
kernel_size=1,
stride=1,
padding=0)
self.relu = nn.ReLU()
self.conv2 = Conv2D(
in_channels=channel // reduction,
out_channels=channel,
kernel_size=1,
stride=1,
padding=0)
self.hardsigmoid = nn.Hardsigmoid()
def forward(self, x):
identity = x
x = self.avg_pool(x)
x = self.conv1(x)
x = self.relu(x)
x = self.conv2(x)
x = self.hardsigmoid(x)
x = paddle.multiply(x=identity, y=x)
return x
@register
@serializable
class LCNet(nn.Layer):
def __init__(self, scale=1.0, feature_maps=[3, 4, 5]):
super().__init__()
self.scale = scale
self.feature_maps = feature_maps
out_channels = []
self.conv1 = ConvBNLayer(
num_channels=3,
filter_size=3,
num_filters=make_divisible(16 * scale),
stride=2)
self.blocks2 = nn.Sequential(* [
DepthwiseSeparable(
num_channels=make_divisible(in_c * scale),
num_filters=make_divisible(out_c * scale),
dw_size=k,
stride=s,
use_se=se)
for i, (k, in_c, out_c, s, se) in enumerate(NET_CONFIG["blocks2"])
])
self.blocks3 = nn.Sequential(* [
DepthwiseSeparable(
num_channels=make_divisible(in_c * scale),
num_filters=make_divisible(out_c * scale),
dw_size=k,
stride=s,
use_se=se)
for i, (k, in_c, out_c, s, se) in enumerate(NET_CONFIG["blocks3"])
])
out_channels.append(
make_divisible(NET_CONFIG["blocks3"][-1][2] * scale))
self.blocks4 = nn.Sequential(* [
DepthwiseSeparable(
num_channels=make_divisible(in_c * scale),
num_filters=make_divisible(out_c * scale),
dw_size=k,
stride=s,
use_se=se)
for i, (k, in_c, out_c, s, se) in enumerate(NET_CONFIG["blocks4"])
])
out_channels.append(
make_divisible(NET_CONFIG["blocks4"][-1][2] * scale))
self.blocks5 = nn.Sequential(* [
DepthwiseSeparable(
num_channels=make_divisible(in_c * scale),
num_filters=make_divisible(out_c * scale),
dw_size=k,
stride=s,
use_se=se)
for i, (k, in_c, out_c, s, se) in enumerate(NET_CONFIG["blocks5"])
])
out_channels.append(
make_divisible(NET_CONFIG["blocks5"][-1][2] * scale))
self.blocks6 = nn.Sequential(* [
DepthwiseSeparable(
num_channels=make_divisible(in_c * scale),
num_filters=make_divisible(out_c * scale),
dw_size=k,
stride=s,
use_se=se)
for i, (k, in_c, out_c, s, se) in enumerate(NET_CONFIG["blocks6"])
])
out_channels.append(
make_divisible(NET_CONFIG["blocks6"][-1][2] * scale))
self._out_channels = [
ch for idx, ch in enumerate(out_channels) if idx + 2 in feature_maps
]
def forward(self, inputs):
x = inputs['image']
outs = []
x = self.conv1(x)
x = self.blocks2(x)
x = self.blocks3(x)
outs.append(x)
x = self.blocks4(x)
outs.append(x)
x = self.blocks5(x)
outs.append(x)
x = self.blocks6(x)
outs.append(x)
outs = [o for i, o in enumerate(outs) if i + 2 in self.feature_maps]
return outs
@property
def out_shape(self):
return [ShapeSpec(channels=c) for c in self._out_channels]
...@@ -21,6 +21,7 @@ import paddle.nn as nn ...@@ -21,6 +21,7 @@ import paddle.nn as nn
from paddle import ParamAttr from paddle import ParamAttr
from paddle.nn import Conv2D, MaxPool2D, AdaptiveAvgPool2D, BatchNorm from paddle.nn import Conv2D, MaxPool2D, AdaptiveAvgPool2D, BatchNorm
from paddle.nn.initializer import KaimingNormal from paddle.nn.initializer import KaimingNormal
from paddle.regularizer import L2Decay
from ppdet.core.workspace import register, serializable from ppdet.core.workspace import register, serializable
from numbers import Integral from numbers import Integral
...@@ -50,7 +51,11 @@ class ConvBNLayer(nn.Layer): ...@@ -50,7 +51,11 @@ class ConvBNLayer(nn.Layer):
weight_attr=ParamAttr(initializer=KaimingNormal()), weight_attr=ParamAttr(initializer=KaimingNormal()),
bias_attr=False) bias_attr=False)
self._batch_norm = BatchNorm(out_channels, act=act) self._batch_norm = BatchNorm(
out_channels,
param_attr=ParamAttr(regularizer=L2Decay(0.0)),
bias_attr=ParamAttr(regularizer=L2Decay(0.0)),
act=act)
def forward(self, inputs): def forward(self, inputs):
y = self._conv(inputs) y = self._conv(inputs)
...@@ -159,14 +164,9 @@ class InvertedResidualDS(nn.Layer): ...@@ -159,14 +164,9 @@ class InvertedResidualDS(nn.Layer):
@register @register
@serializable @serializable
class ShuffleNetV2(nn.Layer): class ShuffleNetV2(nn.Layer):
def __init__(self, def __init__(self, scale=1.0, act="relu", feature_maps=[5, 13, 17]):
scale=1.0,
act="relu",
feature_maps=[5, 13, 17],
with_last_conv=False):
super(ShuffleNetV2, self).__init__() super(ShuffleNetV2, self).__init__()
self.scale = scale self.scale = scale
self.with_last_conv = with_last_conv
if isinstance(feature_maps, Integral): if isinstance(feature_maps, Integral):
feature_maps = [feature_maps] feature_maps = [feature_maps]
self.feature_maps = feature_maps self.feature_maps = feature_maps
...@@ -226,19 +226,6 @@ class ShuffleNetV2(nn.Layer): ...@@ -226,19 +226,6 @@ class ShuffleNetV2(nn.Layer):
self._update_out_channels(stage_out_channels[stage_id + 2], self._update_out_channels(stage_out_channels[stage_id + 2],
self._feature_idx, self.feature_maps) self._feature_idx, self.feature_maps)
if self.with_last_conv:
# last_conv
self._last_conv = ConvBNLayer(
in_channels=stage_out_channels[-2],
out_channels=stage_out_channels[-1],
kernel_size=1,
stride=1,
padding=0,
act=act)
self._feature_idx += 1
self._update_out_channels(stage_out_channels[-1], self._feature_idx,
self.feature_maps)
def _update_out_channels(self, channel, feature_idx, feature_maps): def _update_out_channels(self, channel, feature_idx, feature_maps):
if feature_idx in feature_maps: if feature_idx in feature_maps:
self._out_channels.append(channel) self._out_channels.append(channel)
...@@ -252,9 +239,6 @@ class ShuffleNetV2(nn.Layer): ...@@ -252,9 +239,6 @@ class ShuffleNetV2(nn.Layer):
if i + 2 in self.feature_maps: if i + 2 in self.feature_maps:
outs.append(y) outs.append(y)
if self.with_last_conv:
y = self._last_conv(y)
outs.append(y)
return outs return outs
@property @property
......
...@@ -139,6 +139,13 @@ def match_state_dict(model_state_dict, weight_state_dict): ...@@ -139,6 +139,13 @@ def match_state_dict(model_state_dict, weight_state_dict):
max_id = match_matrix.argmax(1) max_id = match_matrix.argmax(1)
max_len = match_matrix.max(1) max_len = match_matrix.max(1)
max_id[max_len == 0] = -1 max_id[max_len == 0] = -1
not_load_weight_name = []
for match_idx in range(len(max_id)):
if match_idx < len(weight_keys) and max_id[match_idx] == -1:
not_load_weight_name.append(weight_keys[match_idx])
if len(not_load_weight_name) > 0:
logger.info('{} in pretrained weight is not used in the model, '
'and its will not be loaded'.format(not_load_weight_name))
matched_keys = {} matched_keys = {}
result_state_dict = {} result_state_dict = {}
for model_id, weight_id in enumerate(max_id): for model_id, weight_id in enumerate(max_id):
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册