darknet.py 5.5 KB
Newer Older
1 2
#  Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserve.
#
3 4 5
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
6 7 8
#
#    http://www.apache.org/licenses/LICENSE-2.0
#
9 10 11 12 13
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
14

15
import paddle
16 17 18 19
import paddle.fluid as fluid
from paddle.fluid.param_attr import ParamAttr
from paddle.fluid.regularizer import L2Decay

20
from paddle.fluid.dygraph.nn import BatchNorm
21 22 23


class ConvBNLayer(fluid.dygraph.Layer):
24 25 26 27 28 29 30 31 32 33 34
    def __init__(
        self,
        ch_in,
        ch_out,
        filter_size=3,
        stride=1,
        groups=1,
        padding=0,
        act="leaky",
        is_test=True,
    ):
35
        super().__init__()
36

37 38 39 40
        self.conv = paddle.nn.Conv2D(
            in_channels=ch_in,
            out_channels=ch_out,
            kernel_size=filter_size,
41 42 43
            stride=stride,
            padding=padding,
            groups=groups,
44
            weight_attr=ParamAttr(
45 46 47 48
                initializer=fluid.initializer.Normal(0.0, 0.02)
            ),
            bias_attr=False,
        )
49 50 51
        self.batch_norm = BatchNorm(
            num_channels=ch_out,
            is_test=is_test,
52 53 54 55 56 57 58 59 60
            param_attr=ParamAttr(
                initializer=fluid.initializer.Normal(0.0, 0.02),
                regularizer=L2Decay(0.0),
            ),
            bias_attr=ParamAttr(
                initializer=fluid.initializer.Constant(0.0),
                regularizer=L2Decay(0.0),
            ),
        )
61 62 63 64 65 66 67

        self.act = act

    def forward(self, inputs):
        out = self.conv(inputs)
        out = self.batch_norm(out)
        if self.act == 'leaky':
68
            out = paddle.nn.functional.leaky_relu(out, 0.1)
69 70 71 72
        return out


class DownSample(fluid.dygraph.Layer):
73 74 75
    def __init__(
        self, ch_in, ch_out, filter_size=3, stride=2, padding=1, is_test=True
    ):
76

77
        super().__init__()
78

79 80 81 82 83 84 85 86
        self.conv_bn_layer = ConvBNLayer(
            ch_in=ch_in,
            ch_out=ch_out,
            filter_size=filter_size,
            stride=stride,
            padding=padding,
            is_test=is_test,
        )
87 88 89 90 91 92 93 94 95
        self.ch_out = ch_out

    def forward(self, inputs):
        out = self.conv_bn_layer(inputs)
        return out


class BasicBlock(fluid.dygraph.Layer):
    def __init__(self, ch_in, ch_out, is_test=True):
96
        super().__init__()
97

98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113
        self.conv1 = ConvBNLayer(
            ch_in=ch_in,
            ch_out=ch_out,
            filter_size=1,
            stride=1,
            padding=0,
            is_test=is_test,
        )
        self.conv2 = ConvBNLayer(
            ch_in=ch_out,
            ch_out=ch_out * 2,
            filter_size=3,
            stride=1,
            padding=1,
            is_test=is_test,
        )
114 115 116 117 118 119 120 121 122 123

    def forward(self, inputs):
        conv1 = self.conv1(inputs)
        conv2 = self.conv2(conv1)
        out = fluid.layers.elementwise_add(x=inputs, y=conv2, act=None)
        return out


class LayerWarp(fluid.dygraph.Layer):
    def __init__(self, ch_in, ch_out, count, is_test=True):
124
        super().__init__()
125 126 127 128 129 130

        self.basicblock0 = BasicBlock(ch_in, ch_out, is_test=is_test)
        self.res_out_list = []
        for i in range(1, count):
            res_out = self.add_sublayer(
                "basic_block_%d" % (i),
131 132
                BasicBlock(ch_out * 2, ch_out, is_test=is_test),
            )
133 134 135 136 137 138 139 140 141 142 143 144 145 146 147
            self.res_out_list.append(res_out)
        self.ch_out = ch_out

    def forward(self, inputs):
        y = self.basicblock0(inputs)
        for basic_block_i in self.res_out_list:
            y = basic_block_i(y)
        return y


DarkNet_cfg = {53: ([1, 2, 8, 8, 4])}


class DarkNet53_conv_body(fluid.dygraph.Layer):
    def __init__(self, ch_in=3, is_test=True):
148
        super().__init__()
149 150 151
        self.stages = DarkNet_cfg[53]
        self.stages = self.stages[0:5]

152 153 154 155 156 157 158 159
        self.conv0 = ConvBNLayer(
            ch_in=ch_in,
            ch_out=32,
            filter_size=3,
            stride=1,
            padding=1,
            is_test=is_test,
        )
160 161 162 163 164 165 166 167

        self.downsample0 = DownSample(ch_in=32, ch_out=32 * 2, is_test=is_test)
        self.darknet53_conv_block_list = []
        self.downsample_list = []
        ch_in = [64, 128, 256, 512, 1024]
        for i, stage in enumerate(self.stages):
            conv_block = self.add_sublayer(
                "stage_%d" % (i),
168 169
                LayerWarp(int(ch_in[i]), 32 * (2**i), stage, is_test=is_test),
            )
170 171 172 173
            self.darknet53_conv_block_list.append(conv_block)
        for i in range(len(self.stages) - 1):
            downsample = self.add_sublayer(
                "stage_%d_downsample" % i,
174 175 176 177 178 179
                DownSample(
                    ch_in=32 * (2 ** (i + 1)),
                    ch_out=32 * (2 ** (i + 2)),
                    is_test=is_test,
                ),
            )
180 181 182 183 184 185 186 187 188 189 190 191 192
            self.downsample_list.append(downsample)

    def forward(self, inputs):

        out = self.conv0(inputs)
        out = self.downsample0(out)
        blocks = []
        for i, conv_block_i in enumerate(self.darknet53_conv_block_list):
            out = conv_block_i(out)
            blocks.append(out)
            if i < len(self.stages) - 1:
                out = self.downsample_list[i](out)
        return blocks[-1:-4:-1]