test_quant_aware.py 7.5 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
# Copyright (c) 2019  PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import sys
sys.path.append("../")
import unittest
import paddle
from paddleslim.quant import quant_aware, convert
19
from static_case import StaticCase
20 21 22 23 24 25 26 27 28
sys.path.append("../demo")
from models import MobileNet
from layers import conv_bn_layer
import paddle.dataset.mnist as reader
from paddle.fluid.framework import IrGraph
from paddle.fluid import core
import numpy as np


29
class TestQuantAwareCase1(StaticCase):
30
    def get_model(self):
B
Bai Yifan 已提交
31 32 33
        image = paddle.static.data(
            name='image', shape=[None, 1, 28, 28], dtype='float32')
        label = paddle.static.data(name='label', shape=[None, 1], dtype='int64')
34 35
        model = MobileNet()
        out = model.net(input=image, class_dim=10)
B
Bai Yifan 已提交
36 37 38 39
        cost = paddle.nn.functional.loss.cross_entropy(input=out, label=label)
        avg_cost = paddle.mean(x=cost)
        startup_prog = paddle.static.default_startup_program()
        train_prog = paddle.static.default_main_program()
40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55
        return startup_prog, train_prog

    def get_op_number(self, prog):

        graph = IrGraph(core.Graph(prog.desc), for_test=False)
        quant_op_nums = 0
        op_nums = 0
        for op in graph.all_op_nodes():
            if op.name() in ['conv2d', 'depthwise_conv2d', 'mul']:
                op_nums += 1
            elif 'fake_' in op.name():
                quant_op_nums += 1
        return op_nums, quant_op_nums

    def test_quant_op(self):
        startup_prog, train_prog = self.get_model()
B
Bai Yifan 已提交
56
        place = paddle.CUDAPlace(0) if paddle.is_compiled_with_cuda(
B
Bai Yifan 已提交
57 58
        ) else paddle.CPUPlace()
        exe = paddle.static.Executor(place)
59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90
        exe.run(startup_prog)
        config_1 = {
            'weight_quantize_type': 'channel_wise_abs_max',
            'activation_quantize_type': 'moving_average_abs_max',
            'quantize_op_types': ['depthwise_conv2d', 'mul', 'conv2d'],
        }

        quant_prog_1 = quant_aware(
            train_prog, place, config=config_1, for_test=True)
        op_nums_1, quant_op_nums_1 = self.get_op_number(quant_prog_1)
        convert_prog_1 = convert(quant_prog_1, place, config=config_1)
        convert_op_nums_1, convert_quant_op_nums_1 = self.get_op_number(
            convert_prog_1)

        config_1['not_quant_pattern'] = ['last_fc']
        quant_prog_2 = quant_aware(
            train_prog, place, config=config_1, for_test=True)
        op_nums_2, quant_op_nums_2 = self.get_op_number(quant_prog_2)
        convert_prog_2 = convert(quant_prog_2, place, config=config_1)
        convert_op_nums_2, convert_quant_op_nums_2 = self.get_op_number(
            convert_prog_2)

        self.assertTrue(op_nums_1 == op_nums_2)
        # test quant_aware op numbers
        self.assertTrue(op_nums_1 * 4 == quant_op_nums_1)
        # test convert op numbers
        self.assertTrue(convert_op_nums_1 * 2 == convert_quant_op_nums_1)
        # test skip_quant
        self.assertTrue(quant_op_nums_1 - 4 == quant_op_nums_2)
        self.assertTrue(convert_quant_op_nums_1 - 2 == convert_quant_op_nums_2)


91
class TestQuantAwareCase2(StaticCase):
92
    def test_accuracy(self):
B
Bai Yifan 已提交
93 94 95
        image = paddle.static.data(
            name='image', shape=[None, 1, 28, 28], dtype='float32')
        label = paddle.static.data(name='label', shape=[None, 1], dtype='int64')
96 97
        model = MobileNet()
        out = model.net(input=image, class_dim=10)
B
Bai Yifan 已提交
98 99 100 101 102
        cost = paddle.nn.functional.loss.cross_entropy(input=out, label=label)
        avg_cost = paddle.mean(x=cost)
        acc_top1 = paddle.metric.accuracy(input=out, label=label, k=1)
        acc_top5 = paddle.metric.accuracy(input=out, label=label, k=5)
        optimizer = paddle.optimizer.Momentum(
103 104
            momentum=0.9,
            learning_rate=0.01,
B
Bai Yifan 已提交
105
            weight_decay=paddle.regularizer.L2Decay(4e-5))
106
        optimizer.minimize(avg_cost)
B
Bai Yifan 已提交
107
        main_prog = paddle.static.default_main_program()
108 109
        val_prog = main_prog.clone(for_test=True)

B
Bai Yifan 已提交
110
        place = paddle.CUDAPlace(0) if paddle.is_compiled_with_cuda(
B
Bai Yifan 已提交
111 112 113
        ) else paddle.static.CPUPlace()
        exe = paddle.static.Executor(place)
        exe.run(paddle.static.default_startup_program())
B
Bai Yifan 已提交
114 115 116 117 118 119 120 121 122 123 124

        def transform(x):
            return np.reshape(x, [1, 28, 28])

        train_dataset = paddle.vision.datasets.MNIST(
            mode='train', backend='cv2', transform=transform)
        test_dataset = paddle.vision.datasets.MNIST(
            mode='test', backend='cv2', transform=transform)
        train_loader = paddle.io.DataLoader(
            train_dataset,
            places=place,
B
Bai Yifan 已提交
125
            feed_list=[image, label],
B
Bai Yifan 已提交
126
            drop_last=True,
127
            return_list=False,
B
Bai Yifan 已提交
128 129
            batch_size=64)
        valid_loader = paddle.io.DataLoader(
130 131 132 133 134
            test_dataset,
            places=place,
            feed_list=[image, label],
            batch_size=64,
            return_list=False)
135 136 137

        def train(program):
            iter = 0
B
Bai Yifan 已提交
138
            for data in train_loader():
139 140
                cost, top1, top5 = exe.run(
                    program,
B
Bai Yifan 已提交
141
                    feed=data,
142 143 144 145 146 147 148 149 150 151
                    fetch_list=[avg_cost, acc_top1, acc_top5])
                iter += 1
                if iter % 100 == 0:
                    print(
                        'train iter={}, avg loss {}, acc_top1 {}, acc_top5 {}'.
                        format(iter, cost, top1, top5))

        def test(program):
            iter = 0
            result = [[], [], []]
B
Bai Yifan 已提交
152
            for data in valid_loader():
153 154
                cost, top1, top5 = exe.run(
                    program,
B
Bai Yifan 已提交
155
                    feed=data,
156 157 158
                    fetch_list=[avg_cost, acc_top1, acc_top5])
                iter += 1
                if iter % 100 == 0:
159 160
                    print('eval iter={}, avg loss {}, acc_top1 {}, acc_top5 {}'.
                          format(iter, cost, top1, top5))
161 162 163 164 165 166 167 168 169 170 171 172 173 174 175
                result[0].append(cost)
                result[1].append(top1)
                result[2].append(top5)
            print(' avg loss {}, acc_top1 {}, acc_top5 {}'.format(
                np.mean(result[0]), np.mean(result[1]), np.mean(result[2])))
            return np.mean(result[1]), np.mean(result[2])

        train(main_prog)
        top1_1, top5_1 = test(main_prog)

        config = {
            'weight_quantize_type': 'channel_wise_abs_max',
            'activation_quantize_type': 'moving_average_abs_max',
            'quantize_op_types': ['depthwise_conv2d', 'mul', 'conv2d'],
        }
176
        quant_train_prog = quant_aware(main_prog, place, config, for_test=False)
177 178
        quant_eval_prog = quant_aware(val_prog, place, config, for_test=True)
        train(quant_train_prog)
179 180
        quant_eval_prog, int8_prog = convert(
            quant_eval_prog, place, config, save_int8=True)
181 182 183 184 185 186 187 188
        top1_2, top5_2 = test(quant_eval_prog)
        # values before quantization and after quantization should be close
        print("before quantization: top1: {}, top5: {}".format(top1_1, top5_1))
        print("after quantization: top1: {}, top5: {}".format(top1_2, top5_2))


if __name__ == '__main__':
    unittest.main()