test_graph.py 5.9 KB
Newer Older
W
WangZhen 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14
#   copyright (c) 2018 paddlepaddle authors. all rights reserved.
#
# licensed under the apache license, version 2.0 (the "license");
# you may not use this file except in compliance with the license.
# you may obtain a copy of the license at
#
#     http://www.apache.org/licenses/license-2.0
#
# unless required by applicable law or agreed to in writing, software
# distributed under the license is distributed on an "as is" basis,
# without warranties or conditions of any kind, either express or implied.
# see the license for the specific language governing permissions and
# limitations under the license.

15
import os
16
import numpy as np
W
WangZhen 已提交
17
import unittest
18
import paddle
W
WangZhen 已提交
19 20 21 22
import paddle.fluid as fluid
from paddle.fluid.framework import IrGraph
from paddle.fluid import core

P
pangyoki 已提交
23 24
paddle.enable_static()

25 26
os.environ["CUDA_VISIBLE_DEVICES"] = "0"
os.environ["CPU_NUM"] = "1"
W
WangZhen 已提交
27 28


29 30
def conv_block():
    img = fluid.layers.data(name='image', shape=[1, 28, 28], dtype='float32')
W
WangZhen 已提交
31
    label = fluid.layers.data(name='label', shape=[1], dtype='int64')
32 33 34 35 36 37
    conv_pool_1 = fluid.nets.simple_img_conv_pool(input=img,
                                                  filter_size=5,
                                                  num_filters=20,
                                                  pool_size=2,
                                                  pool_stride=2,
                                                  act="relu")
38
    conv_pool_1 = fluid.layers.batch_norm(conv_pool_1)
39 40 41 42 43 44
    conv_pool_2 = fluid.nets.simple_img_conv_pool(input=conv_pool_1,
                                                  filter_size=5,
                                                  num_filters=50,
                                                  pool_size=2,
                                                  pool_stride=2,
                                                  act="relu")
45 46
    prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax')
    loss = fluid.layers.cross_entropy(input=prediction, label=label)
47
    avg_loss = paddle.mean(loss)
48
    return [img, label], avg_loss
W
WangZhen 已提交
49 50 51


class TestGraph(unittest.TestCase):
52

53
    def graph_apis(self, use_cuda=False, for_ci=True):
W
WangZhen 已提交
54 55
        main = fluid.Program()
        startup = fluid.Program()
56 57 58 59 60
        with fluid.unique_name.guard():
            with fluid.program_guard(main, startup):
                feeds, loss = conv_block()
                opt = fluid.optimizer.Adam(learning_rate=0.001)
                opt.minimize(loss)
W
WangZhen 已提交
61
        graph = IrGraph(core.Graph(main.desc), for_test=False)
62 63 64 65 66 67 68 69 70 71 72 73 74 75 76
        backup_graph = graph.clone()
        self.assertEqual(len(graph.all_nodes()), len(backup_graph.all_nodes()))
        build_strategy = fluid.BuildStrategy()
        build_strategy.memory_optimize = False
        build_strategy.enable_inplace = False
        origin_binary = fluid.CompiledProgram(graph.graph).with_data_parallel(
            loss_name=loss.name, build_strategy=build_strategy)
        backup_binary = fluid.CompiledProgram(
            backup_graph.graph).with_data_parallel(
                loss_name=loss.name, build_strategy=build_strategy)
        place = fluid.CUDAPlace(0) if use_cuda else fluid.CPUPlace()
        exe = fluid.Executor(place)
        exe.run(startup)
        iters = 5
        batch_size = 8
77 78
        train_reader = paddle.batch(paddle.dataset.mnist.train(),
                                    batch_size=batch_size)
79 80
        feeder = fluid.DataFeeder(feed_list=feeds, place=place)

81
        def _train(binary):
82 83 84 85 86
            for _ in range(iters):
                data = next(train_reader())
                loss_v = exe.run(binary,
                                 feed=feeder.feed(data),
                                 fetch_list=[loss.name])
87 88
                if not for_ci:
                    print('{}: {}'.format('loss', loss_v))
89

90 91 92 93 94 95 96 97 98 99 100
        _train(origin_binary)
        _train(backup_binary)

        checkponit_dir = "checkpoint_gpu" if use_cuda else "checkpoint_cpu"

        def _set_zero(var_name, scope, place):
            var = scope.find_var(var_name).get_tensor()
            var_array = np.zeros(var._get_dims()).astype("float32")
            var.set(var_array, place)

        sum_before = np.sum(
101 102
            np.array(
                fluid.global_scope().find_var('conv2d_1.w_0').get_tensor()))
103 104 105
        fluid.io._save_persistable_nodes(exe, checkponit_dir, graph)
        _set_zero('conv2d_1.w_0', fluid.global_scope(), place)
        set_after = np.sum(
106 107
            np.array(
                fluid.global_scope().find_var('conv2d_1.w_0').get_tensor()))
108 109 110
        self.assertEqual(set_after, 0)
        fluid.io._load_persistable_nodes(exe, checkponit_dir, graph)
        sum_after = np.sum(
111 112
            np.array(
                fluid.global_scope().find_var('conv2d_1.w_0').get_tensor()))
113
        self.assertEqual(sum_before, sum_after)
114

W
WangZhen 已提交
115
        marked_nodes = set()
116
        for op in graph.all_op_nodes():
W
WangZhen 已提交
117 118
            if op.name().find('conv2d') > -1:
                marked_nodes.add(op)
119 120
        if not for_ci:
            graph.draw('.', 'residual', marked_nodes)
121 122 123 124
            backup_marked_nodes = set()
            for op in backup_graph.all_op_nodes():
                if op.name().find('conv2d') > -1:
                    backup_marked_nodes.add(op)
125
            backup_graph.draw('./origin', 'backup', backup_marked_nodes)
W
WangZhen 已提交
126 127 128
        self.assertFalse(graph.has_circle())
        self.assertEqual(graph.graph_num(), 1)
        nodes = graph.topology_sort()
129
        self.assertEqual(len(nodes), len(graph.all_op_nodes()))
W
WangZhen 已提交
130
        nodes_map = graph.build_adjacency_list()
131
        self.assertEqual(len(nodes_map), len(graph.all_op_nodes()))
W
WangZhen 已提交
132 133 134
        nodes_num = len(graph.all_nodes())
        graph.safe_remove_nodes(marked_nodes)
        self.assertEqual(len(graph.all_nodes()), nodes_num - len(marked_nodes))
135 136 137 138 139 140 141

    def test_graph_apis_cpu(self):
        self.graph_apis(use_cuda=False, for_ci=True)

    def test_graph_apis_cuda(self):
        if fluid.core.is_compiled_with_cuda():
            self.graph_apis(use_cuda=True, for_ci=True)
W
WangZhen 已提交
142 143 144 145


if __name__ == '__main__':
    unittest.main()