test_graph.py 4.5 KB
Newer Older
W
WangZhen 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
#   copyright (c) 2018 paddlepaddle authors. all rights reserved.
#
# licensed under the apache license, version 2.0 (the "license");
# you may not use this file except in compliance with the license.
# you may obtain a copy of the license at
#
#     http://www.apache.org/licenses/license-2.0
#
# unless required by applicable law or agreed to in writing, software
# distributed under the license is distributed on an "as is" basis,
# without warranties or conditions of any kind, either express or implied.
# see the license for the specific language governing permissions and
# limitations under the license.

from __future__ import print_function
16 17
import os
import six
W
WangZhen 已提交
18
import unittest
19
import paddle
W
WangZhen 已提交
20 21 22 23
import paddle.fluid as fluid
from paddle.fluid.framework import IrGraph
from paddle.fluid import core

24 25
os.environ["CUDA_VISIBLE_DEVICES"] = "0"
os.environ["CPU_NUM"] = "1"
W
WangZhen 已提交
26 27


28 29
def conv_block():
    img = fluid.layers.data(name='image', shape=[1, 28, 28], dtype='float32')
W
WangZhen 已提交
30
    label = fluid.layers.data(name='label', shape=[1], dtype='int64')
31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49
    conv_pool_1 = fluid.nets.simple_img_conv_pool(
        input=img,
        filter_size=5,
        num_filters=20,
        pool_size=2,
        pool_stride=2,
        act="relu")
    conv_pool_1 = fluid.layers.batch_norm(conv_pool_1)
    conv_pool_2 = fluid.nets.simple_img_conv_pool(
        input=conv_pool_1,
        filter_size=5,
        num_filters=50,
        pool_size=2,
        pool_stride=2,
        act="relu")
    prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax')
    loss = fluid.layers.cross_entropy(input=prediction, label=label)
    avg_loss = fluid.layers.mean(loss)
    return [img, label], avg_loss
W
WangZhen 已提交
50 51 52


class TestGraph(unittest.TestCase):
53
    def graph_apis(self, use_cuda=False, for_ci=True):
W
WangZhen 已提交
54 55 56
        main = fluid.Program()
        startup = fluid.Program()
        with fluid.program_guard(main, startup):
57
            feeds, loss = conv_block()
W
WangZhen 已提交
58 59 60
            opt = fluid.optimizer.Adam(learning_rate=0.001)
            opt.minimize(loss)
        graph = IrGraph(core.Graph(main.desc), for_test=False)
61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90
        backup_graph = graph.clone()
        self.assertEqual(len(graph.all_nodes()), len(backup_graph.all_nodes()))
        build_strategy = fluid.BuildStrategy()
        build_strategy.memory_optimize = False
        build_strategy.enable_inplace = False
        origin_binary = fluid.CompiledProgram(graph.graph).with_data_parallel(
            loss_name=loss.name, build_strategy=build_strategy)
        backup_binary = fluid.CompiledProgram(
            backup_graph.graph).with_data_parallel(
                loss_name=loss.name, build_strategy=build_strategy)
        place = fluid.CUDAPlace(0) if use_cuda else fluid.CPUPlace()
        exe = fluid.Executor(place)
        exe.run(startup)
        iters = 5
        batch_size = 8
        train_reader = paddle.batch(
            paddle.dataset.mnist.train(), batch_size=batch_size)
        feeder = fluid.DataFeeder(feed_list=feeds, place=place)

        def train(binary):
            for _ in range(iters):
                data = next(train_reader())
                loss_v = exe.run(binary,
                                 feed=feeder.feed(data),
                                 fetch_list=[loss.name])
                print('{}: {}'.format('loss', loss_v))

        train(origin_binary)
        train(backup_binary)

W
WangZhen 已提交
91
        marked_nodes = set()
92
        for op in graph.all_op_nodes():
W
WangZhen 已提交
93 94
            if op.name().find('conv2d') > -1:
                marked_nodes.add(op)
95 96
        if not for_ci:
            graph.draw('.', 'residual', marked_nodes)
97 98 99 100 101
            backup_marked_nodes = set()
            for op in backup_graph.all_op_nodes():
                if op.name().find('conv2d') > -1:
                    backup_marked_nodes.add(op)
            backup_graph.draw('.', 'backup', backup_marked_nodes)
W
WangZhen 已提交
102 103 104
        self.assertFalse(graph.has_circle())
        self.assertEqual(graph.graph_num(), 1)
        nodes = graph.topology_sort()
105
        self.assertEqual(len(nodes), len(graph.all_op_nodes()))
W
WangZhen 已提交
106
        nodes_map = graph.build_adjacency_list()
107
        self.assertEqual(len(nodes_map), len(graph.all_op_nodes()))
W
WangZhen 已提交
108 109 110
        nodes_num = len(graph.all_nodes())
        graph.safe_remove_nodes(marked_nodes)
        self.assertEqual(len(graph.all_nodes()), nodes_num - len(marked_nodes))
111 112 113 114 115 116 117

    def test_graph_apis_cpu(self):
        self.graph_apis(use_cuda=False, for_ci=True)

    def test_graph_apis_cuda(self):
        if fluid.core.is_compiled_with_cuda():
            self.graph_apis(use_cuda=True, for_ci=True)
W
WangZhen 已提交
118 119 120 121


if __name__ == '__main__':
    unittest.main()