From bf807d69a4b5cd797400f34a99867212c4680984 Mon Sep 17 00:00:00 2001 From: Zhen Wang Date: Wed, 6 Mar 2019 17:33:43 +0800 Subject: [PATCH] avoid ce fails on windows. --- .../slim/tests/test_quantization_pass.py | 161 ++++++++++-------- 1 file changed, 89 insertions(+), 72 deletions(-) diff --git a/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py b/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py index 254b73a124..11da352003 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py +++ b/python/paddle/fluid/contrib/slim/tests/test_quantization_pass.py @@ -123,7 +123,7 @@ class TestQuantizationTransformPass(unittest.TestCase): arg_name.endswith('.quantized.dequantized')) self.assertTrue(arg_name in quantized_ops) - def linear_fc_quant(self, quant_type): + def linear_fc_quant(self, quant_type, enable_ce=False): main = fluid.Program() startup = fluid.Program() with fluid.program_guard(main, startup): @@ -138,29 +138,29 @@ class TestQuantizationTransformPass(unittest.TestCase): place=place, activation_quantize_type=quant_type) transform_pass.apply(graph) - marked_nodes = set() - for op in graph.all_op_nodes(): - if op.name().find('quantize') > -1: - marked_nodes.add(op) - graph.draw('.', 'quantize_fc_' + quant_type, marked_nodes) + if not enable_ce: + marked_nodes = set() + for op in graph.all_op_nodes(): + if op.name().find('quantize') > -1: + marked_nodes.add(op) + graph.draw('.', 'quantize_fc_' + quant_type, marked_nodes) program = graph.to_program() self.check_program(transform_pass, program) val_graph = IrGraph(core.Graph(program.desc), for_test=False) - val_marked_nodes = set() - for op in val_graph.all_op_nodes(): - if op.name().find('quantize') > -1: - val_marked_nodes.add(op) - val_graph.draw('.', 'val_fc_' + quant_type, val_marked_nodes) + if not enable_ce: + val_marked_nodes = set() + for op in val_graph.all_op_nodes(): + if op.name().find('quantize') > -1: + val_marked_nodes.add(op) + val_graph.draw('.', 'val_fc_' + quant_type, val_marked_nodes) def test_linear_fc_quant_abs_max(self): - self.act_quant_op_type = 'fake_quantize_abs_max' - self.linear_fc_quant('abs_max') + self.linear_fc_quant('abs_max', enable_ce=True) def test_linear_fc_quant_range_abs_max(self): - self.act_quant_op_type = 'fake_quantize_range_abs_max' - self.linear_fc_quant('range_abs_max') + self.linear_fc_quant('range_abs_max', enable_ce=True) - def residual_block_quant(self, quant_type): + def residual_block_quant(self, quant_type, enable_ce=False): main = fluid.Program() startup = fluid.Program() with fluid.program_guard(main, startup): @@ -175,31 +175,31 @@ class TestQuantizationTransformPass(unittest.TestCase): place=place, activation_quantize_type=quant_type) transform_pass.apply(graph) - marked_nodes = set() - for op in graph.all_op_nodes(): - if op.name().find('quantize') > -1: - marked_nodes.add(op) - graph.draw('.', 'quantize_residual_' + quant_type, marked_nodes) + if not enable_ce: + marked_nodes = set() + for op in graph.all_op_nodes(): + if op.name().find('quantize') > -1: + marked_nodes.add(op) + graph.draw('.', 'quantize_residual_' + quant_type, marked_nodes) program = graph.to_program() self.check_program(transform_pass, program) val_graph = IrGraph(core.Graph(program.desc), for_test=False) - val_marked_nodes = set() - for op in val_graph.all_op_nodes(): - if op.name().find('quantize') > -1: - val_marked_nodes.add(op) - val_graph.draw('.', 'val_residual_' + quant_type, val_marked_nodes) + if not enable_ce: + val_marked_nodes = set() + for op in val_graph.all_op_nodes(): + if op.name().find('quantize') > -1: + val_marked_nodes.add(op) + val_graph.draw('.', 'val_residual_' + quant_type, val_marked_nodes) def test_residual_block_abs_max(self): - self.act_quant_op_type = 'fake_quantize_abs_max' - self.residual_block_quant('abs_max') + self.residual_block_quant('abs_max', enable_ce=True) def test_residual_block_range_abs_max(self): - self.act_quant_op_type = 'fake_quantize_range_abs_max' - self.residual_block_quant('range_abs_max') + self.residual_block_quant('range_abs_max', enable_ce=True) class TestQuantizationFreezePass(unittest.TestCase): - def freeze_graph(self, use_cuda, seed, quant_type): + def freeze_graph(self, use_cuda, seed, quant_type, enable_ce=False): def build_program(main, startup, is_test): main.random_seed = seed startup.random_seed = seed @@ -237,16 +237,17 @@ class TestQuantizationFreezePass(unittest.TestCase): transform_pass.apply(main_graph) transform_pass.apply(test_graph) dev_name = '_gpu_' if use_cuda else '_cpu_' - marked_nodes = set() - for op in main_graph.all_op_nodes(): - if op.name().find('quantize') > -1: - marked_nodes.add(op) - main_graph.draw('.', 'main' + dev_name + quant_type, marked_nodes) - marked_nodes = set() - for op in test_graph.all_op_nodes(): - if op.name().find('quantize') > -1: - marked_nodes.add(op) - test_graph.draw('.', 'test' + dev_name + quant_type, marked_nodes) + if not enable_ce: + marked_nodes = set() + for op in main_graph.all_op_nodes(): + if op.name().find('quantize') > -1: + marked_nodes.add(op) + main_graph.draw('.', 'main' + dev_name + quant_type, marked_nodes) + marked_nodes = set() + for op in test_graph.all_op_nodes(): + if op.name().find('quantize') > -1: + marked_nodes.add(op) + test_graph.draw('.', 'test' + dev_name + quant_type, marked_nodes) quantized_main_program = main_graph.to_program() quantized_test_program = test_graph.to_program() @@ -266,7 +267,9 @@ class TestQuantizationFreezePass(unittest.TestCase): loss_v = exe.run(program=quantized_main_program, feed=feeder.feed(data), fetch_list=[loss]) - print('{}: {}'.format('loss' + dev_name + quant_type, loss_v)) + if not enable_ce: + print('{}: {}'.format('loss' + dev_name + quant_type, + loss_v)) test_data = next(test_reader()) with fluid.program_guard(quantized_test_program): @@ -281,12 +284,13 @@ class TestQuantizationFreezePass(unittest.TestCase): # Freeze graph for inference, but the weight of fc/conv is still float type. freeze_pass = QuantizationFreezePass(scope=scope, place=place) freeze_pass.apply(test_graph) - marked_nodes = set() - for op in test_graph.all_op_nodes(): - if op.name().find('quantize') > -1: - marked_nodes.add(op) - test_graph.draw('.', 'test_freeze' + dev_name + quant_type, - marked_nodes) + if not enable_ce: + marked_nodes = set() + for op in test_graph.all_op_nodes(): + if op.name().find('quantize') > -1: + marked_nodes.add(op) + test_graph.draw('.', 'test_freeze' + dev_name + quant_type, + marked_nodes) server_program = test_graph.to_program() with fluid.scope_guard(scope): @@ -294,24 +298,30 @@ class TestQuantizationFreezePass(unittest.TestCase): feed=feeder.feed(test_data), fetch_list=[loss]) self.assertAlmostEqual(test_loss1, test_loss2, delta=5e-3) - print('{}: {}'.format('test_loss1' + dev_name + quant_type, test_loss1)) - print('{}: {}'.format('test_loss2' + dev_name + quant_type, test_loss2)) + if not enable_ce: + print('{}: {}'.format('test_loss1' + dev_name + quant_type, + test_loss1)) + print('{}: {}'.format('test_loss2' + dev_name + quant_type, + test_loss2)) w_freeze = np.array(scope.find_var('conv2d_1.w_0').get_tensor()) # Maybe failed, this is due to the calculation precision # self.assertAlmostEqual(np.sum(w_freeze), np.sum(w_quant)) - print('{}: {}'.format('w_freeze' + dev_name + quant_type, - np.sum(w_freeze))) - print('{}: {}'.format('w_quant' + dev_name + quant_type, - np.sum(w_quant))) + if not enable_ce: + print('{}: {}'.format('w_freeze' + dev_name + quant_type, + np.sum(w_freeze))) + print('{}: {}'.format('w_quant' + dev_name + quant_type, + np.sum(w_quant))) # Convert parameter to 8-bit. convert_int8_pass = ConvertToInt8Pass(scope=scope, place=place) convert_int8_pass.apply(test_graph) - marked_nodes = set() - for op in test_graph.all_op_nodes(): - if op.name().find('quantize') > -1: - marked_nodes.add(op) - test_graph.draw('.', 'test_int8' + dev_name + quant_type, marked_nodes) + if not enable_ce: + marked_nodes = set() + for op in test_graph.all_op_nodes(): + if op.name().find('quantize') > -1: + marked_nodes.add(op) + test_graph.draw('.', 'test_int8' + dev_name + quant_type, + marked_nodes) server_program_int8 = test_graph.to_program() # Save the 8-bit parameter and model file. with fluid.scope_guard(scope): @@ -325,18 +335,21 @@ class TestQuantizationFreezePass(unittest.TestCase): w_8bit = np.array(scope.find_var('conv2d_1.w_0.int8').get_tensor()) self.assertEqual(w_8bit.dtype, np.int8) self.assertEqual(np.sum(w_8bit), np.sum(w_freeze)) - print('{}: {}'.format('w_8bit' + dev_name + quant_type, np.sum(w_8bit))) - print('{}: {}'.format('w_freeze' + dev_name + quant_type, - np.sum(w_freeze))) + if not enable_ce: + print('{}: {}'.format('w_8bit' + dev_name + quant_type, + np.sum(w_8bit))) + print('{}: {}'.format('w_freeze' + dev_name + quant_type, + np.sum(w_freeze))) mobile_pass = TransformForMobilePass() mobile_pass.apply(test_graph) - marked_nodes = set() - for op in test_graph.all_op_nodes(): - if op.name().find('quantize') > -1: - marked_nodes.add(op) - test_graph.draw('.', 'test_mobile' + dev_name + quant_type, - marked_nodes) + if not enable_ce: + marked_nodes = set() + for op in test_graph.all_op_nodes(): + if op.name().find('quantize') > -1: + marked_nodes.add(op) + test_graph.draw('.', 'test_mobile' + dev_name + quant_type, + marked_nodes) mobile_program = test_graph.to_program() with fluid.scope_guard(scope): @@ -347,20 +360,24 @@ class TestQuantizationFreezePass(unittest.TestCase): def test_freeze_graph_cuda_dynamic(self): if fluid.core.is_compiled_with_cuda(): with fluid.unique_name.guard(): - self.freeze_graph(True, seed=1, quant_type='abs_max') + self.freeze_graph( + True, seed=1, quant_type='abs_max', enable_ce=True) def test_freeze_graph_cpu_dynamic(self): with fluid.unique_name.guard(): - self.freeze_graph(False, seed=2, quant_type='abs_max') + self.freeze_graph( + False, seed=2, quant_type='abs_max', enable_ce=True) def test_freeze_graph_cuda_static(self): if fluid.core.is_compiled_with_cuda(): with fluid.unique_name.guard(): - self.freeze_graph(True, seed=1, quant_type='range_abs_max') + self.freeze_graph( + True, seed=1, quant_type='range_abs_max', enable_ce=True) def test_freeze_graph_cpu_static(self): with fluid.unique_name.guard(): - self.freeze_graph(False, seed=2, quant_type='range_abs_max') + self.freeze_graph( + False, seed=2, quant_type='range_abs_max', enable_ce=True) if __name__ == '__main__': -- GitLab