diff --git a/demo/sensitive_prune/greedy_prune.py b/demo/sensitive_prune/greedy_prune.py index b6703a9adf72893b8098ab492460d535e81d56a9..2534045f98d7ffd4977c1e6dc07f55949f0bf457 100644 --- a/demo/sensitive_prune/greedy_prune.py +++ b/demo/sensitive_prune/greedy_prune.py @@ -217,7 +217,7 @@ def compress(args): train(i, pruned_program) acc = test(i, pruned_val_program) print("iter:{}; pruned FLOPS: {}; acc: {}".format( - iter, float(base_flops - current_flops) / base_flops), acc) + iter, float(base_flops - current_flops) / base_flops, acc)) pruner.save_checkpoint(pruned_program, pruned_val_program) diff --git a/paddleslim/prune/sensitive.py b/paddleslim/prune/sensitive.py index 23338284a1f70584b3a922b64304d26a30fe9b76..ab3a57cbc4389b6ea661a3cf2c83dfe5cb2965f9 100644 --- a/paddleslim/prune/sensitive.py +++ b/paddleslim/prune/sensitive.py @@ -57,12 +57,15 @@ def sensitivity(program, if ratio in sensitivities[name]['pruned_percent']: _logger.debug('{}, {} has computed.'.format(name, ratio)) ratio += step_size + pruned_times += 1 continue if baseline is None: baseline = eval_func(graph.program) param_backup = {} pruner = Pruner() + _logger.info("sensitive - param: {}; ratios: {}".format(name, + ratio)) pruned_program = pruner.prune( program=graph.program, scope=scope,