未验证 提交 1dfba509 编写于 作者: Z Zeng Jinle 提交者: GitHub

set enable_inplace to True in models, test=develop (#3297)

上级 9958875f
......@@ -134,7 +134,6 @@ def main():
# compile program for multi-devices
build_strategy = fluid.BuildStrategy()
build_strategy.enable_inplace = False
sync_bn = getattr(model.backbone, 'norm_type', None) == 'sync_bn'
# only enable sync_bn in multi GPU devices
build_strategy.sync_batch_norm = sync_bn and devices_num > 1 and cfg.use_gpu
......
......@@ -189,8 +189,6 @@ def train(args):
losses = [[], []]
t_time = 0
build_strategy = fluid.BuildStrategy()
build_strategy.enable_inplace = False
build_strategy.memory_optimize = False
exec_strategy = fluid.ExecutionStrategy()
exec_strategy.num_threads = 1
......
......@@ -316,7 +316,6 @@ class AttGAN(object):
### memory optim
build_strategy = fluid.BuildStrategy()
build_strategy.enable_inplace = False
gen_trainer_program = fluid.CompiledProgram(
gen_trainer.program).with_data_parallel(
......
......@@ -245,7 +245,6 @@ class Pix2pix(object):
### memory optim
build_strategy = fluid.BuildStrategy()
build_strategy.enable_inplace = False
gen_trainer_program = fluid.CompiledProgram(
gen_trainer.program).with_data_parallel(
......
......@@ -328,7 +328,6 @@ class STGAN(object):
### memory optim
build_strategy = fluid.BuildStrategy()
build_strategy.enable_inplace = False
gen_trainer_program = fluid.CompiledProgram(
gen_trainer.program).with_data_parallel(
......
......@@ -294,7 +294,6 @@ class StarGAN(object):
### memory optim
build_strategy = fluid.BuildStrategy()
build_strategy.enable_inplace = False
gen_trainer_program = fluid.CompiledProgram(
gen_trainer.program).with_data_parallel(
......
......@@ -171,7 +171,6 @@ def train(args):
build_strategy.enable_inplace = True
if args.model_name in ['CTCN']:
build_strategy.enable_sequential_execution = True
#build_strategy.memory_optimize = True
compiled_train_prog = fluid.compiler.CompiledProgram(
train_prog).with_data_parallel(
......
......@@ -200,7 +200,6 @@ build_strategy = fluid.BuildStrategy()
if args.memory_optimize:
build_strategy.fuse_relu_depthwise_conv = True
build_strategy.enable_inplace = True
build_strategy.memory_optimize = True
place = fluid.CPUPlace()
if args.use_gpu:
......
......@@ -284,7 +284,6 @@ def train_parallel(args):
strategy.num_iteration_per_drop_scope = 30
build_strategy = fluid.BuildStrategy()
build_strategy.enable_inplace = False
build_strategy.memory_optimize = False
build_strategy.enable_sequential_execution = bool(
args.enable_sequential_execution)
......
......@@ -210,7 +210,6 @@ def train(args,
loss.persistable = True
build_strategy = fluid.BuildStrategy()
build_strategy.enable_inplace = True
build_strategy.memory_optimize = True
train_exe = fluid.ParallelExecutor(main_program=train_prog,
use_cuda=use_gpu, loss_name=loss.name, build_strategy=build_strategy)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册