提交 90b45a00 编写于 作者: Z zhangxuefei 提交者: wuzewu

update config (drop the memory optimization)

上级 f0c2925d
......@@ -30,14 +30,14 @@ class RunConfig(object):
def __init__(self,
log_interval=10,
eval_interval=100,
use_pyreader=False,
use_data_parallel=False,
use_pyreader=True,
use_data_parallel=True,
save_ckpt_interval=None,
use_cuda=True,
checkpoint_dir=None,
num_epoch=1,
batch_size=32,
enable_memory_optim=True,
enable_memory_optim=False,
strategy=None):
""" Construct finetune Config """
self._log_interval = log_interval
......@@ -53,7 +53,11 @@ class RunConfig(object):
self._strategy = DefaultStrategy()
else:
self._strategy = strategy
self._enable_memory_optim = enable_memory_optim
if enable_memory_optim:
logger.warning(
"The memory optimization feature has been dropped! PaddleHub now doesn't optimize the memory of the program."
)
self._enable_memory_optim = False
if checkpoint_dir is None:
now = int(time.time())
time_str = time.strftime("%Y%m%d%H%M%S", time.localtime(now))
......
......@@ -134,10 +134,6 @@ class BasicTask(object):
self.exe = fluid.Executor(place=self.place)
self.build_strategy = fluid.BuildStrategy()
if self.config.enable_memory_optim:
self.build_strategy.memory_optimize = True
else:
self.build_strategy.memory_optimize = False
# log item
if not os.path.exists(self.config.checkpoint_dir):
......@@ -281,8 +277,6 @@ class BasicTask(object):
share_vars_from = self._base_compiled_program
if not self.config.use_data_parallel:
if self.config.enable_memory_optim:
fluid.memory_optimize(self.env.main_program)
self.env.main_program_compiled = None
else:
self.env.main_program_compiled = fluid.CompiledProgram(
......
......@@ -13,5 +13,5 @@
# See the License for the specific language governing permissions and
# limitations under the License.
""" PaddleHub version string """
hub_version = "1.3.0"
hub_version = "1.4.0"
module_proto_version = "1.0.0"
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册