From 359c4e69bbd73d8cf7fb52bba1e498872c6d9331 Mon Sep 17 00:00:00 2001 From: wuzewu Date: Wed, 21 Oct 2020 20:05:05 +0800 Subject: [PATCH] Fix incorrect references --- paddlehub/compat/module/nlp_module.py | 10 +++++----- paddlehub/compat/paddle_utils.py | 22 +++++++++++----------- paddlehub/compat/task/base_task.py | 8 ++++---- 3 files changed, 20 insertions(+), 20 deletions(-) diff --git a/paddlehub/compat/module/nlp_module.py b/paddlehub/compat/module/nlp_module.py index e1e60255..d8f37f7b 100644 --- a/paddlehub/compat/module/nlp_module.py +++ b/paddlehub/compat/module/nlp_module.py @@ -50,10 +50,10 @@ class NLPBaseModule(RunModule): class NLPPredictionModule(NLPBaseModule): def _set_config(self): '''predictor config setting''' - cpu_config = paddle.device.core.AnalysisConfig(self.pretrained_model_path) + cpu_config = paddle.fluid.core.AnalysisConfig(self.pretrained_model_path) cpu_config.disable_glog_info() cpu_config.disable_gpu() - self.cpu_predictor = paddle.device.core.create_paddle_predictor(cpu_config) + self.cpu_predictor = paddle.fluid.core.create_paddle_predictor(cpu_config) try: _places = os.environ['CUDA_VISIBLE_DEVICES'] @@ -62,10 +62,10 @@ class NLPPredictionModule(NLPBaseModule): except: use_gpu = False if use_gpu: - gpu_config = paddle.device.core.AnalysisConfig(self.pretrained_model_path) + gpu_config = paddle.fluid.core.AnalysisConfig(self.pretrained_model_path) gpu_config.disable_glog_info() gpu_config.enable_use_gpu(memory_pool_init_size_mb=500, device_id=0) - self.gpu_predictor = paddle.device.core.create_paddle_predictor(gpu_config) + self.gpu_predictor = paddle.fluid.core.create_paddle_predictor(gpu_config) def texts2tensor(self, texts: List[dict]) -> paddle.Tensor: ''' @@ -81,7 +81,7 @@ class NLPPredictionModule(NLPBaseModule): for i, text in enumerate(texts): data += text['processed'] lod.append(len(text['processed']) + lod[i]) - tensor = paddle.device.core.PaddleTensor(np.array(data).astype('int64')) + tensor = paddle.fluid.core.PaddleTensor(np.array(data).astype('int64')) tensor.name = 'words' tensor.lod = [lod] tensor.shape = [lod[-1], 1] diff --git a/paddlehub/compat/paddle_utils.py b/paddlehub/compat/paddle_utils.py index 809ff5b9..39d05393 100644 --- a/paddlehub/compat/paddle_utils.py +++ b/paddlehub/compat/paddle_utils.py @@ -21,20 +21,20 @@ import paddle from paddlehub.utils.utils import Version dtype_map = { - paddle.device.core.VarDesc.VarType.FP32: "float32", - paddle.device.core.VarDesc.VarType.FP64: "float64", - paddle.device.core.VarDesc.VarType.FP16: "float16", - paddle.device.core.VarDesc.VarType.INT32: "int32", - paddle.device.core.VarDesc.VarType.INT16: "int16", - paddle.device.core.VarDesc.VarType.INT64: "int64", - paddle.device.core.VarDesc.VarType.BOOL: "bool", - paddle.device.core.VarDesc.VarType.INT16: "int16", - paddle.device.core.VarDesc.VarType.UINT8: "uint8", - paddle.device.core.VarDesc.VarType.INT8: "int8", + paddle.fluid.core.VarDesc.VarType.FP32: "float32", + paddle.fluid.core.VarDesc.VarType.FP64: "float64", + paddle.fluid.core.VarDesc.VarType.FP16: "float16", + paddle.fluid.core.VarDesc.VarType.INT32: "int32", + paddle.fluid.core.VarDesc.VarType.INT16: "int16", + paddle.fluid.core.VarDesc.VarType.INT64: "int64", + paddle.fluid.core.VarDesc.VarType.BOOL: "bool", + paddle.fluid.core.VarDesc.VarType.INT16: "int16", + paddle.fluid.core.VarDesc.VarType.UINT8: "uint8", + paddle.fluid.core.VarDesc.VarType.INT8: "int8", } -def convert_dtype_to_string(dtype: str) -> paddle.device.core.VarDesc.VarType: +def convert_dtype_to_string(dtype: str) -> paddle.fluid.core.VarDesc.VarType: if dtype in dtype_map: return dtype_map[dtype] raise TypeError("dtype shoule in %s" % list(dtype_map.keys())) diff --git a/paddlehub/compat/task/base_task.py b/paddlehub/compat/task/base_task.py index 37167b58..b5cef5a9 100644 --- a/paddlehub/compat/task/base_task.py +++ b/paddlehub/compat/task/base_task.py @@ -621,7 +621,7 @@ class BaseTask(object): self._eval_end_event(run_states) return run_states - def _create_predictor(self) -> paddle.device.core.PaddlePredictor: + def _create_predictor(self) -> paddle.fluid.core.PaddlePredictor: ''' create high-performance predictor for predict. Returns: @@ -629,7 +629,7 @@ class BaseTask(object): ''' with generate_tempdir() as _dir: self.save_inference_model(dirname=_dir) - predictor_config = paddle.device.core.AnalysisConfig(_dir) + predictor_config = paddle.fluid.core.AnalysisConfig(_dir) predictor_config.disable_glog_info() if self.config.use_cuda: @@ -638,7 +638,7 @@ class BaseTask(object): else: predictor_config.disable_gpu() predictor_config.enable_memory_optim() - return paddle.device.core.create_paddle_predictor(predictor_config) + return paddle.fluid.core.create_paddle_predictor(predictor_config) def _run_with_predictor(self) -> List[RunState]: ''' @@ -671,7 +671,7 @@ class BaseTask(object): tensor_batch = [[] for i in range(len(self.feed_list))] for i in range(len(processed_batch)): processed_batch[i] = np.array(processed_batch[i]).reshape(feed_var_shape[i]).astype(feed_var_type[i]) - tensor_batch[i] = paddle.device.core.PaddleTensor(processed_batch[i]) + tensor_batch[i] = paddle.fluid.core.PaddleTensor(processed_batch[i]) fetch_result = self._predictor.run(tensor_batch) for index, result in enumerate(fetch_result): -- GitLab