diff --git a/python/examples/pipeline/PaddleClas/DarkNet53/benchmark.py b/python/examples/pipeline/PaddleClas/DarkNet53/benchmark.py index 71b5219441a536789e02e4549c84a5cd550bc70f..3e5db19b69fc8693adfe77a84297436bfb497642 100644 --- a/python/examples/pipeline/PaddleClas/DarkNet53/benchmark.py +++ b/python/examples/pipeline/PaddleClas/DarkNet53/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/HRNet_W18_C/benchmark.py b/python/examples/pipeline/PaddleClas/HRNet_W18_C/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/HRNet_W18_C/benchmark.py +++ b/python/examples/pipeline/PaddleClas/HRNet_W18_C/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/MobileNetV1/benchmark.py b/python/examples/pipeline/PaddleClas/MobileNetV1/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/MobileNetV1/benchmark.py +++ b/python/examples/pipeline/PaddleClas/MobileNetV1/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/MobileNetV2/benchmark.py b/python/examples/pipeline/PaddleClas/MobileNetV2/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/MobileNetV2/benchmark.py +++ b/python/examples/pipeline/PaddleClas/MobileNetV2/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/MobileNetV3_large_x1_0/benchmark.py b/python/examples/pipeline/PaddleClas/MobileNetV3_large_x1_0/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/MobileNetV3_large_x1_0/benchmark.py +++ b/python/examples/pipeline/PaddleClas/MobileNetV3_large_x1_0/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/ResNeXt101_vd_64x4d/benchmark.py b/python/examples/pipeline/PaddleClas/ResNeXt101_vd_64x4d/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/ResNeXt101_vd_64x4d/benchmark.py +++ b/python/examples/pipeline/PaddleClas/ResNeXt101_vd_64x4d/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/ResNet50_vd/benchmark.py b/python/examples/pipeline/PaddleClas/ResNet50_vd/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/ResNet50_vd/benchmark.py +++ b/python/examples/pipeline/PaddleClas/ResNet50_vd/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/ResNet50_vd_FPGM/benchmark.py b/python/examples/pipeline/PaddleClas/ResNet50_vd_FPGM/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/ResNet50_vd_FPGM/benchmark.py +++ b/python/examples/pipeline/PaddleClas/ResNet50_vd_FPGM/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/ResNet50_vd_KL/benchmark.py b/python/examples/pipeline/PaddleClas/ResNet50_vd_KL/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/ResNet50_vd_KL/benchmark.py +++ b/python/examples/pipeline/PaddleClas/ResNet50_vd_KL/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/ResNet50_vd_PACT/benchmark.py b/python/examples/pipeline/PaddleClas/ResNet50_vd_PACT/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/ResNet50_vd_PACT/benchmark.py +++ b/python/examples/pipeline/PaddleClas/ResNet50_vd_PACT/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/ResNet_V2_50/benchmark.py b/python/examples/pipeline/PaddleClas/ResNet_V2_50/benchmark.py index 562d159da3ce96233f7f9d2019fbb3061022dc06..4b0336f97c2c520a46d596bf5e435c2b9e3094a9 100644 --- a/python/examples/pipeline/PaddleClas/ResNet_V2_50/benchmark.py +++ b/python/examples/pipeline/PaddleClas/ResNet_V2_50/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleClas/ShuffleNetV2_x1_0/benchmark.py b/python/examples/pipeline/PaddleClas/ShuffleNetV2_x1_0/benchmark.py index 90a3ff9bdda545a01427a26146edcbdf8332da30..c80da12ce36618e75897b33d58e4f4febd382861 100644 --- a/python/examples/pipeline/PaddleClas/ShuffleNetV2_x1_0/benchmark.py +++ b/python/examples/pipeline/PaddleClas/ShuffleNetV2_x1_0/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/PaddleDetection/faster_rcnn/benchmark.py b/python/examples/pipeline/PaddleDetection/faster_rcnn/benchmark.py index 8a25952cdda2e09f0f74794cf8a2226880f29040..f8d5f2b4fd196048a139867a893b06f47d2778bb 100644 --- a/python/examples/pipeline/PaddleDetection/faster_rcnn/benchmark.py +++ b/python/examples/pipeline/PaddleDetection/faster_rcnn/benchmark.py @@ -33,7 +33,7 @@ def cv2_to_base64(image): def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -46,7 +46,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 30} if device == "gpu": diff --git a/python/examples/pipeline/PaddleDetection/ppyolo_mbv3/benchmark.py b/python/examples/pipeline/PaddleDetection/ppyolo_mbv3/benchmark.py index 45853c065b013754d0d591686a9a03ad0aeb6a3d..611712b6754efd88fc7b51027e99b9bb3e82cf7d 100644 --- a/python/examples/pipeline/PaddleDetection/ppyolo_mbv3/benchmark.py +++ b/python/examples/pipeline/PaddleDetection/ppyolo_mbv3/benchmark.py @@ -33,7 +33,7 @@ def cv2_to_base64(image): def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -46,7 +46,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 30} if device == "gpu": diff --git a/python/examples/pipeline/PaddleDetection/yolov3/benchmark.py b/python/examples/pipeline/PaddleDetection/yolov3/benchmark.py index 62732613dbfc6ab0b119609a547ea36c18b11ede..cb73d2f932c12d0559af307b3ecf12ecf7986390 100644 --- a/python/examples/pipeline/PaddleDetection/yolov3/benchmark.py +++ b/python/examples/pipeline/PaddleDetection/yolov3/benchmark.py @@ -33,7 +33,7 @@ def cv2_to_base64(image): def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -46,7 +46,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device, gpu_id): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 30} if device == "gpu": diff --git a/python/examples/pipeline/bert/benchmark.py b/python/examples/pipeline/bert/benchmark.py index 5abc646bffffff118ab24414e3a50f06668729d9..ccdbbdf599943ebf757d336b96d4f19b92e1b94a 100644 --- a/python/examples/pipeline/bert/benchmark.py +++ b/python/examples/pipeline/bert/benchmark.py @@ -54,7 +54,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -67,7 +67,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/ocr/benchmark.py b/python/examples/pipeline/ocr/benchmark.py index 1e39176436b0be11093031ddfc4727ee68671c62..3c1243a1c327a5f94544c7fa56524321cad2892f 100644 --- a/python/examples/pipeline/ocr/benchmark.py +++ b/python/examples/pipeline/ocr/benchmark.py @@ -28,7 +28,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def parse_benchmark(filein, fileout): with open(filein, "r") as fin: - res = yaml.load(fin) + res = yaml.load(fin, yaml.FullLoader) del_list = [] for key in res["DAG"].keys(): if "call" in key: @@ -41,7 +41,7 @@ def parse_benchmark(filein, fileout): def gen_yml(device): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 10} if device == "gpu": diff --git a/python/examples/pipeline/simple_web_service/benchmark.py b/python/examples/pipeline/simple_web_service/benchmark.py index c2c612dd2740d7c97da4289a0913270b03611e7a..88c3ea21722ad9e6420e193a69299b2cf8e443a4 100644 --- a/python/examples/pipeline/simple_web_service/benchmark.py +++ b/python/examples/pipeline/simple_web_service/benchmark.py @@ -27,7 +27,7 @@ from paddle_serving_client.utils import benchmark_args, show_latency def gen_yml(): fin = open("config.yml", "r") - config = yaml.load(fin) + config = yaml.load(fin, yaml.FullLoader) fin.close() config["dag"]["tracer"] = {"interval_s": 5} with open("config2.yml", "w") as fout: diff --git a/python/paddle_serving_server/parse_profile.py b/python/paddle_serving_server/parse_profile.py index 37e801c255272778c6926642beabdcf2f3f92cf0..e718e4685e13ed35f8dba16eb0d5f8a3ff6fd305 100644 --- a/python/paddle_serving_server/parse_profile.py +++ b/python/paddle_serving_server/parse_profile.py @@ -96,7 +96,7 @@ if __name__ == "__main__": args = parse_args() benchmark_cfg_filename = args.benchmark_cfg f = open(benchmark_cfg_filename, 'r') - benchmark_config = yaml.load(f) + benchmark_config = yaml.load(f, yaml.FullLoader) f.close() benchmark_log_filename = args.benchmark_log f = open(benchmark_log_filename, 'r') diff --git a/python/pipeline/analyse.py b/python/pipeline/analyse.py index 814b43acaf52bbf0c066ff4bbdce2a0165508a2d..a571ccfe9018fac70523803c40d05df1cf16e271 100644 --- a/python/pipeline/analyse.py +++ b/python/pipeline/analyse.py @@ -274,7 +274,7 @@ class OpAnalyst(object): """ import yaml with open(op_config_yaml) as f: - op_config = yaml.load(f) + op_config = yaml.load(f, yaml.FullLoader) # check that each model is deployed on a different card card_set = set() diff --git a/python/pipeline/pipeline_server.py b/python/pipeline/pipeline_server.py index c3a904690dcd4a7a044f6afd309553fd1446aa49..5d3fa3540149412186b9335741964910a7ed56d2 100644 --- a/python/pipeline/pipeline_server.py +++ b/python/pipeline/pipeline_server.py @@ -341,7 +341,7 @@ class ServerYamlConfChecker(object): " or yml_dict can be selected as the parameter.") if yml_file is not None: with io.open(yml_file, encoding='utf-8') as f: - conf = yaml.load(f.read()) + conf = yaml.load(f.read(), yaml.FullLoader) elif yml_dict is not None: conf = yml_dict else: diff --git a/python/requirements.txt b/python/requirements.txt index 3dd93093ad3a8f3fd6ff8c3b3e41c8a84ffebd81..ba7cf42d9e0a6b4cd713ef245108bb45e7244dda 100644 --- a/python/requirements.txt +++ b/python/requirements.txt @@ -7,7 +7,7 @@ protobuf>=3.12.2 grpcio-tools>=1.28.1 grpcio>=1.28.1 func-timeout>=4.3.5 -pyyaml>=1.3.0, <6.0 +pyyaml>=5.1 flask>=1.1.2 click==7.1.2 itsdangerous==1.1.0 diff --git a/python/requirements_mac.txt b/python/requirements_mac.txt index fb289aea74257195b74ce600082f69905eafe774..6a396239c5e68e545bd5af0928b3e6f42b19c82b 100644 --- a/python/requirements_mac.txt +++ b/python/requirements_mac.txt @@ -6,7 +6,7 @@ google>=2.0.3 opencv-python==4.2.0.32 protobuf>=3.12.2 func-timeout>=4.3.5 -pyyaml>=1.3.0, <6.0 +pyyaml>=5.1 flask>=1.1.2 click==7.1.2 itsdangerous==1.1.0