diff --git a/test_tipc/config/VisionTransformer/ViT_huge_patch16_224_train_infer_python.txt b/test_tipc/config/VisionTransformer/ViT_huge_patch16_224_train_infer_python.txt deleted file mode 100644 index 8b83f4d59266694badc0292d053151c82c9a56fd..0000000000000000000000000000000000000000 --- a/test_tipc/config/VisionTransformer/ViT_huge_patch16_224_train_infer_python.txt +++ /dev/null @@ -1,52 +0,0 @@ -===========================train_params=========================== -model_name:ViT_huge_patch16_224 -python:python3.7 -gpu_list:0|0,1 --o Global.device:gpu --o Global.auto_cast:null --o Global.epochs:lite_train_lite_infer=2|whole_train_whole_infer=120 --o Global.output_dir:./output/ --o DataLoader.Train.sampler.batch_size:8 --o Global.pretrained_model:null -train_model_name:latest -train_infer_img_dir:./dataset/ILSVRC2012/val -null:null -## -trainer:norm_train -norm_train:tools/train.py -c ppcls/configs/ImageNet/VisionTransformer/ViT_huge_patch16_224.yaml -o Global.seed=1234 -o DataLoader.Train.sampler.shuffle=False -o DataLoader.Train.loader.num_workers=0 -o DataLoader.Train.loader.use_shared_memory=False -pact_train:null -fpgm_train:null -distill_train:null -null:null -null:null -## -===========================eval_params=========================== -eval:tools/eval.py -c ppcls/configs/ImageNet/VisionTransformer/ViT_huge_patch16_224.yaml -null:null -## -===========================infer_params========================== --o Global.save_inference_dir:./inference --o Global.pretrained_model: -norm_export:tools/export_model.py -c ppcls/configs/ImageNet/VisionTransformer/ViT_huge_patch16_224.yaml -quant_export:null -fpgm_export:null -distill_export:null -kl_quant:null -export2:null -pretrained_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/ViT_huge_patch16_224_pretrained.pdparams -infer_model:../inference/ -infer_export:True -infer_quant:Fasle -inference:python/predict_cls.py -c configs/inference_cls.yaml --o Global.use_gpu:True|False --o Global.enable_mkldnn:True|False --o Global.cpu_num_threads:1|6 --o Global.batch_size:1|16 --o Global.use_tensorrt:True|False --o Global.use_fp16:True|False --o Global.inference_model_dir:../inference --o Global.infer_imgs:../dataset/ILSVRC2012/val --o Global.save_log_path:null --o Global.benchmark:True -null:null -null:null diff --git a/test_tipc/config/VisionTransformer/ViT_huge_patch32_384_train_infer_python.txt b/test_tipc/config/VisionTransformer/ViT_huge_patch32_384_train_infer_python.txt deleted file mode 100644 index 942302f2cd170a2f4c94f9051e28bb11c0740cb9..0000000000000000000000000000000000000000 --- a/test_tipc/config/VisionTransformer/ViT_huge_patch32_384_train_infer_python.txt +++ /dev/null @@ -1,52 +0,0 @@ -===========================train_params=========================== -model_name:ViT_huge_patch32_384 -python:python3.7 -gpu_list:0|0,1 --o Global.device:gpu --o Global.auto_cast:null --o Global.epochs:lite_train_lite_infer=2|whole_train_whole_infer=120 --o Global.output_dir:./output/ --o DataLoader.Train.sampler.batch_size:8 --o Global.pretrained_model:null -train_model_name:latest -train_infer_img_dir:./dataset/ILSVRC2012/val -null:null -## -trainer:norm_train -norm_train:tools/train.py -c ppcls/configs/ImageNet/VisionTransformer/ViT_huge_patch32_384.yaml -o Global.seed=1234 -o DataLoader.Train.sampler.shuffle=False -o DataLoader.Train.loader.num_workers=0 -o DataLoader.Train.loader.use_shared_memory=False -pact_train:null -fpgm_train:null -distill_train:null -null:null -null:null -## -===========================eval_params=========================== -eval:tools/eval.py -c ppcls/configs/ImageNet/VisionTransformer/ViT_huge_patch32_384.yaml -null:null -## -===========================infer_params========================== --o Global.save_inference_dir:./inference --o Global.pretrained_model: -norm_export:tools/export_model.py -c ppcls/configs/ImageNet/VisionTransformer/ViT_huge_patch32_384.yaml -quant_export:null -fpgm_export:null -distill_export:null -kl_quant:null -export2:null -pretrained_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/ViT_huge_patch32_384_pretrained.pdparams -infer_model:../inference/ -infer_export:True -infer_quant:Fasle -inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 --o Global.use_gpu:True|False --o Global.enable_mkldnn:True|False --o Global.cpu_num_threads:1|6 --o Global.batch_size:1|16 --o Global.use_tensorrt:True|False --o Global.use_fp16:True|False --o Global.inference_model_dir:../inference --o Global.infer_imgs:../dataset/ILSVRC2012/val --o Global.save_log_path:null --o Global.benchmark:True -null:null -null:null