diff --git a/fluid/image_classification/dist_train/README.md b/fluid/image_classification/dist_train/README.md index 02bbea17f423fe2e16fd3115058ba92805a313ab..282a026acf1ee6d5b1c17aa05a2a8f734047c006 100644 --- a/fluid/image_classification/dist_train/README.md +++ b/fluid/image_classification/dist_train/README.md @@ -52,7 +52,7 @@ In this example, we launched 4 parameter server instances and 4 trainer instance 1. launch trainer process ``` python - PADDLE_TRAINING_ROLE=PSERVER \ + PADDLE_TRAINING_ROLE=TRAINER \ PADDLE_TRAINERS=4 \ PADDLE_PSERVER_IPS=192.168.0.100,192.168.0.101,192.168.0.102,192.168.0.103 \ PADDLE_TRAINER_ID=0 \ @@ -110,4 +110,4 @@ Training acc1 curves ### Performance -TBD \ No newline at end of file +TBD diff --git a/fluid/image_classification/train.py b/fluid/image_classification/train.py index bfc5f8b1412a11606d54b020f29bef969bae2a62..238c322bea9abf1ce086a0228b491e82cb69ae45 100644 --- a/fluid/image_classification/train.py +++ b/fluid/image_classification/train.py @@ -33,7 +33,7 @@ add_arg('lr', float, 0.1, "set learning rate.") add_arg('lr_strategy', str, "piecewise_decay", "Set the learning rate decay strategy.") add_arg('model', str, "SE_ResNeXt50_32x4d", "Set the network to use.") add_arg('enable_ce', bool, False, "If set True, enable continuous evaluation job.") -add_arg('data_dir' str, "./data/ILSVRC2012", "The ImageNet dataset root dir.") +add_arg('data_dir', str, "./data/ILSVRC2012", "The ImageNet dataset root dir.") # yapf: enable model_list = [m for m in dir(models) if "__" not in m]