getting_started_en.md 11.3 KB
Newer Older
1 2
# Getting Started
---
T
Tingquan Gao 已提交
3
Please refer to [Installation](install_en.md) to setup environment at first, and prepare flower102 dataset by following the instruction mentioned in the [Quick Start](quick_start_en.md).
4

T
Tingquan Gao 已提交
5
## 1. Training and Evaluation on CPU or Single GPU
6

T
Tingquan Gao 已提交
7 8
If training and evaluation are performed on CPU or single GPU, it is recommended to use the `tools/train.py` and `tools/eval.py`.
For training and evaluation in multi-GPU environment on Linux, please refer to [2. Training and evaluation on Linux+GPU](#2-training-and-evaluation-on-linuxgpu).
9

T
Tingquan Gao 已提交
10
<a name="1.1"></a>
11 12 13 14 15
## 1.1 Model training

After preparing the configuration file, The training process can be started in the following way.

```
T
Tingquan Gao 已提交
16 17 18 19
python tools/train.py \
    -c configs/quick_start/MobileNetV3_large_x1_0_finetune.yaml \
    -o pretrained_model="" \
    -o use_gpu=False
20 21
```

T
Tingquan Gao 已提交
22 23
Among them, `-c` is used to specify the path of the configuration file, `-o` is used to specify the parameters needed to be modified or added, `-o pretrained_model=""` means to not using pre-trained models.
`-o use_gpu=True` means to use GPU for training. If you want to use the CPU for training, you need to set `use_gpu` to `False`.
24 25


T
Tingquan Gao 已提交
26
Of course, you can also directly modify the configuration file to update the configuration. For specific configuration parameters, please refer to [Configuration Document](config_en.md).
27 28 29 30 31 32 33 34 35 36 37 38 39 40

* The output log examples are as follows:
    * If mixup or cutmix is used in training, only loss, lr (learning rate) and training time of the minibatch will be printed in the log.

    ```
    train step:890  loss:  6.8473 lr: 0.100000 elapse: 0.157s
    ```

    * If mixup or cutmix is not used during training, in addition to loss, lr (learning rate) and the training time of the minibatch, top-1 and top-k( The default is 5) will also be printed in the log.

    ```
    epoch:0    train    step:13    loss:7.9561    top1:0.0156    top5:0.1094    lr:0.100000    elapse:0.193s
    ```

T
Tingquan Gao 已提交
41
During training, you can view loss changes in real time through `VisualDL`,  see [VisualDL](https://github.com/PaddlePaddle/VisualDL) for details.
42 43 44

### 1.2 Model finetuning

T
Tingquan Gao 已提交
45
After configuring the configuration file, you can finetune it by loading the pretrained weights, The command is as shown below.
46 47

```
T
Tingquan Gao 已提交
48 49 50 51
python tools/train.py \
    -c configs/quick_start/MobileNetV3_large_x1_0_finetune.yaml \
    -o pretrained_model="./pretrained/MobileNetV3_large_x1_0_pretrained" \
    -o use_gpu=True
52 53
```

T
Tingquan Gao 已提交
54 55 56
Among them, `-o pretrained_model` is used to set the address to load the pretrained weights. When using it, you need to replace it with your own pretrained weights' path, or you can modify the path directly in the configuration file.

We also provide a lot of pre-trained models trained on the ImageNet-1k dataset. For the model list and download address, please refer to the [model library overview](../models/models_intro_en.md).
57 58 59

### 1.3 Resume Training

T
Tingquan Gao 已提交
60
If the training process is terminated for some reasons, you can also load the checkpoints to continue training.
61 62

```
T
Tingquan Gao 已提交
63 64 65 66 67
python tools/train.py \
    -c configs/quick_start/MobileNetV3_large_x1_0_finetune.yaml \
    -o checkpoints="./output/MobileNetV3_large_x1_0_gpupaddle/0/ppcls" \
    -o last_epoch=5 \
    -o use_gpu=True
68 69
```

T
Tingquan Gao 已提交
70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89
The configuration file does not need to be modified. You only need to add the `checkpoints` parameter during training, which represents the path of the checkpoints. The parameter weights, learning rate, optimizer and other information will be loaded using this parameter.

**Note**:
* The parameter `-o last_epoch=5` means to record the number of the last training epoch as `5`, that is, the number of this training epoch starts from `6`, , and the parameter defaults to `-1`, which means the number of this training epoch starts from `0`.

* The `-o checkpoints` parameter does not need to include the suffix of the checkpoints. The above training command will generate the checkpoints as shown below during the training process. If you want to continue training from the epoch `0`, Just set the `checkpoints` to `./output/MobileNetV3_large_x1_0_gpupaddle/0/ppcls`, PaddleClas will automatically fill in the `pdopt` and `pdparams` suffixes.

    ```shell
    output/
    └── MobileNetV3_large_x1_0
        ├── 0
        │   ├── ppcls.pdopt
        │   └── ppcls.pdparams
        ├── 1
        │   ├── ppcls.pdopt
        │   └── ppcls.pdparams
        .
        .
        .
    ```
90 91 92 93


### 1.4 Model evaluation

T
Tingquan Gao 已提交
94
The model evaluation process can be started as follows.
95 96

```bash
T
Tingquan Gao 已提交
97
python tools/eval.py \
98
    -c ./configs/eval.yaml \
T
Tingquan Gao 已提交
99
    -o ARCHITECTURE.name="MobileNetV3_large_x1_0" \
100 101 102
    -o pretrained_model=path_to_pretrained_models
```

T
Tingquan Gao 已提交
103
You can modify the `ARCHITECTURE.name` field and `pretrained_model` field in `configs/eval.yaml` to configure the evaluation model, and you also can update the configuration through the `-o` parameter.
104 105


T
Tingquan Gao 已提交
106
**Note:** When loading the pre-trained model, you need to specify the prefix of the pretrained model, such as  [1.3 Resume Training](#13-resume-training).
107

T
Tingquan Gao 已提交
108
<a name="2"></a>
109 110
### 2. Training and evaluation on Linux+GPU

T
Tingquan Gao 已提交
111
If you want to run PaddleClas on Linux with GPU, it is highly recommended to use `paddle.distributed.launch` to start the model training script(`tools/train.py`) and evaluation script(`tools/eval.py`), which can start on multi-GPU environment more conveniently.
112 113 114

### 2.1 Model training

T
Tingquan Gao 已提交
115
After preparing the configuration file, The training process can be started in the following way. `paddle.distributed.launch` specifies the GPU running card number by setting `selected_gpus`:
116 117

```bash
T
Tingquan Gao 已提交
118 119
export CUDA_VISIBLE_DEVICES=0,1,2,3

120 121 122
python -m paddle.distributed.launch \
    --selected_gpus="0,1,2,3" \
    tools/train.py \
T
Tingquan Gao 已提交
123
        -c ./configs/quick_start/MobileNetV3_large_x1_0_finetune.yaml \
124 125 126 127 128 129 130 131
```

The configuration can be updated by adding the `-o` parameter.

```bash
python -m paddle.distributed.launch \
    --selected_gpus="0,1,2,3" \
    tools/train.py \
T
Tingquan Gao 已提交
132 133 134
        -c ./configs/quick_start/MobileNetV3_large_x1_0_finetune.yaml \
        -o pretrained_model="" \
        -o use_gpu=True
135 136
```

T
Tingquan Gao 已提交
137
The format of output log information is the same as above, see [1.1 Model training](#11-model-training) for details.
138 139 140

### 2.2 Model finetuning

T
Tingquan Gao 已提交
141
After configuring the configuration file, you can finetune it by loading the pretrained weights, The command is as shown below.
142 143

```
T
Tingquan Gao 已提交
144 145
export CUDA_VISIBLE_DEVICES=0,1,2,3

146 147 148 149 150 151 152 153 154
python -m paddle.distributed.launch \
    --selected_gpus="0,1,2,3" \
    tools/train.py \
        -c configs/ResNet/ResNet50.yaml \
        -o pretrained_model="./pretrained/ResNet50_pretrained"
```

Among them, `pretrained_model` is used to set the address to load the pretrained weights. When using it, you need to replace it with your own pretrained weights' path, or you can modify the path directly in the configuration file.

T
Tingquan Gao 已提交
155
There contains a lot of examples of model finetuning in [Quick Start](./quick_start_en.md). You can refer to this tutorial to finetune the model on a specific dataset.
156

T
Tingquan Gao 已提交
157
<a name="model_resume"></a>
158 159
### 2.3 Resume Training

T
Tingquan Gao 已提交
160
If the training process is terminated for some reasons, you can also load the checkpoints to continue training.
161 162

```
T
Tingquan Gao 已提交
163 164
export CUDA_VISIBLE_DEVICES=0,1,2,3

165 166 167
python -m paddle.distributed.launch \
    --selected_gpus="0,1,2,3" \
    tools/train.py \
T
Tingquan Gao 已提交
168
        -c ./configs/quick_start/MobileNetV3_large_x1_0_finetune.yaml \
169
        -o checkpoints="./output/ResNet/0/ppcls"
T
Tingquan Gao 已提交
170 171
        -o last_epoch=5 \
        -o use_gpu=True
172 173
```

T
Tingquan Gao 已提交
174
The configuration file does not need to be modified. You only need to add the `checkpoints` parameter during training, which represents the path of the checkpoints. The parameter weights, learning rate, optimizer and other information will be loaded using this parameter. About `last_epoch` parameter, please refer [1.3 Resume training](#13-resume-training) for details.
175 176 177

### 2.4 Model evaluation

T
Tingquan Gao 已提交
178
The model evaluation process can be started as follows.
179 180

```bash
T
Tingquan Gao 已提交
181
python tools/eval.py \
182
    -c ./configs/eval.yaml \
T
Tingquan Gao 已提交
183
    -o ARCHITECTURE.name="MobileNetV3_large_x1_0" \
184 185 186
    -o pretrained_model=path_to_pretrained_models
```

T
Tingquan Gao 已提交
187
You can modify the `ARCHITECTURE.name` field and `pretrained_model` field in `configs/eval.yaml` to configure the evaluation model, and you also can update the configuration through the `-o` parameter.
188

T
Tingquan Gao 已提交
189 190 191
<a name="model_infer"></a>
## 3. Use the pre-trained model to predict
After the training is completed, you can predict by using the pre-trained model obtained by the training, as follows:
192

T
Tingquan Gao 已提交
193 194 195 196 197 198 199 200
```python
python tools/infer/infer.py \
    --i image path \
    --m model name \
    --p path to persistence model \
    --use_gpu True \
    --load_static_weights False
```
201

T
Tingquan Gao 已提交
202 203 204 205 206 207 208 209 210 211 212 213 214 215 216
Among them:
+ `image_file`(i): The path of the image file to be predicted, such as `./test.jpeg`;
+ `model`(m): Model name, such as `MobileNetV3_large_x1_0`;
+ `pretrained_model`(p): Weight file path, such as `./pretrained/MobileNetV3_large_x1_0_pretrained/`;
+ `use_gpu`: Whether to use the GPU, default by `True`;
+ `load_static_weights`: Whether to load the pre-trained model obtained from static image training, default by `False`;
+ `pre_label_image`: Whether to pre-label the image data, default value: `False`;
+ `pre_label_out_idr`: The output path of pre-labeled image data. When `pre_label_image=True`, a lot of subfolders will be generated under the path, each subfolder represent a category, which stores all the images predicted by the model to belong to the category.

About more detailed infomation, you can refer to [infer.py](../../../tools/infer/infer.py).

<a name="model_inference"></a>
## 4. Use the inference model to predict

PaddlePaddle supports inference using prediction engines, which will be introduced next.
217 218 219 220 221

Firstly, you should export inference model using `tools/export_model.py`.

```bash
python tools/export_model.py \
T
Tingquan Gao 已提交
222 223 224 225 226 227 228 229
    --model=MobileNetV3_large_x1_0 \
    --pretrained_model=./output/MobileNetV3_large_x1_0/best_model/ppcls \
    --output_path=./exported_model

```
Among them, the `--model` parameter is used to specify the model name, `--pretrained_model` parameter is used to specify the model file path, the path does not need to include the model file suffix name, and `--output_path` is used to specify the storage path of the converted model .

**Note**: In the file `export_model.py:line53`, the `shape` parameter is the shape of the model input image, the default is `224*224`. Please modify it according to the actual situation, as shown below:
230

T
Tingquan Gao 已提交
231 232 233 234 235 236
```python
50 # Please modify the 'shape' according to actual needs
51 @to_static(input_spec=[
52     paddle.static.InputSpec(
53         shape=[None, 3, 224, 224], dtype='float32')
54 ])
237 238
```

T
Tingquan Gao 已提交
239
The above command will generate the model structure file (`__model__`) and the model weight file (`__variables__`), and then the inference engine can be used for inference:
240 241 242 243

```bash
python tools/infer/predict.py \
    -i image path \
T
Tingquan Gao 已提交
244 245
    -m path of __model__ \
    -p path of __variables__ \
246 247 248
    --use_gpu=1 \
    --use_tensorrt=True
```
T
Tingquan Gao 已提交
249 250 251 252 253 254 255 256
Among them:
+ `image_file`(i): The path of the image file to be predicted, such as `./test.jpeg`;
+ `model_file`(m): Model file path, such as `./MobileNetV3_large_x1_0/__model__`;
+ `params_file`(p): Weight file path, such as `./MobileNetV3_large_x1_0/__variables__`;
+ `use_tensorrt`: Whether to use the TesorRT, default by `True`;
+ `use_gpu`: Whether to use the GPU, default by `True`.

If you want to evaluate the speed of the model, it is recommended to use [predict.py](../../../tools/infer/predict.py), and enable TensorRT to accelerate.