@@ -163,7 +163,7 @@ In the annotation text, each line is describing a bounding box and has the follo
...
@@ -163,7 +163,7 @@ In the annotation text, each line is describing a bounding box and has the follo
```
```
**Notes:**
**Notes:**
-`class` should be `0`. Only single-class multi-object tracking is supported now.
-`class` should be `0`. Only single-class multi-object tracking is supported now.
-`identity` is an integer from `0` to `num_identities - 1`(`num_identities` is the total number of instances of objects in the dataset), or `-1` if this box has no identity annotation.
-`identity` is an integer from `1` to `num_identities`(`num_identities` is the total number of instances of objects in the dataset), or `-1` if this box has no identity annotation.
-`[x_center] [y_center] [width] [height]` are the center coordinates, width and height, note that they are normalized by the width/height of the image, so they are floating point numbers ranging from 0 to 1.
-`[x_center] [y_center] [width] [height]` are the center coordinates, width and height, note that they are normalized by the width/height of the image, so they are floating point numbers ranging from 0 to 1.
python deploy/python/mot_infer.py --model_dir=output_inference/fairmot_dla34_30e_1088x608 --video_file={your video name}.mp4 --device=GPU --save_mot_txts
python deploy/python/mot_infer.py --model_dir=output_inference/fairmot_dla34_30e_1088x608 --video_file={your video name}.mp4 --device=GPU --save_mot_txts
```
```
**Notes:**
**Notes:**
The tracking model is used to predict the video, and does not support the prediction of a single image. The visualization video of the tracking results is saved by default. You can add `--save_mot_txts` to save the txt result file, or `--save_images` to save the visualization images.
The tracking model is used to predict the video, and does not support the prediction of a single image. The visualization video of the tracking results is saved by default. You can add `--save_mot_txts` to save the txt result file, or `--save_images` to save the visualization images.
### 6. Using exported MOT and keypoint model for unite python inference
### 6. Using exported MOT and keypoint model for unite python inference
...
@@ -265,7 +265,7 @@ The tracking model is used to predict the video, and does not support the predic
...
@@ -265,7 +265,7 @@ The tracking model is used to predict the video, and does not support the predic
```bash
```bash
python deploy/python/mot_keypoint_unite_infer.py --mot_model_dir=output_inference/fairmot_dla34_30e_1088x608/ --keypoint_model_dir=output_inference/higherhrnet_hrnet_w32_512/ --video_file={your video name}.mp4 --device=GPU
python deploy/python/mot_keypoint_unite_infer.py --mot_model_dir=output_inference/fairmot_dla34_30e_1088x608/ --keypoint_model_dir=output_inference/higherhrnet_hrnet_w32_512/ --video_file={your video name}.mp4 --device=GPU
```
```
**Notes:**
**Notes:**
Keypoint model export tutorial: `configs/keypoint/README.md`.
Keypoint model export tutorial: `configs/keypoint/README.md`.
@@ -39,7 +39,7 @@ In the annotation text, each line is describing a bounding box and has the follo
...
@@ -39,7 +39,7 @@ In the annotation text, each line is describing a bounding box and has the follo
```
```
**Notes:**
**Notes:**
-`class` should be `0`. Only single-class multi-object tracking is supported now.
-`class` should be `0`. Only single-class multi-object tracking is supported now.
-`identity` is an integer from `0` to `num_identities - 1`(`num_identities` is the total number of instances of objects in the dataset), or `-1` if this box has no identity annotation.
-`identity` is an integer from `1` to `num_identities`(`num_identities` is the total number of instances of objects in the dataset), or `-1` if this box has no identity annotation.
-`[x_center] [y_center] [width] [height]` are the center coordinates, width and height, note that they are normalized by the width/height of the image, so they are floating point numbers ranging from 0 to 1.
-`[x_center] [y_center] [width] [height]` are the center coordinates, width and height, note that they are normalized by the width/height of the image, so they are floating point numbers ranging from 0 to 1.
...
@@ -124,16 +124,18 @@ imExt=.jpg
...
@@ -124,16 +124,18 @@ imExt=.jpg
Each line in `gt.txt` describes a bounding box, with the format as follows:
Each line in `gt.txt` describes a bounding box, with the format as follows:
-`identity` is an integer from `0` to `num_identities - 1`(`num_identities` is the total number of instances of objects in the dataset), or `-1` if this box has no identity annotation.
-`identity` is an integer from `1` to `num_identities`(`num_identities` is the total number of instances of objects in the dataset), or `-1` if this box has no identity annotation.
-`bb_left` is the x coordinate of the left boundary of the target box
-`bb_left` is the x coordinate of the left boundary of the target box
-`bb_top` is the Y coordinate of the upper boundary of the target box
-`bb_top` is the Y coordinate of the upper boundary of the target box
-`width, height` are the pixel width and height
-`width, height` are the pixel width and height
-`x,y,z` are only used in 3D, default to `-1` in 2D.
-`score` acts as a flag whether the entry is to be considered. A value of 0 means that this particular instance is ignored in the evaluation, while a value of 1 is used to mark it as active. `1` by default.
-`label` is the type of object annotated, use `1` as default because only single-class multi-object tracking is supported now. There are other classes of object in MOT-16, but they are treated as ignore.
-`vis_ratio` is the visibility ratio of each bounding box. This can be due to occlusion by another
static or moving object, or due to image border cropping. `1` by default.
#### labels_with_ids
#### labels_with_ids
Annotations of these datasets are provided in a unified format. Every image has a corresponding annotation text. Given an image path, the annotation text path can be generated by replacing the string `images` with `labels_with_ids` and replacing `.jpg` with `.txt`.
Annotations of these datasets are provided in a unified format. Every image has a corresponding annotation text. Given an image path, the annotation text path can be generated by replacing the string `images` with `labels_with_ids` and replacing `.jpg` with `.txt`.
...
@@ -144,7 +146,7 @@ In the annotation text, each line is describing a bounding box and has the follo
...
@@ -144,7 +146,7 @@ In the annotation text, each line is describing a bounding box and has the follo
```
```
**Notes:**
**Notes:**
-`class` should be `0`. Only single-class multi-object tracking is supported now.
-`class` should be `0`. Only single-class multi-object tracking is supported now.
-`identity` is an integer from `0` to `num_identities - 1`(`num_identities` is the total number of instances of objects in the dataset), or `-1` if this box has no identity annotation.
-`identity` is an integer from `1` to `num_identities`(`num_identities` is the total number of instances of objects in the dataset), or `-1` if this box has no identity annotation.
-`[x_center] [y_center] [width] [height]` are the center coordinates, width and height, note that they are normalized by the width/height of the image, so they are floating point numbers ranging from 0 to 1.
-`[x_center] [y_center] [width] [height]` are the center coordinates, width and height, note that they are normalized by the width/height of the image, so they are floating point numbers ranging from 0 to 1.
Generate the corresponding `labels_with_ids` with following command:
Generate the corresponding `labels_with_ids` with following command: