diff --git a/modules/image/object_detection/faster_rcnn_resnet50_coco2017/README_en.md b/modules/image/object_detection/faster_rcnn_resnet50_coco2017/README_en.md
index 90341db0821fbb8c8a85ed5dba3e722e8477a586..8dcf806a3b88abc8cea6110937e655b4db10c68f 100644
--- a/modules/image/object_detection/faster_rcnn_resnet50_coco2017/README_en.md
+++ b/modules/image/object_detection/faster_rcnn_resnet50_coco2017/README_en.md
@@ -23,7 +23,9 @@
- ### Module Introduction
- - Faster_RCNN是两阶段目标检测器,对图像生成候选区域、提取特征、判别特征类别并修正候选框位置.Faster_RCNN整体网络可以分为4部分,一是ResNet-50作为基础卷积层,二是区域生成网络,三是Rol Align,四是检测层.Faster_RCNN是在MS-COCO数据集上预训练的模型.目前仅提供预测功能.
+ - Faster_RCNN is a two-stage detector, it consists of feature extraction, proposal, classification and refinement processes. This module is trained on COCO2017 dataset, and can be used for object detection.
+
+
## II.Installation
@@ -73,7 +75,7 @@
visualization=True)
```
- - 预测API,检测输入图片中的所有目标的位置.
+ - Detection API, detect positions of all objects in image
- **Parameters**
@@ -82,22 +84,22 @@
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- output_dir (str): save path of images;
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
**NOTE:** choose one parameter to provide data from paths and images
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
diff --git a/modules/image/object_detection/faster_rcnn_resnet50_fpn_coco2017/README_en.md b/modules/image/object_detection/faster_rcnn_resnet50_fpn_coco2017/README_en.md
index 9e925e31a6230f62bd8eaf7c97b24df9a47e17f7..7e452db4d2eda34bdd586a793ccf863a2d9dfab1 100644
--- a/modules/image/object_detection/faster_rcnn_resnet50_fpn_coco2017/README_en.md
+++ b/modules/image/object_detection/faster_rcnn_resnet50_fpn_coco2017/README_en.md
@@ -22,7 +22,7 @@
- ### Module Introduction
- - Faster_RCNN是两阶段目标检测器,对图像生成候选区域、提取特征、判别特征类别并修正候选框位置.Faster_RCNN整体网络可以分为4个部分,一是ResNet-50作为基础卷积层,二是区域生成网络,三是Rol Align,四是检测层.Faster_RCNN是在MS-COCO数据集上预训练的模型.目前仅支持预测.
+ - Faster_RCNN is a two-stage detector, it consists of feature extraction, proposal, classification and refinement processes. This module is trained on COCO2017 dataset, and can be used for object detection.
## II.Installation
@@ -73,7 +73,7 @@
visualization=True)
```
- - 预测API,检测输入图片中的所有目标的位置.
+ - Detection API, detect positions of all objects in image
- **Parameters**
@@ -82,22 +82,22 @@
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- output_dir (str): save path of images;
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
**NOTE:** choose one parameter to provide data from paths and images
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
diff --git a/modules/image/object_detection/faster_rcnn_resnet50_fpn_venus/README_en.md b/modules/image/object_detection/faster_rcnn_resnet50_fpn_venus/README_en.md
index bd8ebf093ea33d7d358ae215dc4a8feb00e54050..c58fc8f38d4b8ec970bc629e742a2c5ec58fcfdc 100644
--- a/modules/image/object_detection/faster_rcnn_resnet50_fpn_venus/README_en.md
+++ b/modules/image/object_detection/faster_rcnn_resnet50_fpn_venus/README_en.md
@@ -4,7 +4,7 @@
| :--- | :---: |
|Category|object detection|
|Network|faster_rcnn|
-|Dataset|百度自建Dataset|
+|Dataset|Baidu Detection Dataset|
|Fine-tuning supported or not|Yes|
|Module Size|317MB|
|Latest update date|2021-02-26|
@@ -15,7 +15,7 @@
- ### Module Introduction
- - Faster_RCNN是两阶段目标检测器,对图像生成候选区域、提取特征、判别特征类别并修正候选框位置.Faster_RCNN整体网络可以分为4个部分,一是ResNet-50作为基础卷积层,二是区域生成网络,三是Rol Align,四是检测层.该PaddleHub Module是由800+tag,170w图片,1000w+检测框训练的大规模通用检测模型,在8个数据集上MAP平均提升2.06%,iou=0.5的准确率平均提升1.78%.对比于其他通用检测模型,使用该Module进行finetune,可以更快收敛,达到较优效果.
+ - Faster_RCNN is a two-stage detector, it consists of feature extraction, proposal, classification and refinement processes. This module is trained on Baidu Detection Dataset, which contains 170w pictures and 1000w+ boxes, and improve the accuracy on 8 test datasets with average 2.06%. Besides, this module supports to fine-tune model, and can achieve faster convergence and better performance.
## II.Installation
@@ -44,38 +44,38 @@
phase='train')
```
- - 提取特征,用于迁移学习.
+ - Extract features, and do transfer learning
- **Parameters**
- - num\_classes (int): 类别数;
- - trainable (bool): Parameters是否可训练;
- - pretrained (bool): 是否加载预训练模型;
- - get\_prediction (bool): 可选值为 'train'/'predict','train' 用于训练,'predict' 用于预测.
+ - num\_classes (int): number of classes;
+ - trainable (bool): whether parameters trainable or not;
+ - pretrained (bool): whether load pretrained model or not
+ - get\_prediction (bool): optional, 'train' or 'predict','train' is used for training,'predict' used for prediction.
- **Return**
- - inputs (dict): 模型的输入,相应的取值为:
- 当phase为'train'时,包含:
- - image (Variable): 图像变量
- - im\_size (Variable): 图像的尺寸
- - im\_info (Variable): 图像缩放信息
- - gt\_class (Variable): 检测框类别
- - gt\_box (Variable): 检测框坐标
- - is\_crowd (Variable): 单个框内是否包含多个物体
- 当 phase 为 'predict'时,包含:
- - image (Variable): 图像变量
- - im\_size (Variable): 图像的尺寸
- - im\_info (Variable): 图像缩放信息
- - outputs (dict): 模型的输出,响应的取值为:
- 当 phase 为 'train'时,包含:
- - head_features (Variable): 所提取的特征
- - rpn\_cls\_loss (Variable): 检测框分类损失
- - rpn\_reg\_loss (Variable): 检测框回归损失
- - generate\_proposal\_labels (Variable): 图像信息
- 当 phase 为 'predict'时,包含:
- - head_features (Variable): 所提取的特征
- - rois (Variable): 提取的roi
- - bbox\_out (Variable): 预测结果
- - context\_prog (Program): 用于迁移学习的 Program
+ - inputs (dict): inputs, a dict:
+ if phase is 'train', keys are:
+ - image (Variable): image variable
+ - im\_size (Variable): image size
+ - im\_info (Variable): image information
+ - gt\_class (Variable): box class
+ - gt\_box (Variable): box coordination
+ - is\_crowd (Variable): if multiple objects in box
+ if phase 为 'predict',keys are:
+ - image (Variable): image variable
+ - im\_size (Variable): image size
+ - im\_info (Variable): image information
+ - outputs (dict): model output
+ if phase is 'train', keys are:
+ - head_features (Variable): features extracted
+ - rpn\_cls\_loss (Variable): classfication loss in box
+ - rpn\_reg\_loss (Variable): regression loss in box
+ - generate\_proposal\_labels (Variable): proposal labels
+ if phase 为 'predict',keys are:
+ - head_features (Variable): features extracted
+ - rois (Variable): roi
+ - bbox\_out (Variable): prediction results
+ - program for transfer learning
- ```python
def save_inference_model(dirname,
diff --git a/modules/image/object_detection/ssd_mobilenet_v1_pascal/README_en.md b/modules/image/object_detection/ssd_mobilenet_v1_pascal/README_en.md
index 7082dbbdbf3856f3478ce40ee4da6fdcf3903cb5..b8cf5fc9a9f9f1cc2c3482f466b79b26cf523676 100644
--- a/modules/image/object_detection/ssd_mobilenet_v1_pascal/README_en.md
+++ b/modules/image/object_detection/ssd_mobilenet_v1_pascal/README_en.md
@@ -22,7 +22,7 @@
- ### Module Introduction
- - Single Shot MultiBox Detector (SSD) 是一种单阶段的目标检测器.与两阶段的检测方法不同,单阶段目标检测并不进行区域推荐,而是直接从特征图回归出目标的边界框和分类概率.SSD 运用了这种单阶段检测的思想,并且对其进行改进:在不同尺度的特征图上检测对应尺度的目标.该PaddleHub Module的基网络为MobileNet-v1模型,在Pascal数据集上预训练得到,目前仅支持预测.
+ - Single Shot MultiBox Detector (SSD) is a one-stage detector. Different from two-stage detector, SSD frames object detection as a re- gression problem to spatially separated bounding boxes and associated class probabilities. This module is based on MobileNet-v1, trained on Pascal dataset, and can be used for object detection.
## II.Installation
@@ -73,7 +73,7 @@
)
```
- - 预测API,检测输入图片中的所有目标的位置.
+ - Detection API, detect positions of all objects in image
- **Parameters**
@@ -90,15 +90,15 @@
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
def save_inference_model(dirname,
diff --git a/modules/image/object_detection/ssd_vgg16_512_coco2017/README_en.md b/modules/image/object_detection/ssd_vgg16_512_coco2017/README_en.md
index d6e20e107d4db4313d4648309ea4e257c29eff87..efe8655009fcb46dc19f126d97f8d6b9423b66ee 100644
--- a/modules/image/object_detection/ssd_vgg16_512_coco2017/README_en.md
+++ b/modules/image/object_detection/ssd_vgg16_512_coco2017/README_en.md
@@ -22,7 +22,8 @@
- ### Module Introduction
- - Single Shot MultiBox Detector (SSD) 是一种单阶段的目标检测器.与两阶段的检测方法不同,单阶段目标检测并不进行区域推荐,而是直接从特征图回归出目标的边界框和分类概率.SSD 运用了这种单阶段检测的思想,并且对其进行改进:在不同尺度的特征图上检测对应尺度的目标.该PaddleHub Module的基网络为VGG16模型,在Pascal数据集上预训练得到,目前仅支持预测.
+ - Single Shot MultiBox Detector (SSD) is a one-stage detector. Different from two-stage detector, SSD frames object detection as a re- gression problem to spatially separated bounding boxes and associated class probabilities. This module is based on VGG16, trained on COCO2017 dataset, and can be used for object detection.
+
## II.Installation
@@ -72,7 +73,7 @@
visualization=True)
```
- - 预测API,检测输入图片中的所有目标的位置.
+ - Detection API, detect positions of all objects in image
- **Parameters**
@@ -81,22 +82,22 @@
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- output_dir (str): save path of images;
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
**NOTE:** choose one parameter to provide data from paths and images
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
def save_inference_model(dirname,
diff --git a/modules/image/object_detection/yolov3_darknet53_coco2017/README_en.md b/modules/image/object_detection/yolov3_darknet53_coco2017/README_en.md
index b4a50e27645bf5973ed5fe21c72aaa9735473543..db292af2bb1d81b128c6be92884c03a950dc2bfb 100644
--- a/modules/image/object_detection/yolov3_darknet53_coco2017/README_en.md
+++ b/modules/image/object_detection/yolov3_darknet53_coco2017/README_en.md
@@ -22,7 +22,7 @@
- ### Module Introduction
- - YOLOv3是由Joseph Redmon和Ali Farhadi提出的单阶段检测器, 该检测器与达到同样精度的传统目标检测方法相比,推断速度能达到接近两倍. YOLOv3将输入图像划分格子,并对每个格子预测bounding box.YOLOv3的loss函数由三部分组成:Location误差,Confidence误差和分类误差.该PaddleHub Module预训练数据集为COCO2017,目前仅支持预测.
+ - YOLOv3 is a one-stage detector proposed by Joseph Redmon and Ali Farhadi, which can reach comparable accuracy but twice as fast as traditional methods. This module is based on YOLOv3, trained on COCO2017, and can be used for object detection.
## II.Installation
@@ -72,7 +72,7 @@
visualization=True)
```
- - 预测API,检测输入图片中的所有目标的位置.
+ - Detection API, detect positions of all objects in image
- **Parameters**
@@ -81,22 +81,22 @@
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- output_dir (str): save path of images;
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
**NOTE:** choose one parameter to provide data from paths and images
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
def save_inference_model(dirname,
diff --git a/modules/image/object_detection/yolov3_darknet53_pedestrian/README_en.md b/modules/image/object_detection/yolov3_darknet53_pedestrian/README_en.md
index fd36da1aa7d139cade7ea899de613cf4709ae5eb..0642406af7ca5caa8b3ede914a4c97a656743da2 100644
--- a/modules/image/object_detection/yolov3_darknet53_pedestrian/README_en.md
+++ b/modules/image/object_detection/yolov3_darknet53_pedestrian/README_en.md
@@ -4,7 +4,7 @@
| :--- | :---: |
|Category|object detection|
|Network|YOLOv3|
-|Dataset|百度自建大规模行人Dataset|
+|Dataset|Baidu Pedestrian Dataset|
|Fine-tuning supported or not|No|
|Module Size|238MB|
|Latest update date|2021-03-15|
@@ -22,7 +22,7 @@
- ### Module Introduction
- - 行人检测是计算机视觉技术中的目标检测问题,用于判断图像中是否存在行人并给予精确定位,定位结果用矩形框表示.行人检测技术有很强的使用价值,它可以与行人跟踪、行人重识别等技术结合,应用于汽车无人驾驶系统、智能视频监控、人体行为分析、客流统计系统、智能交通等领域.yolov3_darknet53_pedestrian Module的网络为YOLOv3, 其中backbone为DarkNet53, 采用百度自建大规模车辆数据集训练得到,目前仅支持预测.
+ - YOLOv3 is a one-stage detector proposed by Joseph Redmon and Ali Farhadi, which can reach comparable accuracy but twice as fast as traditional methods. This module is based on YOLOv3, trained on Baidu Pedestrian Dataset, and can be used for pedestrian detection.
## II.Installation
@@ -72,7 +72,7 @@
visualization=True)
```
- - 预测API,检测输入图片中的所有行人的位置.
+ - Detection API, detect positions of all pedestrian in image
- **Parameters**
@@ -81,7 +81,7 @@
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- output_dir (str): save path of images;
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
**NOTE:** choose one parameter to provide data from paths and images
@@ -89,15 +89,15 @@
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
def save_inference_model(dirname,
diff --git a/modules/image/object_detection/yolov3_darknet53_vehicles/README_en.md b/modules/image/object_detection/yolov3_darknet53_vehicles/README_en.md
index ce6922670ff1d9b929907f265977bf4a9f2b0828..e27ec19078aae8b9f90e7df5d02c876aba61d5b3 100644
--- a/modules/image/object_detection/yolov3_darknet53_vehicles/README_en.md
+++ b/modules/image/object_detection/yolov3_darknet53_vehicles/README_en.md
@@ -4,7 +4,7 @@
| :--- | :---: |
|Category|object detection|
|Network|YOLOv3|
-|Dataset|百度自建大规模车辆Dataset|
+|Dataset|Baidu Vehicle Dataset|
|Fine-tuning supported or not|No|
|Module Size|238MB|
|Latest update date|2021-03-15|
@@ -22,7 +22,8 @@
- ### Module Introduction
- - 车辆检测是城市交通监控中非常重要并且具有挑战性的任务,该任务的难度在于对复杂场景中相对较小的车辆进行精准地定位和分类.该 PaddleHub Module 的网络为 YOLOv3, 其中 backbone 为 DarkNet53,采用百度自建大规模车辆数据集训练得到,支持car (汽车)、truck (卡车)、bus (公交车)、motorbike (摩托车)、tricycle (三轮车)等车型的识别.目前仅支持预测.
+ - YOLOv3 is a one-stage detector proposed by Joseph Redmon and Ali Farhadi, which can reach comparable accuracy but twice as fast as traditional methods. This module is based on YOLOv3, trained on Baidu Vehicle Dataset, and can be used for vehicle detection.
+
## II.Installation
@@ -72,7 +73,7 @@
visualization=True)
```
- - 预测API,检测输入图片中的所有车辆的位置.
+ - Detection API, detect positions of all vehicles in image
- **Parameters**
@@ -81,22 +82,22 @@
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- output_dir (str): save path of images;
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
**NOTE:** choose one parameter to provide data from paths and images
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
def save_inference_model(dirname,
diff --git a/modules/image/object_detection/yolov3_darknet53_venus/README_en.md b/modules/image/object_detection/yolov3_darknet53_venus/README_en.md
index aaa9a1c6c3040a73a0c892328e9f9e402e5b8bab..09d6fcde6ff57cd213231b026738533e85dc4984 100644
--- a/modules/image/object_detection/yolov3_darknet53_venus/README_en.md
+++ b/modules/image/object_detection/yolov3_darknet53_venus/README_en.md
@@ -4,7 +4,7 @@
| :--- | :---: |
|Category|object detection|
|Network|YOLOv3|
-|Dataset|百度自建Dataset|
+|Dataset|Baidu Detection Dataset|
|Fine-tuning supported or not|Yes|
|Module Size|501MB|
|Latest update date|2021-02-26|
@@ -15,7 +15,7 @@
- ### Module Introduction
- - YOLOv3是由Joseph Redmon和Ali Farhadi提出的单阶段检测器, 该检测器与达到同样精度的传统目标检测方法相比,推断速度能达到接近两倍. YOLOv3将输入图像划分格子,并对每个格子预测bounding box.YOLOv3的loss函数由三部分组成:Location误差,Confidence误差和分类误差.该PaddleHub Module是由800+tag,170w图片,1000w+检测框训练的大规模通用检测模型,在8个数据集上MAP平均提升5.36%,iou=0.5的准确率提升4.53%.对比于其他通用检测模型,使用该Module进行finetune,可以更快收敛,达到较优效果.
+ - YOLOv3 is a one-stage detector proposed by Joseph Redmon and Ali Farhadi, which can reach comparable accuracy but twice as fast as traditional methods. This module is based on YOLOv3, trained on Baidu Vehicle Dataset which consists of 170w pictures and 1000w+ boxes, improve the accuracy on 8 test datasets for average 5.36%, and can be used for vehicle detection.
## II.Installation
@@ -43,20 +43,20 @@
get_prediction=False)
```
- - 提取特征,用于迁移学习.
+ - Extract features, and do transfer learning
- **Parameters**
- - trainable(bool): Parameters是否可训练;
- - pretrained (bool): 是否加载预训练模型;
- - get\_prediction (bool): 是否执行预测.
+ - trainable(bool): whether parameters trainable or not
+ - pretrained (bool): whether load pretrained model or not
+ - get\_prediction (bool): whether perform prediction
- **Return**
- - inputs (dict): 模型的输入,keys 包括 'image', 'im\_size',相应的取值为:
- - image (Variable): 图像变量
- - im\_size (Variable): 图片的尺寸
- - outputs (dict): 模型的输出.如果 get\_prediction 为 False,输出 'head\_features'、'body\_features',否则输出 'bbox\_out'
- - context\_prog (Program): 用于迁移学习的 Program
+ - inputs (dict): inputs, a dict, include two keys: "image" and "im\_size"
+ - image (Variable): image variable
+ - im\_size (Variable): image size
+ - outputs (dict): model output
+ - program for transfer learning
- ```python
def object_detection(paths=None,
@@ -68,7 +68,7 @@
output_dir='detection_result')
```
- - 预测API,检测输入图片中的所有目标的位置.
+ - Detection API, detect positions of all objects in image
- **Parameters**
@@ -76,7 +76,7 @@
- images (list\[numpy.ndarray\]): image data, ndarray.shape is in the format [H, W, C], BGR;
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
- output_dir (str): save path of images;
@@ -85,12 +85,12 @@
- res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
def save_inference_model(dirname,
diff --git a/modules/image/object_detection/yolov3_mobilenet_v1_coco2017/README_en.md b/modules/image/object_detection/yolov3_mobilenet_v1_coco2017/README_en.md
index a1be6c9787ed45afa9bff7fdb0b6724e2b9b79ae..e04db2657a703319f782d9e189cc93555917b7ca 100644
--- a/modules/image/object_detection/yolov3_mobilenet_v1_coco2017/README_en.md
+++ b/modules/image/object_detection/yolov3_mobilenet_v1_coco2017/README_en.md
@@ -23,7 +23,7 @@
- ### Module Introduction
- - YOLOv3是由Joseph Redmon和Ali Farhadi提出的单阶段检测器, 该检测器与达到同样精度的传统目标检测方法相比,推断速度能达到接近两倍.YOLOv3将输入图像划分格子,并对每个格子预测bounding box.YOLOv3的loss函数由三部分组成:Location误差,Confidence误差和分类误差.该PaddleHub Module预训练数据集为COCO2017,目前仅支持预测.
+ - YOLOv3 is a one-stage detector proposed by Joseph Redmon and Ali Farhadi, which can reach comparable accuracy but twice as fast as traditional methods. This module is based on YOLOv3, trained on COCO2017, and can be used for object detection.
## II.Installation
@@ -73,7 +73,7 @@
visualization=True)
```
- - 预测API,检测输入图片中的所有目标的位置.
+ - Detection API, detect positions of all objects in image
- **Parameters**
@@ -82,22 +82,22 @@
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- output_dir (str): save path of images;
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
**NOTE:** choose one parameter to provide data from paths and images
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
def save_inference_model(dirname,
diff --git a/modules/image/object_detection/yolov3_resnet34_coco2017/README_en.md b/modules/image/object_detection/yolov3_resnet34_coco2017/README_en.md
index 73dcd0c3ef68bdb92e3e2c0af3ac3898c6d6b9ad..28a147c2c2048ddf03f3b03ecc1a6dd6c990f738 100644
--- a/modules/image/object_detection/yolov3_resnet34_coco2017/README_en.md
+++ b/modules/image/object_detection/yolov3_resnet34_coco2017/README_en.md
@@ -22,7 +22,7 @@
- ### Module Introduction
- - YOLOv3是由Joseph Redmon和Ali Farhadi提出的单阶段检测器, 该检测器与达到同样精度的传统目标检测方法相比,推断速度能达到接近两倍. YOLOv3将输入图像划分格子,并对每个格子预测bounding box.YOLOv3的loss函数由三部分组成:Location误差,Confidence误差和分类误差.该PaddleHub Module预训练数据集为COCO2017,目前仅支持预测.
+ - YOLOv3 is a one-stage detector proposed by Joseph Redmon and Ali Farhadi, which can reach comparable accuracy but twice as fast as traditional methods. This module is based on YOLOv3, trained on COCO2017, and can be used for object detection.
## II.Installation
@@ -72,7 +72,7 @@
visualization=True)
```
- - 预测API,检测输入图片中的所有目标的位置.
+ - Detection API, detect positions of all objects in image
- **Parameters**
@@ -81,22 +81,22 @@
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- output_dir (str): save path of images;
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
**NOTE:** choose one parameter to provide data from paths and images
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
def save_inference_model(dirname,
diff --git a/modules/image/object_detection/yolov3_resnet50_vd_coco2017/README_en.md b/modules/image/object_detection/yolov3_resnet50_vd_coco2017/README_en.md
index 4400b0eb8de89cf58276c9ebf0adb73f0ac1edd8..f84c68718b7377317331ab265a525c1baf5fbdd0 100644
--- a/modules/image/object_detection/yolov3_resnet50_vd_coco2017/README_en.md
+++ b/modules/image/object_detection/yolov3_resnet50_vd_coco2017/README_en.md
@@ -22,7 +22,7 @@
- ### Module Introduction
- - YOLOv3是由Joseph Redmon和Ali Farhadi提出的单阶段检测器, 该检测器与达到同样精度的传统目标检测方法相比,推断速度能达到接近两倍. YOLOv3将输入图像划分格子,并对每个格子预测bounding box.YOLOv3的loss函数由三部分组成:Location误差,Confidence误差和分类误差.该PaddleHub Module预训练数据集为COCO2017,目前仅支持预测.
+ - YOLOv3 is a one-stage detector proposed by Joseph Redmon and Ali Farhadi, which can reach comparable accuracy but twice as fast as traditional methods. This module is based on YOLOv3, trained on COCO2017, and can be used for object detection.
## II.Installation
@@ -72,7 +72,7 @@
visualization=True)
```
- - 预测API,检测输入图片中的所有目标的位置.
+ - Detection API, detect positions of all objects in image
- **Parameters**
@@ -81,22 +81,22 @@
- batch_size (int): the size of batch;
- use_gpu (bool): use GPU or not; **set the CUDA_VISIBLE_DEVICES environment variable first if you are using GPU**
- output_dir (str): save path of images;
- - score\_thresh (float): 识别置信度的阈值;
+ - score\_thresh (float): confidence threshold;
- visualization (bool): Whether to save the results as picture files;
**NOTE:** choose one parameter to provide data from paths and images
- **Return**
- - res (list\[dict\]): classication results, each element in the list is dict, key is the label name, and value is the corresponding probability
+ - res (list\[dict\]): results
- data (list): detection results, each element in the list is dict
- confidence (float): the confidence of the result
- - label (str): 标签
+ - label (str): label
- left (int): the upper left corner x coordinate of the detection box
- top (int): the upper left corner y coordinate of the detection box
- right (int): the lower right corner x coordinate of the detection box
- bottom (int): the lower right corner y coordinate of the detection box
- - save\_path (str, optional): 识别结果的保存路径 (仅当visualization=True时存在)
+ - save\_path (str, optional): output path for saving results
- ```python
def save_inference_model(dirname,