README.md 13.0 KB
Newer Older
N
Nikita Manovich 已提交
1 2
# Computer Vision Annotation Tool (CVAT)

3 4 5 6 7 8
[![CI][ci-img]][ci-url]
[![Gitter chat][gitter-img]][gitter-url]
[![Coverage Status][coverage-img]][coverage-url]
[![server pulls][docker-server-pulls-img]][docker-server-image-url]
[![ui pulls][docker-ui-pulls-img]][docker-ui-image-url]
[![DOI][doi-img]][doi-url]
9

10 11 12
CVAT is free, online, interactive video and image annotation
tool for computer vision. It is being used by our team to
annotate million of objects with different properties. Many UI
13 14
and UX decisions are based on feedbacks from professional data
annotation team. Try it online [cvat.org](https://cvat.org).
N
Nikita Manovich 已提交
15

16
![CVAT screenshot](site/content/en/images/cvat.jpg)
N
Nikita Manovich 已提交
17 18 19

## Documentation

20 21 22 23 24 25 26 27 28
- [Contributing](https://cvat-ai.github.io/cvat/docs/contributing/)
- [Installation guide](https://cvat-ai.github.io/cvat/docs/administration/basics/installation/)
- [Manual](https://cvat-ai.github.io/cvat/docs/manual/)
- [Django REST API documentation](https://cvat-ai.github.io/cvat/docs/administration/basics/rest_api_guide/)
- [Datumaro dataset framework](https://github.com/cvat-ai/datumaro/blob/develop/README.md)
- [Command line interface](https://cvat-ai.github.io/cvat/docs/manual/advanced/cli/)
- [XML annotation format](https://cvat-ai.github.io/cvat/docs/manual/advanced/xml_format/)
- [AWS Deployment Guide](https://cvat-ai.github.io/cvat/docs/administration/basics/aws-deployment-guide/)
- [Frequently asked questions](https://cvat-ai.github.io/cvat/docs/faq/)
N
Nikita Manovich 已提交
29
- [Questions](#questions)
N
Nikita Manovich 已提交
30 31 32

## Screencasts

T
TOsmanov 已提交
33 34 35 36
- [Introduction](https://youtu.be/JERohTFp-NI)
- [Annotation mode](https://youtu.be/vH_639N67HI)
- [Interpolation of bounding boxes](https://youtu.be/Hc3oudNuDsY)
- [Interpolation of polygons](https://youtu.be/K4nis9lk92s)
B
Boris Sekachev 已提交
37
- [Tag annotation video](https://youtu.be/62bI4mF-Xfk)
T
TOsmanov 已提交
38
- [Attribute mode](https://youtu.be/iIkJsOkDzVA)
T
TOsmanov 已提交
39 40 41
- [Segmentation mode](https://youtu.be/9Fe_GzMLo3E)
- [Tutorial for polygons](https://youtu.be/C7-r9lZbjBw)
- [Semi-automatic segmentation](https://youtu.be/9HszWP_qsRQ)
42

43
## Supported annotation formats
A
Andrey Zhavoronkov 已提交
44

A
Anastasia Yasakova 已提交
45
Format selection is possible after clicking on the Upload annotation and Dump
46
annotation buttons. [Datumaro](https://github.com/cvat-ai/datumaro)
A
Anastasia Yasakova 已提交
47 48 49 50
dataset framework allows additional dataset transformations via its command
line tool and Python library.

For more information about supported formats look at the
51
[documentation](https://cvat-ai.github.io/cvat/docs/manual/advanced/formats/).
52 53 54

<!--lint disable maximum-line-length-->

55 56
| Annotation format                                                                                         | Import | Export |
| --------------------------------------------------------------------------------------------------------- | ------ | ------ |
57 58 59
| [CVAT for images](https://cvat-ai.github.io/cvat/docs/manual/advanced/xml_format/#annotation)     | X      | X      |
| [CVAT for a video](https://cvat-ai.github.io/cvat/docs/manual/advanced/xml_format/#interpolation) | X      | X      |
| [Datumaro](https://github.com/cvat-ai/datumaro)                                                   |        | X      |
60 61 62 63 64 65 66 67 68 69 70 71 72
| [PASCAL VOC](http://host.robots.ox.ac.uk/pascal/VOC/)                                                     | X      | X      |
| Segmentation masks from [PASCAL VOC](http://host.robots.ox.ac.uk/pascal/VOC/)                             | X      | X      |
| [YOLO](https://pjreddie.com/darknet/yolo/)                                                                | X      | X      |
| [MS COCO Object Detection](http://cocodataset.org/#format-data)                                           | X      | X      |
| [TFrecord](https://www.tensorflow.org/tutorials/load_data/tfrecord)                                       | X      | X      |
| [MOT](https://motchallenge.net/)                                                                          | X      | X      |
| [LabelMe 3.0](http://labelme.csail.mit.edu/Release3.0)                                                    | X      | X      |
| [ImageNet](http://www.image-net.org)                                                                      | X      | X      |
| [CamVid](http://mi.eng.cam.ac.uk/research/projects/VideoRec/CamVid/)                                      | X      | X      |
| [WIDER Face](http://shuoyang1213.me/WIDERFACE/)                                                           | X      | X      |
| [VGGFace2](https://github.com/ox-vgg/vgg_face2)                                                           | X      | X      |
| [Market-1501](https://www.aitribune.com/dataset/2018051063)                                               | X      | X      |
| [ICDAR13/15](https://rrc.cvc.uab.es/?ch=2)                                                                | X      | X      |
73 74 75 76
| [Open Images V6](https://storage.googleapis.com/openimages/web/index.html)                                | X      | X      |
| [Cityscapes](https://www.cityscapes-dataset.com/login/)                                                   | X      | X      |
| [KITTI](http://www.cvlibs.net/datasets/kitti/)                                                            | X      | X      |
| [LFW](http://vis-www.cs.umass.edu/lfw/)                                                                   | X      | X      |
77 78

<!--lint enable maximum-line-length-->
A
Andrey Zhavoronkov 已提交
79

80 81 82
## Deep learning serverless functions for automatic labeling

<!--lint disable maximum-line-length-->
83

A
Ali Jahani 已提交
84 85
| Name                                                                                                    | Type       | Framework  | CPU | GPU |
| ------------------------------------------------------------------------------------------------------- | ---------- | ---------- | --- | --- |
86 87 88 89 90 91 92
| [Deep Extreme Cut](/serverless/openvino/dextr/nuclio)                                                   | interactor | OpenVINO   | X   |     |
| [Faster RCNN](/serverless/openvino/omz/public/faster_rcnn_inception_v2_coco/nuclio)                     | detector   | OpenVINO   | X   |     |
| [Mask RCNN](/serverless/openvino/omz/public/mask_rcnn_inception_resnet_v2_atrous_coco/nuclio)           | detector   | OpenVINO   | X   |     |
| [YOLO v3](/serverless/openvino/omz/public/yolo-v3-tf/nuclio)                                            | detector   | OpenVINO   | X   |     |
| [Object reidentification](/serverless/openvino/omz/intel/person-reidentification-retail-300/nuclio)     | reid       | OpenVINO   | X   |     |
| [Semantic segmentation for ADAS](/serverless/openvino/omz/intel/semantic-segmentation-adas-0001/nuclio) | detector   | OpenVINO   | X   |     |
| [Text detection v4](/serverless/openvino/omz/intel/text-detection-0004/nuclio)                          | detector   | OpenVINO   | X   |     |
93
| [YOLO v5](/serverless/pytorch/ultralytics/yolov5/nuclio)                                                | detector   | PyTorch    | X   |     |
T
Timur Osmanov 已提交
94
| [SiamMask](/serverless/pytorch/foolwood/siammask/nuclio)                                                | tracker    | PyTorch    | X   | X   |
95
| [f-BRS](/serverless/pytorch/saic-vul/fbrs/nuclio)                                                       | interactor | PyTorch    | X   |     |
96
| [HRNet](/serverless/pytorch/saic-vul/hrnet/nuclio)                                                      | interactor | PyTorch    |     | X   |
97
| [Inside-Outside Guidance](/serverless/pytorch/shiyinzhang/iog/nuclio)                                   | interactor | PyTorch    | X   |     |
A
Ali Jahani 已提交
98
| [Faster RCNN](/serverless/tensorflow/faster_rcnn_inception_v2_coco/nuclio)                              | detector   | TensorFlow | X   | X   |
99
| [Mask RCNN](/serverless/tensorflow/matterport/mask_rcnn/nuclio)                                         | detector   | TensorFlow | X   | X   |
100
| [RetinaNet](serverless/pytorch/facebookresearch/detectron2/retinanet/nuclio)                            | detector   | PyTorch    | X   | X   |
101
| [Face Detection](/serverless/openvino/omz/intel/face-detection-0205/nuclio)                             | detector   | OpenVINO   | X   |     |
102 103

<!--lint enable maximum-line-length-->
N
Nikita Manovich 已提交
104

105
## Online demo: [cvat.org](https://cvat.org)
106

107 108 109
This is an online demo with the latest version of the annotation tool.
Try it online without local installation. Only own or assigned tasks
are visible to users.
110

111
Disabled features:
V
Vitaliy Nishukov 已提交
112

113
- [Analytics: management and monitoring of data annotation team](https://cvat-ai.github.io/cvat/docs/administration/advanced/analytics/)
114 115

Limitations:
V
Vitaliy Nishukov 已提交
116

117 118
- No more than 10 tasks per user
- Uploaded data is limited to 500Mb
119

120 121 122 123
## Prebuilt Docker images

Prebuilt docker images for CVAT releases are available on Docker Hub:

K
Kirill Sizov 已提交
124 125
- [cvat_server](https://hub.docker.com/r/cvat/server)
- [cvat_ui](https://hub.docker.com/r/cvat/ui)
126

127 128 129 130
## REST API
The current REST API version is `2.0-alpha`. We focus on its improvement and therefore
REST API may be changed in the next release.

N
Nikita Manovich 已提交
131 132 133 134
## LICENSE

Code released under the [MIT License](https://opensource.org/licenses/MIT).

135 136 137 138 139 140
This software uses LGPL licensed libraries from the [FFmpeg](https://www.ffmpeg.org) project.
The exact steps on how FFmpeg was configured and compiled can be found in the [Dockerfile](Dockerfile).

FFmpeg is an open source framework licensed under LGPL and GPL.
See [https://www.ffmpeg.org/legal.html](https://www.ffmpeg.org/legal.html). You are solely responsible
for determining if your use of FFmpeg requires any
141
additional licenses. CVAT.ai is not responsible for obtaining any
142 143 144
such licenses, nor liable for any licensing fees due in
connection with your use of FFmpeg.

145 146
## Partners

A
Andrey Zhavoronkov 已提交
147 148
- [ATLANTIS](https://github.com/smhassanerfani/atlantis) is an open-source dataset for semantic segmentation
  of waterbody images, depevoped by [iWERS](http://ce.sc.edu/iwers/) group in the
149 150 151
  Department of Civil and Environmental Engineering at University of South Carolina, using CVAT.
  For developing a semantic segmentation dataset using CVAT, please check
  [ATLANTIS published article](https://www.sciencedirect.com/science/article/pii/S1364815222000391),
A
Andrey Zhavoronkov 已提交
152
  [ATLANTIS Development Kit](https://github.com/smhassanerfani/atlantis/tree/master/adk)
153
  and [annotation tutorial videos](https://www.youtube.com/playlist?list=PLIfLGY-zZChS5trt7Lc3MfNhab7OWl2BR).
154 155 156 157 158 159 160 161
- [Onepanel](https://github.com/onepanelio/core) is an open source
  vision AI platform that fully integrates CVAT with scalable data processing
  and parallelized training pipelines.
- [DataIsKey](https://dataiskey.eu/annotation-tool/) uses CVAT as their prime data labeling tool
  to offer annotation services for projects of any size.
- [Human Protocol](https://hmt.ai) uses CVAT as a way of adding annotation service to the human protocol.
- [Cogito Tech LLC](https://bit.ly/3klT0h6), a Human-in-the-Loop Workforce Solutions Provider, used CVAT
  in annotation of about 5,000 images for a brand operating in the fashion segment.
162 163 164 165
- [FiftyOne](https://fiftyone.ai) is an open-source dataset curation and model analysis
tool for visualizing, exploring, and improving computer vision datasets and models that is
[tightly integrated](https://voxel51.com/docs/fiftyone/integrations/cvat.html) with CVAT
for annotation and label refinement.
166

I
idriss 已提交
167 168
## Questions

N
Nikita Manovich 已提交
169 170 171
CVAT usage related questions or unclear concepts can be posted in our
[Gitter chat](https://gitter.im/opencv-cvat) for **quick replies** from
contributors and other users.
I
idriss 已提交
172

N
Nikita Manovich 已提交
173 174
However, if you have a feature request or a bug report that can reproduced,
feel free to open an issue (with steps to reproduce the bug if it's a bug
V
Vitaliy Nishukov 已提交
175
report) on [GitHub\* issues](https://github.com/opencv/cvat/issues).
I
idriss 已提交
176

N
Nikita Manovich 已提交
177
If you are not sure or just want to browse other users common questions,
178 179 180
[Gitter chat](https://gitter.im/opencv-cvat) is the way to go.

Other ways to ask questions and get our support:
V
Vitaliy Nishukov 已提交
181 182

- [\#cvat](https://stackoverflow.com/search?q=%23cvat) tag on StackOverflow\*
183 184

## Links
V
Vitaliy Nishukov 已提交
185

186 187 188
- [Intel AI blog: New Computer Vision Tool Accelerates Annotation of Digital Images and Video](https://www.intel.ai/introducing-cvat)
- [Intel Software: Computer Vision Annotation Tool: A Universal Approach to Data Annotation](https://software.intel.com/en-us/articles/computer-vision-annotation-tool-a-universal-approach-to-data-annotation)
- [VentureBeat: Intel open-sources CVAT, a toolkit for data labeling](https://venturebeat.com/2019/03/05/intel-open-sources-cvat-a-toolkit-for-data-labeling/)
189

190 191
  <!-- prettier-ignore-start -->
  <!-- Badges -->
192 193 194

[docker-server-pulls-img]: https://img.shields.io/docker/pulls/openvino/cvat_server.svg?style=flat-square&label=server%20pulls
[docker-server-image-url]: https://hub.docker.com/r/openvino/cvat_server
A
Andrey Zhavoronkov 已提交
195

196 197
[docker-ui-pulls-img]: https://img.shields.io/docker/pulls/openvino/cvat_ui.svg?style=flat-square&label=UI%20pulls
[docker-ui-image-url]: https://hub.docker.com/r/openvino/cvat_ui
A
Andrey Zhavoronkov 已提交
198 199 200 201

[ci-img]: https://github.com/cvat-ai/cvat/workflows/CI/badge.svg?branch=develop
[ci-url]: https://github.com/cvat-ai/cvat/actions

202 203
[gitter-img]: https://badges.gitter.im/opencv-cvat/gitter.png
[gitter-url]: https://gitter.im/opencv-cvat
A
Andrey Zhavoronkov 已提交
204 205 206 207

[coverage-img]: https://coveralls.io/repos/github/cvat-ai/cvat/badge.svg?branch=develop
[coverage-url]: https://coveralls.io/github/cvat-ai/cvat?branch=develop

208 209
[doi-img]: https://zenodo.org/badge/139156354.svg
[doi-url]: https://zenodo.org/badge/latestdoi/139156354