README.md 15.1 KB
Newer Older
1
![CVAT logo](site/content/en/images/cvat_poster_with_name.png)
N
Nikita Manovich 已提交
2 3
# Computer Vision Annotation Tool (CVAT)

4 5
<a href="https://www.producthunt.com/posts/cvat-computer-vision-annotation-tool?utm_source=badge-featured&utm_medium=badge&utm_souce=badge-cvat&#0045;computer&#0045;vision&#0045;annotation&#0045;tool" target="_blank"><img src="https://api.producthunt.com/widgets/embed-image/v1/featured.svg?post_id=353415&theme=light" alt="CVAT&#0032;–&#0032;Computer&#0032;Vision&#0032;Annotation&#0032;Tool - The&#0032;open&#0032;data&#0032;annotation&#0032;platform&#0032;for&#0032;AI | Product Hunt" style="width: 250px; height: 54px;" width="250" height="54" /></a>

6 7
[![CI][ci-img]][ci-url]
[![Gitter chat][gitter-img]][gitter-url]
8
[![Discord][discord-img]][discord-url]
9 10 11 12
[![Coverage Status][coverage-img]][coverage-url]
[![server pulls][docker-server-pulls-img]][docker-server-image-url]
[![ui pulls][docker-ui-pulls-img]][docker-ui-image-url]
[![DOI][doi-img]][doi-url]
13

14
CVAT is an interactive video and image annotation
15
tool for computer vision. It is used by tens of thousands of users and
16
companies around the world. CVAT is free and open-source.
N
Nikita Manovich 已提交
17

18 19 20
**A new repo**: CVAT core team moved the active development of the tool
to this new repository. Our mission is to help developers, companies and
organizations around the world to solve real problems using the Data-centric
21
AI approach.
N
Nikita Manovich 已提交
22

23
Start using CVAT online for free: [cvat.ai](https://cvat.ai). Or set it up as a self-hosted solution:
W
William Schmitt 已提交
24
[read here](https://opencv.github.io/cvat/docs/administration/basics/installation/).
25 26 27 28

![CVAT screencast](site/content/en/images/cvat-ai-screencast.gif)

## Quick start ⚡
N
Nikita Manovich 已提交
29

W
William Schmitt 已提交
30 31 32 33
- [Installation guide](https://opencv.github.io/cvat/docs/administration/basics/installation/)
- [Manual](https://opencv.github.io/cvat/docs/manual/)
- [Contributing](https://opencv.github.io/cvat/docs/contributing/)
- [Django REST API documentation](https://opencv.github.io/cvat/docs/administration/basics/rest_api_guide/)
34
- [Datumaro dataset framework](https://github.com/cvat-ai/datumaro/blob/develop/README.md)
W
William Schmitt 已提交
35 36 37 38
- [Command line interface](https://opencv.github.io/cvat/docs/manual/advanced/cli/)
- [XML annotation format](https://opencv.github.io/cvat/docs/manual/advanced/xml_format/)
- [AWS Deployment Guide](https://opencv.github.io/cvat/docs/administration/basics/aws-deployment-guide/)
- [Frequently asked questions](https://opencv.github.io/cvat/docs/faq/)
39
- [Where to ask questions](#where-to-ask-questions)
N
Nikita Manovich 已提交
40

41 42 43 44 45
## Partners ❤️

CVAT is used by teams all over the world. If you use us, please drop us a line at
[contact@cvat.ai](mailto:contact+github@cvat.ai) - and we'll add you to this list.

46
- [ATLANTIS](https://github.com/smhassanerfani/atlantis), an open-source dataset for semantic segmentation
47
  of waterbody images, depeloped by [iWERS](http://ce.sc.edu/iwers/) group in the
48 49 50 51 52
  Department of Civil and Environmental Engineering at University of South Carolina, is using CVAT.
  For developing a semantic segmentation dataset using CVAT, please check
  [ATLANTIS published article](https://www.sciencedirect.com/science/article/pii/S1364815222000391),
  [ATLANTIS Development Kit](https://github.com/smhassanerfani/atlantis/tree/master/adk)
  and [annotation tutorial videos](https://www.youtube.com/playlist?list=PLIfLGY-zZChS5trt7Lc3MfNhab7OWl2BR).
53
- [Onepanel](https://github.com/onepanelio/core) is an open-source
54 55 56 57
  vision AI platform that fully integrates CVAT with scalable data processing
  and parallelized training pipelines.
- [DataIsKey](https://dataiskey.eu/annotation-tool/) uses CVAT as their prime data labeling tool
  to offer annotation services for projects of any size.
58
- [Human Protocol](https://hmt.ai) uses CVAT as a way of adding annotation service to the Human Protocol.
59 60 61 62 63 64 65 66 67 68 69
- [Cogito Tech LLC](https://bit.ly/3klT0h6), a Human-in-the-Loop Workforce Solutions Provider, used CVAT
  in annotation of about 5,000 images for a brand operating in the fashion segment.
- [FiftyOne](https://fiftyone.ai) is an open-source dataset curation and model analysis
tool for visualizing, exploring, and improving computer vision datasets and models that is
[tightly integrated](https://voxel51.com/docs/fiftyone/integrations/cvat.html) with CVAT
for annotation and label refinement.

## CVAT online: [cvat.ai](https://cvat.ai)

This is an online version of CVAT. It's free, efficient, and easy to use.

70 71
[cvat.ai](https://cvat.ai) runs the latest version of the tool. You can create up
to 10 tasks there and upload up to 500Mb of data to annotate. It will only be
72 73
visible to you or people you assign to it.

W
William Schmitt 已提交
74
For now, it does not have [analytics features](https://opencv.github.io/cvat/docs/administration/advanced/analytics/)
75 76 77 78 79 80 81 82 83 84 85 86 87 88 89
like management and monitoring the data annotation team.

We plan to enhance [cvat.ai](https://cvat.ai) with new powerful features. Stay tuned!

## Prebuilt Docker images 🐳

Prebuilt docker images are the easiest way to start using CVAT locally. They are available on Docker Hub:

- [cvat/server](https://hub.docker.com/r/cvat/server)
- [cvat/ui](https://hub.docker.com/r/cvat/ui)

The images have been downloaded more than 1M times so far.

## REST API

W
William Schmitt 已提交
90
CVAT has a REST API: [documentation](https://opencv.github.io/cvat/docs/administration/basics/rest_api_guide/).
91
Its current version is `2.0-alpha`. We focus on its improvement, and the API may be changed in the next releases.
92 93 94 95

## Screencasts 🎦

Here are some screencasts showing how to use CVAT.
N
Nikita Manovich 已提交
96

T
TOsmanov 已提交
97 98 99 100
- [Introduction](https://youtu.be/JERohTFp-NI)
- [Annotation mode](https://youtu.be/vH_639N67HI)
- [Interpolation of bounding boxes](https://youtu.be/Hc3oudNuDsY)
- [Interpolation of polygons](https://youtu.be/K4nis9lk92s)
B
Boris Sekachev 已提交
101
- [Tag annotation video](https://youtu.be/62bI4mF-Xfk)
T
TOsmanov 已提交
102
- [Attribute mode](https://youtu.be/iIkJsOkDzVA)
T
TOsmanov 已提交
103 104 105
- [Segmentation mode](https://youtu.be/9Fe_GzMLo3E)
- [Tutorial for polygons](https://youtu.be/C7-r9lZbjBw)
- [Semi-automatic segmentation](https://youtu.be/9HszWP_qsRQ)
106

107
## Supported annotation formats
A
Andrey Zhavoronkov 已提交
108

109 110
CVAT supports multiple annotation formats. You can select the format after clicking the "Upload annotation" and "Dump
annotation" buttons. [Datumaro](https://github.com/cvat-ai/datumaro)
A
Anastasia Yasakova 已提交
111 112 113
dataset framework allows additional dataset transformations via its command
line tool and Python library.

114
For more information about the supported formats, look at the
W
William Schmitt 已提交
115
[documentation](https://opencv.github.io/cvat/docs/manual/advanced/formats/).
116 117 118

<!--lint disable maximum-line-length-->

119 120
| Annotation format                                                                                         | Import | Export |
| --------------------------------------------------------------------------------------------------------- | ------ | ------ |
W
William Schmitt 已提交
121 122
| [CVAT for images](https://opencv.github.io/cvat/docs/manual/advanced/xml_format/#annotation)     |   ✔️    |   ✔️    |
| [CVAT for a video](https://opencv.github.io/cvat/docs/manual/advanced/xml_format/#interpolation) |   ✔️    |   ✔️    |
123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140
| [Datumaro](https://github.com/cvat-ai/datumaro)                                                   |        |   ✔️    |
| [PASCAL VOC](http://host.robots.ox.ac.uk/pascal/VOC/)                                                     |   ✔️    |   ✔️    |
| Segmentation masks from [PASCAL VOC](http://host.robots.ox.ac.uk/pascal/VOC/)                             |   ✔️    |   ✔️    |
| [YOLO](https://pjreddie.com/darknet/yolo/)                                                                |   ✔️    |   ✔️    |
| [MS COCO Object Detection](http://cocodataset.org/#format-data)                                           |   ✔️    |   ✔️    |
| [TFrecord](https://www.tensorflow.org/tutorials/load_data/tfrecord)                                       |   ✔️    |   ✔️    |
| [MOT](https://motchallenge.net/)                                                                          |   ✔️    |   ✔️    |
| [LabelMe 3.0](http://labelme.csail.mit.edu/Release3.0)                                                    |   ✔️    |   ✔️    |
| [ImageNet](http://www.image-net.org)                                                                      |   ✔️    |   ✔️    |
| [CamVid](http://mi.eng.cam.ac.uk/research/projects/VideoRec/CamVid/)                                      |   ✔️    |   ✔️    |
| [WIDER Face](http://shuoyang1213.me/WIDERFACE/)                                                           |   ✔️    |   ✔️    |
| [VGGFace2](https://github.com/ox-vgg/vgg_face2)                                                           |   ✔️    |   ✔️    |
| [Market-1501](https://www.aitribune.com/dataset/2018051063)                                               |   ✔️    |   ✔️    |
| [ICDAR13/15](https://rrc.cvc.uab.es/?ch=2)                                                                |   ✔️    |   ✔️    |
| [Open Images V6](https://storage.googleapis.com/openimages/web/index.html)                                |   ✔️    |   ✔️    |
| [Cityscapes](https://www.cityscapes-dataset.com/login/)                                                   |   ✔️    |   ✔️    |
| [KITTI](http://www.cvlibs.net/datasets/kitti/)                                                            |   ✔️    |   ✔️    |
| [LFW](http://vis-www.cs.umass.edu/lfw/)                                                                   |   ✔️    |   ✔️    |
141 142

<!--lint enable maximum-line-length-->
A
Andrey Zhavoronkov 已提交
143

144 145
## Deep learning serverless functions for automatic labeling

146 147
CVAT supports automatic labelling. It can speed up the annotation process
up to 10x. Here is a list of the algorithms we support, and the platforms they
148 149
can be ran on:

150
<!--lint disable maximum-line-length-->
151

A
Ali Jahani 已提交
152 153
| Name                                                                                                    | Type       | Framework  | CPU | GPU |
| ------------------------------------------------------------------------------------------------------- | ---------- | ---------- | --- | --- |
154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169
| [Deep Extreme Cut](/serverless/openvino/dextr/nuclio)                                                   | interactor | OpenVINO   |  ✔️  |     |
| [Faster RCNN](/serverless/openvino/omz/public/faster_rcnn_inception_v2_coco/nuclio)                     | detector   | OpenVINO   |  ✔️  |     |
| [Mask RCNN](/serverless/openvino/omz/public/mask_rcnn_inception_resnet_v2_atrous_coco/nuclio)           | detector   | OpenVINO   |  ✔️  |     |
| [YOLO v3](/serverless/openvino/omz/public/yolo-v3-tf/nuclio)                                            | detector   | OpenVINO   |  ✔️  |     |
| [Object reidentification](/serverless/openvino/omz/intel/person-reidentification-retail-300/nuclio)     | reid       | OpenVINO   |  ✔️  |     |
| [Semantic segmentation for ADAS](/serverless/openvino/omz/intel/semantic-segmentation-adas-0001/nuclio) | detector   | OpenVINO   |  ✔️  |     |
| [Text detection v4](/serverless/openvino/omz/intel/text-detection-0004/nuclio)                          | detector   | OpenVINO   |  ✔️  |     |
| [YOLO v5](/serverless/pytorch/ultralytics/yolov5/nuclio)                                                | detector   | PyTorch    |  ✔️  |     |
| [SiamMask](/serverless/pytorch/foolwood/siammask/nuclio)                                                | tracker    | PyTorch    |  ✔️  |  ✔️  |
| [f-BRS](/serverless/pytorch/saic-vul/fbrs/nuclio)                                                       | interactor | PyTorch    |  ✔️  |     |
| [HRNet](/serverless/pytorch/saic-vul/hrnet/nuclio)                                                      | interactor | PyTorch    |     |  ✔️  |
| [Inside-Outside Guidance](/serverless/pytorch/shiyinzhang/iog/nuclio)                                   | interactor | PyTorch    |  ✔️  |     |
| [Faster RCNN](/serverless/tensorflow/faster_rcnn_inception_v2_coco/nuclio)                              | detector   | TensorFlow |  ✔️  |  ✔️  |
| [Mask RCNN](/serverless/tensorflow/matterport/mask_rcnn/nuclio)                                         | detector   | TensorFlow |  ✔️  |  ✔️  |
| [RetinaNet](serverless/pytorch/facebookresearch/detectron2/retinanet/nuclio)                            | detector   | PyTorch    |  ✔️  |  ✔️  |
| [Face Detection](/serverless/openvino/omz/intel/face-detection-0205/nuclio)                             | detector   | OpenVINO   |  ✔️  |     |
170 171

<!--lint enable maximum-line-length-->
N
Nikita Manovich 已提交
172

173
## License
174

175
The code is released under the [MIT License](https://opensource.org/licenses/MIT).
N
Nikita Manovich 已提交
176

177 178 179 180 181 182
This software uses LGPL licensed libraries from the [FFmpeg](https://www.ffmpeg.org) project.
The exact steps on how FFmpeg was configured and compiled can be found in the [Dockerfile](Dockerfile).

FFmpeg is an open source framework licensed under LGPL and GPL.
See [https://www.ffmpeg.org/legal.html](https://www.ffmpeg.org/legal.html). You are solely responsible
for determining if your use of FFmpeg requires any
183
additional licenses. CVAT.ai Corporation is not responsible for obtaining any
184 185 186
such licenses, nor liable for any licensing fees due in
connection with your use of FFmpeg.

187
## Where to ask questions
I
idriss 已提交
188

189
[Gitter chat][gitter-url]: you can post CVAT usage related questions there.
190
Typically they get answered fast by the core team or community. There you can also browse other common questions.
I
idriss 已提交
191

192
[Discord][discord-url] is the place to also ask questions or discuss any other stuff related to CVAT.
193 194

[GitHub issues](https://github.com/cvat-ai/cvat/issues): please post them for feature requests or bug reports.
195
If it's a bug, please add the steps to reproduce it.
196

197 198
[\#cvat](https://stackoverflow.com/search?q=%23cvat) tag on StackOverflow is one more way to ask
questions and get our support.
V
Vitaliy Nishukov 已提交
199

200
[contact@cvat.ai](mailto:contact+github@cvat.ai): reach out to us with feedback, comments, or inquiries.
201 202

## Links
V
Vitaliy Nishukov 已提交
203

204 205 206
- [Intel AI blog: New Computer Vision Tool Accelerates Annotation of Digital Images and Video](https://www.intel.ai/introducing-cvat)
- [Intel Software: Computer Vision Annotation Tool: A Universal Approach to Data Annotation](https://software.intel.com/en-us/articles/computer-vision-annotation-tool-a-universal-approach-to-data-annotation)
- [VentureBeat: Intel open-sources CVAT, a toolkit for data labeling](https://venturebeat.com/2019/03/05/intel-open-sources-cvat-a-toolkit-for-data-labeling/)
207

208 209
  <!-- prettier-ignore-start -->
  <!-- Badges -->
210

211 212
[docker-server-pulls-img]: https://img.shields.io/docker/pulls/cvat/server.svg?style=flat-square&label=server%20pulls
[docker-server-image-url]: https://hub.docker.com/r/cvat/server
A
Andrey Zhavoronkov 已提交
213

214 215
[docker-ui-pulls-img]: https://img.shields.io/docker/pulls/cvat/ui.svg?style=flat-square&label=UI%20pulls
[docker-ui-image-url]: https://hub.docker.com/r/cvat/ui
A
Andrey Zhavoronkov 已提交
216 217 218 219

[ci-img]: https://github.com/cvat-ai/cvat/workflows/CI/badge.svg?branch=develop
[ci-url]: https://github.com/cvat-ai/cvat/actions

K
Kirill Sizov 已提交
220
[gitter-img]: https://img.shields.io/gitter/room/opencv-cvat/public?style=flat
221
[gitter-url]: https://gitter.im/opencv-cvat
A
Andrey Zhavoronkov 已提交
222 223 224 225

[coverage-img]: https://coveralls.io/repos/github/cvat-ai/cvat/badge.svg?branch=develop
[coverage-url]: https://coveralls.io/github/cvat-ai/cvat?branch=develop

226 227
[doi-img]: https://zenodo.org/badge/139156354.svg
[doi-url]: https://zenodo.org/badge/latestdoi/139156354
228 229 230

[discord-img]: https://img.shields.io/discord/1000789942802337834?label=discord
[discord-url]: https://discord.gg/fNR3eXfk6C