Welcome to the documentation for the Dataset Management Framework (Datumaro).
The Datumaro is a free framework and CLI tool for building, transforming,
and analyzing datasets.
It is developed and used by Intel to build, transform, and analyze annotations
and datasets in a large number of supported formats.
Our documentation provides information for AI researchers, developers,
and teams, who are working with datasets and annotations.
Datuaro as a standalone tool allows to do various dataset operations from
the command line interface:
datum --help
python -m datumaro --help
Python module
Datumaro can be used in custom scripts as a Python module. Used this way, it
allows to use its features from an existing codebase, enabling dataset
reading, exporting and iteration capabilities, simplifying integration of custom
formats and providing high performance operations:
from datumaro.components.project import Project
# load a Datumaro project
project = Project('directory')# create a dataset
dataset = project.working_tree.make_dataset()# keep only annotated images
dataset.select(lambda item:len(item.annotations)!=0)# change dataset labels
dataset.transform('remap_labels',{'cat':'dog',# rename cat to dog'truck':'car',# rename truck to car'person':'',# remove this label}, default='delete')# remove everything else# iterate over dataset elementsfor item in dataset:print(item.id, item.annotations)# export the resulting dataset in COCO format
dataset.export('dst/dir','coco')# optionally, release the project resources
project.close()
Convert only non-occluded annotations from a
CVAT project to TFrecord:
# export Datumaro dataset in CVAT UI, extract somewhere, go to the project dir
datum filter -e '/item/annotation[occluded="False"]' --mode items+anno
datum export --format tf_detection_api -- --save-images
Annotate MS COCO dataset, extract image subset, re-annotate it in
CVAT, update old dataset:
# Download COCO dataset http://cocodataset.org/#download# Put images to coco/images/ and annotations to coco/annotations/
datum create
datum import --format coco <path/to/coco>
datum export --filter '/image[images_I_dont_like]' --format cvat
# import dataset and images to CVAT, re-annotate# export Datumaro project, extract to 'reannotation-upd'
datum project update reannotation-upd
datum export --format coco
Annotate instance polygons in
CVAT, export as masks in COCO:
Apply an OpenVINO detection model to some COCO-like dataset,
then compare annotations with ground truth and visualize in TensorBoard:
datum create
datum import --format coco <path/to/coco># create model results interpretation script
datum model add -n mymodel openvino \
--weights model.bin --description model.xml \
--interpretation-script parse_results.py
datum model run --model -n mymodel --output-dir mymodel_inference/
datum diff mymodel_inference/ --format tensorboard --output-dir diff
Change colors in PASCAL VOC-like .png masks:
datum create
datum import --format voc <path/to/voc/dataset># Create a color map file with desired colors:## label : color_rgb : parts : actions# cat:0,0,255::# dog:255,0,0::## Save as mycolormap.txt
datum export --format voc_segmentation -- --label-map mycolormap.txt
# add "--apply-colormap=0" to save grayscale (indexed) masks# check "--help" option for more info# use "datum --loglevel debug" for extra conversion info
Create a custom COCO-like dataset:
import numpy as np
from datumaro.components.annotation import(
AnnotationType, Bbox, LabelCategories,)from datumaro.components.extractor import DatasetItem
from datumaro.components.dataset import Dataset
dataset = Dataset([
DatasetItem(id=0, image=np.ones((5,5,3)),
annotations=[
Bbox(1,2,3,4, label=0),]),# ...], categories=['cat','dog'])
dataset.export('test_dataset/','coco')
2 - Datumaro Design
Concept
Datumaro is:
a tool to build composite datasets and iterate over them
a tool to create and maintain datasets
Version control of annotations and images
Publication (with removal of sensitive information)
Editing
Joining and splitting
Exporting, format changing
Image preprocessing
a dataset storage
a tool to debug datasets
A network can be used to generate
informative data subsets (e.g. with false-positives)
to be analyzed further
Requirements
User interfaces
a library
a console tool with visualization means
Targets: single datasets, composite datasets, single images / videos
Built-in support for well-known annotation formats and datasets:
CVAT, COCO, PASCAL VOC, Cityscapes, ImageNet
Extensibility with user-provided components
Lightweightness - it should be easy to start working with Datumaro
Minimal dependency on environment and configuration
It should be easier to use Datumaro than writing own code
for computation of statistics or dataset manipulations
exploration of network prediction uncertainty (aka Bayessian approach)
Use case: explanation of network “quality”, “stability”, “certainty”
adversarial attacks on networks
dataset minification / reduction
Use case: removal of redundant information to reach the same network quality
with lesser training time
dataset expansion and filtration of additions
Use case: add only important data
guidance for key frame selection for tracking (paper)
Use case: more effective annotation, better predictions
RC 1 vision
CVAT integration
Datumaro needs to be integrated with CVAT,
extending CVAT UI capabilities regarding task and project operations.
It should be capable of downloading and processing data from CVAT.
User
|
v
+------------------+
| CVAT |
+--------v---------+ +------------------+ +--------------+
| Datumaro module | ----> | Datumaro project | <---> | Datumaro CLI | <--- User
+------------------+ +------------------+ +--------------+
Interfaces
Python API for user code
Installation as a package
Installation with pip by name
A command-line tool for dataset manipulations
Features
Dataset format support (reading, writing)
Own format
CVAT
COCO
PASCAL VOC
YOLO
TF Detection API
Cityscapes
ImageNet
Dataset visualization (show)
Ability to visualize a dataset
with TensorBoard
Calculation of statistics for datasets
Pixel mean, std
Object counts (detection scenario)
Image-Class distribution (classification scenario)
Pixel-Class distribution (segmentation scenario)
Image similarity clusters
Custom statistics
Dataset building
Composite dataset building
Class remapping
Subset splitting
Dataset filtering (extract)
Dataset merging (merge)
Dataset item editing (edit)
Dataset comparison (diff)
Annotation-annotation comparison
Annotation-inference comparison
Annotation quality estimation (for CVAT)
Provide a simple method to check
annotation quality with a model and generate summary
Ability to run a model on a dataset and read the results
CVAT-integration features
Task export
Datumaro project export
Dataset export
Original raw data (images, a video file) can be downloaded (exported)
together with annotations or just have links
on CVAT server (in future, support S3, etc)
Be able to use local files instead of remote links
Specify cache directory
Use case “annotate for model training”
create a task
annotate
export the task
convert to a training format
train a DL model
Use case “annotate - reannotate problematic images - merge”
# From PyPI:
pip install datumaro[default]# From the GitHub repository:
pip install'git+https://github.com/openvinotoolkit/datumaro[default]'
Read more about choosing between datumaro and datumaro[default]here.
Plugins
Datumaro has many plugins, which are responsible for dataset formats,
model launchers and other optional components. If a plugin has dependencies,
they can require additional installation. You can find the list of all the
plugin dependencies in the plugins section.
Customizing installation
Datumaro has the following installation options:
pip install datumaro - for core library functionality
pip install datumaro[default] - for normal CLI experience
In restricted installation environments, where some dependencies are
not available, or if you need only the core library functionality,
you can install Datumaro without extra plugins.
In some cases, installing just the core library may be not enough,
because there can be limited options of installing graphical libraries
in the system (various Docker environments, servers etc). You can select
between using opencv-python and opencv-python-headless by setting the
DATUMARO_HEADLESS environment variable to 0 or 1 before installing
the package. It requires installation from sources (using --no-binary):
This option can’t be covered by extras due to Python packaging system
limitations.
When installing directly from the repository, you can change the
installation branch with ...@<branch_name>. Also use --force-reinstall
parameter in this case. It can be useful for testing of unreleased
versions from GitHub pull requests.
from datumaro.components.project import Project
from datumaro.components.dataset import Dataset
from datumaro.components.extractor import Label, Bbox, DatasetItem
...
dataset = Dataset.import_from(path,format)...
Glossary
Basic concepts:
Dataset - A collection of dataset items, which consist of media and
associated annotations.
Dataset item - A basic single element of the dataset. Also known as
“sample”, “entry”. In different datasets it can be an image, a video
frame, a whole video, a 3d point cloud etc. Typically, has corresponding
annotations.
(Datumaro) Project - A combination of multiple datasets, plugins,
models and metadata.
Project versioning concepts:
Data source - A link to a dataset or a copy of a dataset inside a project.
Basically, a URL + dataset format name.
Project revision - A commit or a reference from Git (branch, tag,
HEAD~3 etc.). A revision is referenced by data hash. The HEAD
revision is the currently selected revision of the project.
Revision tree - A project build tree and plugins at
a specified revision.
Working tree - The revision tree in the working directory of a project.
data source revision - a state of a data source at a specific stage.
A revision is referenced by the data hash.
Object - The data of a revision tree or a data source revision.
An object is referenced by the data hash.
Dataset path concepts:
Dataset revpath - A path to a dataset in a special format. They are
supposed to specify paths to files, directories or data source revisions
in a uniform way in the CLI.
dataset path - a path to a dataset in the following format:
<dataset path>:<format>
format is optional. If not specified, will try to detect automatically
revision path - a path to a data source revision in a project.
The syntax is:
<project path>@<revision>:<target name>, any part can be omitted.
Default project is the current project (-p/--project CLI arg.)
Local revpaths imply that the current project is used and this part
should be omitted.
Default revision is the working tree of the project
Default build target is project
If a path refers to project (i.e. target name is not set, or
this target is exactly specified), the target dataset is the result of
joining all the project data sources.
Otherwise, if the path refers to a data source revision, the
corresponding stage from the revision build tree will be used.
Dataset building concepts:
Stage - A revision of a dataset - the original dataset or its modification
after transformation, filtration or something else. A build tree node.
A stage is referred by a name.
Build tree - A directed graph (tree) with root nodes at data sources
and a single top node called project, which represents
a joined dataset.
Each data source has a starting root node, which corresponds to the
original dataset. The internal graph nodes are stages.
Build target - A data source or a stage name. Data source names correspond
to the last stages of data sources.
Pipeline - A subgraph of a stage, which includes all the ancestors.
Other:
Transform - A transformation operation over dataset elements. Examples
are image renaming, image flipping, image and subset renaming,
label remapping etc. Corresponds to the transform command.
Command-line workflow
In Datumaro, most command-line commands operate on projects, but there are
also few commands operating on datasets directly. There are 2 basic ways
to use Datumaro from the command-line:
Use the convert, diff, merge
commands directly on existing datasets
Basically, a project is a combination of datasets, models and environment.
A project can contain an arbitrary number of datasets (data sources).
A project acts as a manager for them and allows to manipulate them
separately or as a whole, in which case it combines dataset items
from all the sources into one composite dataset. You can manage separate
datasets in a project by commands in the datum source
command line context.
Note that modifying operations (transform, filter, patch)
are applied in-place to the datasets by default.
If you want to interact with models, you need to add them to the project
first using the model add command.
A typical way to obtain Datumaro projects is to export tasks in
CVAT UI.
Project data model
Datumaro tries to combine a “Git for datasets” and a build system like
make or CMake for datasets in a single solution. Currently, Project
represents a Version Control System for datasets, which is based on Git and DVC
projects. Each project Revision describes a build tree of a dataset
with all the related metadata. A build tree consists of a number of data
sources and transformation stages. Each data source has its own set of build
steps (stages). Datumaro supposes copying of datasets and working in-place by
default. Modifying operations are recorded in the project, so any of the
dataset revisions can be reproduced when needed. Multiple dataset versions can
be stored in different branches with the common data shared.
Let’s consider an example of a build tree:
There are 2 data sources in the example project. The resulting dataset
is obtained by simple merging (joining) the results of the input datasets.
“Source 1” and “Source 2” are the names of data sources in the project. Each
source has several stages with their own names. The first stage (called “root”)
represents the original contents of a data source - the data at the
user-provided URL. The following stages represent operations, which needs to
be done with the data source to prepare the resulting dataset.
Roughly, such build tree can be created by the following commands (arguments
are omitted for simplicity):
datum create
# describe the first source
datum import<...> -n source1
datum filter <...> source1
datum transform <...> source1
datum transform <...> source1
# describe the second source
datum import<...> -n source2
datum model add<...>
datum transform <...> source2
datum transform <...> source2
Now, the resulting dataset can be built with:
datum export<...>
Project layout
project/
├── .dvc/
├── .dvcignore
├── .git/
├── .gitignore
├── .datumaro/
│ ├── cache/ # object cache
│ │ └── <2 leading symbols of obj hash>/
│ │ └── <remaining symbols of obj hash>/
│ │ └── <object data>
│ │
│ ├── models/ # project-specific models
│ │
│ ├── plugins/ # project-specific plugins
│ │ ├── plugin1/ # composite plugin, a directory
│ │ | ├── __init__.py
│ │ | └── file2.py
│ │ ├── plugin2.py # simple plugin, a file
│ │ └── ...
│ │
│ ├── tmp/ # temp files
│ └── tree/ # working tree metadata
│ ├── config.yml
│ └── sources/
│ ├── <source name 1>.dvc
│ ├── <source name 2>.dvc
│ └── ...
│
├── <source name 1>/ # working directory for the source 1
│ └── <source data>
└── <source name 2>/ # working directory for the source 2
└── <source data>
Datasets and Data Sources
A project can contain an arbitrary number of Data Sources. Each Data Source
describes a dataset in a specific format. A project acts as a manager for
the data sources and allows to manipulate them separately or as a whole, in
which case it combines dataset items from all the sources into one composite
dataset. You can manage separate sources in a project by commands in
the datum source command
line context.
Datasets come in a wide variety of formats. Each dataset
format defines its own data structure and rules on how to
interpret the data. For example, the following data structure
is used in COCO format:
/dataset/
- /images/<id>.jpg
- /annotations/
Datumaro supports complete datasets, having both image data and
annotations, or incomplete ones, having annotations only.
Incomplete datasets can be used to prepare images and annotations
independently of each other, or to analyze or modify just the lightweight
annotations without the need to download the whole dataset.
Check supported formats for more info
about format specifications, supported import and export options and other
details. The list of formats can be extended by custom plugins,
check extending tips for information on this
topic.
Use cases
Let’s consider few examples describing what Datumaro does for you behind the
scene.
The first example explains how working trees, working directories and the
cache interact. Suppose, there is a dataset which we want to modify and
export in some other format. To do it with Datumaro, we need to create a
project and register the dataset as a data source:
datum create
datum import<...> -n source1
The dataset will be copied to the working directory inside the project. It
will be added to the project working tree.
After the dataset is added, we want to transform it and filter out some
irrelevant samples, so we run the following commands:
datum transform <...> source1
datum filter <...> source1
The commands modify the data source inside the working directory, inplace.
The operations done are recorded in the working tree.
Now, we want to make a new version of the dataset and make a snapshot in the
project cache. So we commit the working tree:
datum commit <...>
At this time, the data source is copied into the project cache and a new
project revision is created. The dataset operation history is saved, so
the dataset can be reproduced even if it is removed from the cache and the
working directory. Note, however, that the original dataset hash was not
computed, so Datumaro won’t be able to compare dataset hash on re-downloading.
If it is desired, consider making a commit with an unmodified data source.
After this, we do some other modifications to the dataset and make a new
commit. Note that the dataset is not cached, until a commit is done.
When the dataset is ready and all the required operations are done, we
can export it to the required format. We can export the resulting dataset,
or any previous stage.
datum export<...> source1
datum export<...> source1.stage3
Let’s extend the example. Imagine we have a project with 2 data sources.
Roughly, it corresponds to the following set of commands:
datum create
datum import<...> -n source1
datum import<...> -n source2
datum transform <...> source1 # used 3 times
datum transform <...> source2 # used 5 times
Then, for some reasons, the project cache was cleaned from source1 revisions.
We also don’t have anything in the project working directories - suppose,
the user removed them to save disk space.
Let’s see what happens, if we call the diff command with 2 different
revisions now.
Datumaro needs to reproduce 2 dataset revisions requested so that they could
be read and compared. Let’s see how the first dataset is reproduced
step-by-step:
source1.stage2 will be looked for in the project cache. It won’t be
found, since the cache was cleaned.
Then, Datumaro will look for previous source revisions in the cache
and won’t find any.
The project can be marked read-only, if we are not working with the
“current” project (which is specified by the -p/--project command
parameter). In the example, the command is datum diff rev1:... rev2:...,
which means there is a project in the current directory, so the project
we are working with is not read-only. If a command target was specified as
datum diff <project>@<rev>:<source>, the project would be loaded
as read-only. If a project is read-only, we can’t do anything more to
reproduce the dataset and can only exit with an error (3a). The reason for
such behavior is that the dataset downloading can be quite expensive (in
terms of time, disk space etc.). It is supposed, that such side-effects
should be controlled manually.
If the project is not read-only (3b), Datumaro will try to download
the original dataset and reproduce the resulting dataset. The data hash
will be computed and hashes will be compared (if the data source had hash
computed on addition). On success, the data will be put into the cache.
The downloaded dataset will be read and the remaining operations from the
source history will be re-applied.
The resulting dataset might be cached in some cases.
The resulting dataset is returned.
The source2 will be looked for the same way. In our case, it will be found
in the cache and returned. Once both datasets are restored and read, they
are compared.
Consider other situation. Let’s try to export the source1. Suppose
we have a clear project cache and the source1 has a copy in the working
directory.
Again, Datumaro needs to reproduce a dataset revision (stage) requested.
It looks for the dataset in the working directory and finds some data. If
there is no source working directory, Datumaro will try to reproduce the
source using the approach described above (1b).
The data hash is computed and compared with the one saved in the history.
If the hashes match, the dataset is read and returned (4).
Note: we can’t use the cached hash stored in the working tree info -
it can be outdated, so we need to compute it again.
Otherwise, Datumaro tries to detect the stage by the data hash.
If the current stage is not cached, the tree is the working tree and the
working directory is not empty, the working copy is hashed and matched
against the source stage list. If there is a matching stage, it will be
read and the missing stages will be added. The result might be cached in
some cases.
If there is no matching stage in the source history, the situation can
be contradictory. Currently, an error is raised (3b).
The resulting dataset is returned.
After the requested dataset is obtained, it is exported in the requested
format.
To sum up, Datumaro tries to restore a dataset from the project cache or
reproduce it from sources. It can be done as long as the source operations
are recorded and any step data is available. Note that cache objects share
common files, so if there are only annotation differences between datasets,
or data sources contain the same images, there will only be a single copy
of the related media files. This helps to keep storage use reasonable and
avoid unnecessary data copies.
Examples
Example: create a project, add dataset, modify, restore an old version
datum create
datum import<path/to/dataset> -f coco -n source1
datum commit -m "Added a dataset"
datum transform -t shapes_to_boxes
datum filter -e '/item/annotation[label="cat" or label="dog"]' -m i+a
datum commit -m "Transformed"
datum checkout HEAD~1 -- source1 # restore a previous revision
datum status # prints "modified source1"
datum checkout source1 # restore the last revision
datum export -f voc -- --save-images
3.3 - Supported Formats
List of supported formats:
MS COCO
(image_info, instances, person_keypoints, captions, labels,panoptic, stuff)
Datumaro does not separate datasets by tasks like classification, detection
etc. Instead, datasets can have any annotations. When a dataset is exported
in a specific format, only relevant annotations are exported.
Dataset meta info file
It is possible to use classes that are not original to the format.
To do this, use dataset_meta.json.
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('directory/path/','image_dir')
This will search for images in the directory recursively and add
them as dataset entries with names like <subdir1>/<subsubdir1>/<image_name1>.
The list of formats matches the list of supported image formats in OpenCV:
Once there is a Dataset instance, its items can be split into subsets,
renamed, filtered, joined with annotations, exported in various formats etc.
To use a video as an input, one should either create a plugin,
which splits a video into frames, or split the video manually and import images.
3.5 - Command reference
The command line is split into the separate commands and command contexts.
Contexts group multiple commands related to a specific topic, e.g.
project operations, data source operations etc. Almost all the commands
operate on projects, so the project context and commands without a context
are mostly the same. By default, commands look for a project in the current
directory. If the project you’re working on is located somewhere else, you
can pass the -p/--project <path> argument to the command.
Note: command behavior is subject to change, so this text might be
outdated,
always check the --help output of the specific command
Note: command parameters must be passed prior to the positional arguments.
Datumaro functionality is available with the datum command.
Usage:
datum [-h][--version][--loglevel LOGLEVEL][command][command args]
Parameters:
--loglevel (string) - Logging level, one of
debug, info, warning, error, critical (default: info)
--version - Print the version number and exit.
-h, --help - Print the help message and exit.
3.5.1 - Convert datasets
This command allows to convert a dataset from one format to another.
The command is a usability alias for create,
add and export and just provides
a simpler way to obtain the same results in simple cases. A list of supported
formats can be found in the --help output of this command.
This command downloads a publicly available dataset and saves it to a local
directory.
In terms of syntax, this command is similar to convert,
but instead of taking a local directory as the source, it takes a dataset ID.
A list of supported datasets and output formats can be found in the --help
output of this command.
Currently, the only source of datasets is the TensorFlow Datasets library.
Therefore, to use this command you must install TensorFlow & TFDS, which you can
do as follows:
datum download [-h] -i DATASET_ID [-f OUTPUT_FORMAT][-o DST_DIR][--overwrite][-- EXTRA_EXPORT_ARGS]
Parameters:
-h, --help - Print the help message and exit.
-i, --dataset-id (string) - ID of the dataset to download.
-f, --output-format (string) - Output format. By default, the format
of the original dataset is used.
-o, --output-dir (string) - Output directory. By default, a subdirectory
in the current directory is used.
--overwrite - Allows overwriting existing files in the output directory,
when it is not empty.
-- <extra export args> - Additional arguments for the format writer
(use -- -h for help). Must be specified after the main command arguments.
Example: download the MNIST dataset, saving it in the ImageNet text format:
datum download -i tfds:mnist -f imagenet_txt -- --save-images
3.5.3 - Create project
The command creates an empty project. A project is required for the most of
Datumaro functionality.
By default, the project is created in the current directory. To specify
another output directory, pass the -o/--output-dir parameter. If output
already directory contains a Datumaro project, an error is raised, unless
--overwrite is used.
Usage:
datum create [-h][-o DST_DIR][--overwrite]
Parameters:
-o, --output-dir (string) - Allows to specify an output directory.
The current directory is used by default.
--overwrite - Allows to overwrite existing project files in the output
directory. Any other files are not touched.
-h, --help - Print the help message and exit.
Examples:
Example: create an empty project in the my_dataset directory
datum create -o my_dataset/
Example: create a new empty project in the current directory, remove the
existing one
datum create
...
datum create --overwrite
3.5.4 - Export Datasets
This command exports a project or a source as a dataset in some format.
Check supported formats for more info
about format specifications, supported options and other details.
The list of formats can be extended by custom plugins, check extending tips
for information on this topic.
Available formats are listed in the command help output.
Dataset format writers support additional export options. To pass
such options, use the -- separator after the main command arguments.
The usage information can be printed with datum import -f <format> -- --help.
Common export options:
Most formats (where applicable) support the --save-images option, which
allows to export dataset images along with annotations. The option is
disabled be default.
If --save-images is used, the image-ext option can be passed to
specify the output image file extension (.jpg, .png etc.). By default,
tries to Datumaro keep the original image extension. This option
allows to convert all the images from one format into another.
This command allows to use the -f/--filter parameter to select dataset
elements needed for exporting. Read the filter command
description for more info about this functionality.
The command can only be applied to a project build target, a stage
or the combined project target, in which case all the targets will
be affected.
Usage:
datum export[-h][-e FILTER][--filter-mode FILTER_MODE][-o DST_DIR][--overwrite][-p PROJECT_DIR] -f FORMAT [target][-- EXTRA_FORMAT_ARGS]
Parameters:
<target> (string) - A project build target to be exported.
By default, all project targets are affected.
-f, --format (string) - Output format.
-e, --filter (string) - XML XPath filter expression for dataset items
--filter-mode (string) - The filtering mode. Default is the i mode.
-o, --output-dir (string) - Output directory. By default, a subdirectory
in the current directory is used.
--overwrite - Allows overwriting existing files in the output directory,
when it is not empty.
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
-- <extra format args> - Additional arguments for the format writer
(use -- -h for help). Must be specified after the main command arguments.
Example: save a project as a VOC-like dataset, include images, convert
images to PNG from other formats.
This command allows to extract a sub-dataset from a dataset. The new dataset
includes only items satisfying some condition. The XML XPath
is used as a query format.
The command can be applied to a dataset or a project build target,
a stage or the combined project target, in which case all the project
targets will be affected. A build tree stage will be recorded
if --stage is enabled, and the resulting dataset(-s) will be
saved if --apply is enabled.
By default, datasets are updated in-place. The -o/--output-dir
option can be used to specify another output directory. When
updating in-place, use the --overwrite parameter (in-place
updates fail by default to prevent data loss), unless a project
target is modified.
The current project (-p/--project) is also used as a context for
plugins, so it can be useful for dataset paths having custom formats.
When not specified, the current project’s working tree is used.
There are several filtering modes available (the -m/--mode parameter).
Supported modes:
i, items
a, annotations
i+a, a+i, items+annotations, annotations+items
When filtering annotations, use the items+annotations
mode to point that annotation-less dataset items should be
removed, otherwise they will be kept in the resulting dataset.
To select an annotation, write an XPath that returns annotation
elements (see examples).
Item representations can be printed with the --dry-run parameter:
The command can only be applied to a project build target, a stage or the
combined project target, in which case all the targets will be affected.
A build tree stage will be added if --stage is enabled, and the resulting
dataset(-s) will be saved if --apply is enabled.
Usage:
datum filter [-h][-e FILTER][-m MODE][--dry-run][--stage STAGE][--apply APPLY][-o DST_DIR][--overwrite][-p PROJECT_DIR][target]
Parameters:
<target> (string) - Target
dataset revpath.
By default, filters all targets of the current project.
-e, --filter (string) - XML XPath filter expression for dataset items
-m, --mode (string) - The filtering mode. Default is the i mode.
--dry-run - Print XML representations of the filtered dataset and exit.
--stage (bool) - Include this action as a project build step.
If true, this operation will be saved in the project
build tree, allowing to reproduce the resulting dataset later.
Applicable only to main project targets (i.e. data sources
and the project target, but not intermediate stages). Enabled by default.
--apply (bool) - Run this command immediately. If disabled, only the
build tree stage will be written. Enabled by default.
-o, --output-dir (string) - Output directory. Can be omitted for
main project targets (i.e. data sources and the project target, but not
intermediate stages) and dataset targets. If not specified, the results
will be saved inplace.
--overwrite - Allows to overwrite existing files in the output directory,
when it is specified and is not empty.
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
Example: extract a dataset with images with width < height
datum filter \
-p test_project \
-e '/item[image/width < image/height]'
Example: extract a dataset with images of the train subset
datum filter \
-p test_project \
-e '/item[subset="train"]'
Example: extract a dataset with only large annotations of the cat class and
any non-persons
datum filter \
-p test_project \
--mode annotations \
-e '/item/annotation[(label="cat" and area > 99.5) or label!="person"]'
Example: extract a dataset with non-occluded annotations, remove empty images.
Use data only from the “s1” source of the project.
datum create
datum import --format voc -i <path/to/dataset1/> --name s1
datum import --format voc -i <path/to/dataset2/> --name s2
datum filter s1 \
-m i+a -e '/item/annotation[occluded="False"]'
3.5.6 - Merge Datasets
Consider the following task: there is a set of images (the original dataset)
we want to annotate. Suppose we did this manually and/or automated it
using models, and now we have few sets of annotations for the same images.
We want to merge them and produce a single set of high-precision annotations.
Another use case: there are few datasets with different sets of images
and labels, which we need to combine in a single dataset. If the labels
were the same, we could just join the datasets. But in this case we need
to merge labels and adjust the annotations in the resulting dataset.
In Datumaro, it can be done with the merge command. This command merges 2
or more datasets and checks annotations for errors.
In simple cases, when dataset images do not intersect and new
labels are not added, the recommended way of merging is using
the patch command.
It will offer better performance and provide the same results.
Datasets are merged by items, and item annotations are merged by finding the
unique ones across datasets. Annotations are matched between matching dataset
items by distance. Spatial annotations are compared by the applicable distance
measure (IoU, OKS, PDJ etc.), labels and annotation attributes are selected
by voting. Each set of matching annotations produces a single annotation in
the resulting dataset. The score (a number in the range [0; 1]) attribute
indicates the agreement between different sources in the produced annotation.
The working time of the function can be estimated as
O( (summary dataset length) * (dataset count) ^ 2 * (item annotations) ^ 2 )
This command also allows to merge datasets with different, or partially
overlapping sets of labels (which is impossible by simple joining).
During the process, some merge conflicts can appear. For example,
it can be mismatching dataset images having the same ids, label voting
can be unsuccessful if quorum is not reached (the --quorum parameter),
bboxes may be too close (the -iou parameter) etc. Found merge
conflicts, missing items or annotations, and other errors are saved into
an output .json file.
In Datumaro, annotations can be grouped. It can be useful to represent
different parts of a single object - for example, it can be different parts
of a human body, parts of a vehicle etc. This command allows to check
annotation groups for completeness with the -g/--groups option. If used,
this parameter must specify a list of labels for annotations that must be
in the same group. It can be particularly useful to check if separate
keypoints are grouped and all the necessary object components in the same
group.
This command has multiple forms:
1) datum merge <revpath>2) datum merge <revpath><revpath>...
1 - Merges the current project’s main target (“project”)
in the working tree with the specified dataset.
2 - Merges the specified datasets.
Note that the current project is not included in the list of merged
sources automatically.
The command supports passing extra exporting options for the output
dataset. The format can be specified with the -f/--format option.
Extra options should be passed after the main arguments
and after the -- separator. Particularly, this is useful to include
images in the output dataset with --save-images.
-iou, --iou-thresh (number) - IoU matching threshold for spatial
annotations (both maximum inter-cluster and pairwise). Default is 0.25.
--quorum (number) - Minimum count of votes for a label or attribute
to be counted. Default is 0.
-g, --groups (string) - A comma-separated list of label names in
annotation groups to check. The ? postfix can be added to a label to
make it optional in the group (repeatable)
-oconf, --output-conf-thresh (number) - Confidence threshold for output
annotations to be included in the resulting dataset. Default is 0.
-o, --output-dir (string) - Output directory. By default, a new directory
is created in the current directory.
--overwrite - Allows to overwrite existing files in the output directory,
when it is specified and is not empty.
-f, --format (string) - Output format. The default format is datumaro.
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
-- <extra format args> - Additional arguments for the format writer
(use -- -h for help). Must be specified after the main command arguments.
Examples:
Merge 4 (partially-)intersecting projects,
consider voting successful when there are no less than 3 same votes
consider shapes intersecting when IoU >= 0.6
check annotation groups to have person, hand, head and foot
(? is used for optional parts)
Merge images and annotations from 2 datasets in COCO format:
datum merge dataset1/:image_dir dataset2/:coco dataset3/:coco
Check groups of the merged dataset for consistency:
look for groups consisting of person, handhead, footdatum merge project1/ project2/ -g 'person,hand?,head,foot?'
Merge two datasets, specify formats:
datum merge path/to/dataset1:voc path/to/dataset2:coco
Merge the current working tree and a dataset:
datum merge path/to/dataset2:coco
Merge a source from a previous revision and a dataset:
datum merge HEAD~2:source-2 path/to/dataset2:yolo
Merge datasets and save in different format:
datum merge -f voc dataset1/:yolo path2/:coco -- --save-images
3.5.7 - Patch Datasets
Updates items of the first dataset with items from the second one.
By default, datasets are updated in-place. The -o/--output-dir
option can be used to specify another output directory. When
updating in-place, use the --overwrite parameter along with the
--save-images export option (in-place updates fail by default
to prevent data loss).
Unlike the regular project data source joining,
the datasets are not required to have the same labels. The labels from
the “patch” dataset are projected onto the labels of the patched dataset,
so only the annotations with the matching labels are used, i.e.
all the annotations having unknown labels are ignored. Currently,
this command doesn’t allow to update the label information in the
patched dataset.
The command supports passing extra exporting options for the output
dataset. The extra options should be passed after the main arguments
and after the -- separator. Particularly, this is useful to include
images in the output dataset with --save-images.
This command can be applied to the current project targets or
arbitrary datasets outside a project. Note that if the target dataset
is read-only (e.g. if it is a project, stage or a cache entry),
the output directory must be provided.
The current project (-p/--project) is also used as a context for
plugins, so it can be useful for dataset paths having custom formats.
When not specified, the current project’s working tree is used.
-o, --output-dir (string) - Output directory. By default, a new directory
is created in the current directory.
--overwrite - Allows to overwrite existing files in the output directory,
when it is specified and is not empty.
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
Distance comparison options:
--iou-thresh (number) - The IoU threshold for spatial annotations
(default is 0.5).
-f, --format (string) - Output format, one of simple
(text files and images) and tensorboard (a TB log directory)
Equality comparison options:
-iia, --ignore-item-attr (string) - Ignore an item attribute (repeatable)
-ia, --ignore-attr (string) - Ignore an annotation attribute (repeatable)
-if, --ignore-field (string) - Ignore an annotation field (repeatable)
Default is id and group
--match-images - Match dataset items by image pixels instead of ids
--all - Include matches in the output. By default, only differences are
printed.
Examples:
Compare two projects by distance, match boxes if IoU > 0.7,
save results to TensorBoard:
datum diff other/project -o diff/ -f tensorboard --iou-thresh 0.7
Compare two projects for equality, exclude annotation groups
and the is_crowd attribute from comparison:
datum diff other/project/ -if group -ia is_crowd
Compare two datasets, specify formats:
datum diff path/to/dataset1:voc path/to/dataset2:coco
Compare the current working tree and a dataset:
datum diff path/to/dataset2:coco
Compare a source from a previous revision and a dataset:
datum diff HEAD~2:source-2 path/to/dataset2:yolo
Compare a dataset with model inference
datum create
datum import<...>
datum model add mymodel <...>
datum transform <...> -o inference
datum diff inference -o diff
3.5.9 - Print dataset info
This command outputs high level dataset information such as sample count,
categories and subsets.
Usage:
datum info [-h][--all][-p PROJECT_DIR][revpath]
Parameters:
<target> (string) - Target dataset revpath.
By default, prints info about the joined project dataset.
--all - Print all the information: do not fold long lists of labels etc.
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
Examples:
Print info about a project dataset:
datum info -p test_project/
Print info about a COCO-like dataset:
datum info path/to/dataset:coco
This command allows to fix the current state of a project and
create a new revision from the working tree.
By default, this command checks sources in the working tree for
changes. If there are unknown changes found, an error will be raised,
unless --allow-foreign is used. If such changes are committed,
the source will only be available for reproduction from the project
cache, because Datumaro will not know how to repeat them.
The command will add the sources into the project cache. If you only
need to record revision metadata, you can use the --no-cache parameter.
This can be useful if you want to save disk space and/or have a backup copy
of datasets used in the project.
If there are no changes found, the command will stop. To allow empty
commits, use --allow-empty.
Usage:
datum commit [-h] -m MESSAGE [--allow-empty][--allow-foreign][--no-cache][-p PROJECT_DIR]
Parameters:
--allow-empty - Allow commits with no changes
--allow-foreign - Allow commits with changes made not by Datumaro
--no-cache - Don’t put committed datasets into cache, save only metadata
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
Example:
datum create
datum import -f coco <path/to/coco/>
datum commit -m "Added COCO"
3.5.13 - Transform Dataset
Often datasets need to be modified during preparation for model training and
experimenting. In trivial cases it can be done manually - e.g. image renaming
or label renaming. However, in more complex cases even simple modifications
can require too much efforts, distracting the user from the real work.
Datumaro provides the datum transform command to help in such cases.
This command allows to modify dataset images or annotations all at once.
This command is designed for batch dataset processing, so if you only
need to modify few elements of a dataset, you might want to use
other approaches for better performance. A possible solution can be
a simple script, which uses Datumaro API.
The command can be applied to a dataset or a project build target,
a stage or the combined project target, in which case all the project
targets will be affected. A build tree stage will be recorded
if --stage is enabled, and the resulting dataset(-s) will be
saved if --apply is enabled.
By default, datasets are updated in-place. The -o/--output-dir
option can be used to specify another output directory. When
updating in-place, use the --overwrite parameter (in-place
updates fail by default to prevent data loss), unless a project
target is modified.
The current project (-p/--project) is also used as a context for
plugins, so it can be useful for dataset paths having custom formats.
When not specified, the current project’s working tree is used.
<target> (string) - Target
dataset revpath.
By default, transforms all targets of the current project.
-t, --transform (string) - Transform method name
--stage (bool) - Include this action as a project build step.
If true, this operation will be saved in the project
build tree, allowing to reproduce the resulting dataset later.
Applicable only to main project targets (i.e. data sources
and the project target, but not intermediate stages). Enabled by default.
--apply (bool) - Run this command immediately. If disabled, only the
build tree stage will be written. Enabled by default.
-o, --output-dir (string) - Output directory. Can be omitted for
main project targets (i.e. data sources and the project target, but not
intermediate stages) and dataset targets. If not specified, the results
will be saved inplace.
--overwrite - Allows to overwrite existing files in the output directory,
when it is specified and is not empty.
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
<extra args> - The list of extra transformation parameters. Should be
passed after the -- separator after the main command arguments. See
transform descriptions for info about extra parameters. Use the --help
option to print parameter info.
Examples:
Split a VOC-like dataset randomly:
datum transform -t random_split --overwrite path/to/dataset:voc
Rename images in a project data source by a regex from frame_XXX to XXX:
datum create <...>
datum import<...> -n source-1
datum transform -t rename source-1 -- -e '|frame_(\d+)|\\1|'
Built-in transforms
Basic dataset item manipulations:
rename - Renames dataset items by regular expression
id_from_image_name - Renames dataset items to their image filenames
reindex - Renames dataset items with numbers
ndr - Removes duplicated images from dataset
sampler - Runs inference and leaves only the most representative images
resize - Resizes images and annotations in the dataset
Subset manipulations:
random_split - Splits dataset into subsets randomly
split - Splits dataset into subsets for classification, detection,
segmentation or re-identification
map_subsets - Renames and removes subsets
Annotation manipulations:
remap_labels - Renames, adds or removes labels in dataset
project_labels - Sets dataset labels to the requested sequence
shapes_to_boxes - Replaces spatial annotations with bounding boxes
boxes_to_masks - Converts bounding boxes to instance masks
polygons_to_masks - Converts polygons to instance masks
masks_to_polygons - Converts instance masks to polygons
anns_to_labels - Replaces annotations having labels with label annotations
merge_instance_segments - Merges grouped spatial annotations into a mask
crop_covered_segments - Removes occluded segments of covered masks
bbox_value_decrement - Subtracts 1 from bbox coordinates
Examples:
Split a dataset randomly to train and test subsets, ratio is 2:1
datum transform -t random_split -- --subset train:.67 --subset test:.33
Split a dataset for a specific task. The tasks supported are
classification, detection, segmentation and re-identification.
datum transform -t boxes_to_masks
datum transform -t masks_to_polygons
datum transform -t polygons_to_masks
datum transform -t shapes_to_boxes
Set dataset labels to {person, cat, dog}, remove others, add missing.
Original labels (can be any): cat, dog, elephant, human
New labels: person (added), cat (kept), dog (kept)
datum transform -t project_labels -- -l person -l cat -l dog
Remap dataset labels, person to car and cat to dog,
keep bus, remove others
datum transform -t rename -- -e '|pattern|replacement|'
datum transform -t rename -- -e '|frame_(\d+)|\\1|'
Create a dataset from K the most hard items for a model. The dataset will
be split into the sampled and unsampled subsets, based on the model
confidence, which is stored in the scores annotation attribute.
There are five methods of sampling (the -m/--method option):
topk - Return the k with high uncertainty data
lowk - Return the k with low uncertainty data
randk - Return the random k data
mixk - Return half to topk method and the rest to lowk method
randtopk - First, select 3 times the number of k randomly, and return
the topk among them.
Remove duplicated images from a dataset. Keep at most N resulting images.
Available sampling options (the -e parameter):
random - sample from removed data randomly
similarity - sample from removed data with ascending
Available sampling methods (the -u parameter):
uniform - sample data with uniform distribution
inverse - sample data with reciprocal of the number
datum transform -t ndr -- \
-w train \
-a gradient \
-k 100\
-e random \
-u uniform
Resize dataset images and annotations. Supports upscaling, downscaling
and mixed variants.
datum transform -t resize -- -dw 256 -dh 256
3.5.14 - Checkout
This command allows to restore a specific project revision in the project
tree or to restore separate revisions of sources. A revision can be a commit
hash, branch, tag, or any relative reference in the Git format.
This command has multiple forms:
1) datum checkout <revision>2) datum checkout [--]<source1>...
3) datum checkout <revision>[--]<source1><source2>...
1 - Restores a revision and all the corresponding sources in the
working directory. If there are conflicts between modified files in the
working directory and the target revision, an error is raised, unless
--force is used.
2, 3 - Restores only selected sources from the specified revision.
The current revision is used, when not set.
“–” can be used to separate source names and revisions:
datum checkout name - will look for revision “name”
datum checkout -- name - will look for source “name” in the current
revision
Usage:
datum checkout [-h][-f][-p PROJECT_DIR][rev][--][sources [sources ...]]
Parameters:
--force - Allows to overwrite unsaved changes in case of conflicts
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
Examples:
Restore the previous revision:
datum checkout HEAD~1
Restore the saved version of a source in the working tree
datum checkout -- source-1
Restore a previous version of a source
datum checkout 33fbfbe my-source
3.5.15 - Status
This command prints the summary of the source changes between
the working tree of a project and its HEAD revision.
Prints lines in the following format:
<status> <source name>
The list of possible status values:
modified - the source data exists and it is changed
foreign_modified - the source data exists and it is changed,
but Datumaro does not know about the way the differences were made.
If changes are committed, they will only be available for reproduction
from the project cache.
added - the source was added in the working tree
removed - the source was removed from the working tree. This status won’t
be reported if just the source data is removed in the working tree.
In such situation the status will be missing.
missing - the source data is removed from the working directory.
The source still can be restored from the project cache or reproduced.
Usage:
datum status [-h][-p PROJECT_DIR]
Parameters:
-p, --project (string) - Directory of the project to operate on
(default: current directory).
3.5.17 - Run model inference explanation (explain)
Runs an explainable AI algorithm for a model.
This tool is supposed to help an AI developer to debug a model and a dataset.
Basically, it executes model inference and tries to find relation between
inputs and outputs of the trained model, i.e. determine decision boundaries
and belief intervals for the classifier.
Currently, the only available algorithm is RISE (article),
which runs model a single time and then re-runs a model multiple times on
each image to produce a heatmap of activations for each output of the
first inference. Each time a part of the input image is masked. As a result,
we obtain a number heatmaps, which show, how specific image pixels affected
the inference result. This algorithm doesn’t require any special information
about the model, but it requires the model to return all the outputs and
confidences. The original algorithm supports only classification scenario,
but Datumaro extends it for detection models.
The following use cases available:
RISE for classification
RISE for object detection
Usage:
datum explain [-h] -m MODEL [-o SAVE_DIR][-p PROJECT_DIR][target]{rise}[RISE_ARGS]
Parameters:
<target> (string) - Target
dataset revpath.By default,
uses the whole current project. An image path can be specified instead.
<image path> - a path to the file.
<revpath> - a dataset path or a revision path.
<method> (string) - The algorithm to use. Currently, only rise
is supported.
-m, --model (string) - The model to use for inference
-o, --output-dir (string) - Directory to save results to
(default: display only)
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
RISE options:
-s, --max-samples (number) - Number of algorithm model runs per image
(default: mask size ^ 2).
--mw, --mask-width (number) - Mask width in pixels (default: 7)
--mh, --mask-height (number) - Mask height in pixels (default: 7)
--iou, --iou-thresh (number) - IoU match threshold for detections
(default: 0.9)
--nms, --nms-iou-thresh (number) - IoU match threshold for detections
for non-maxima suppression (default: no NMS)
--conf, --det-conf-thresh (number) - Confidence threshold for
detections (default: include all)
-b, --batch-size (number) - Batch size for inference (default: 1)
--display - Visualize results during computations
Examples:
Run RISE on an image, display results:
datum explain path/to/image.jpg -m mymodel rise --max-samples 50
Run RISE on a source revision:
datum explain HEAD~1:source-1 -m model rise
Run inference explanation on a single image with online visualization
datum create <...>
datum model add mymodel <...>
datum explain -t image.png -m mymodel \
rise --max-samples 1000 --display
Note: this algorithm requires the model to return
all (or a reasonable amount) the outputs and confidences unfiltered,
i.e. all the Label annotations for classification models and
all the Bboxes for detection models.
You can find examples of the expected model outputs in tests/test_RISE.py
For OpenVINO models the output processing script would look like this:
from datumaro.components.extractor import*# return a significant number of output boxes to make multiple runs# statistically correct and meaningful
max_det =1000defprocess_outputs(inputs, outputs):# inputs = model input, array or images, shape = (N, C, H, W)# outputs = model output, shape = (N, 1, K, 7)# results = conversion result, [ [ Annotation, ... ], ... ]
results =[]forinput, output inzip(inputs, outputs):
input_height, input_width =input.shape[:2]
detections = output[0]
image_results =[]for i, det inenumerate(detections):
label =int(det[1])
conf =float(det[2])
x =max(int(det[3]* input_width),0)
y =max(int(det[4]* input_height),0)
w =min(int(det[5]* input_width - x), input_width)
h =min(int(det[6]* input_height - y), input_height)
image_results.append(Bbox(x, y, w, h,
label=label, attributes={'score': conf}))
results.append(image_results[:max_det])return results
3.5.18 - Models
Register model
Datumaro can execute deep learning models in various frameworks. Check
the plugins section
for more info.
Supported frameworks:
OpenVINO
Custom models via custom launchers
Models need to be added to the Datumaro project first. It can be done with
the datum model add command.
Usage:
datum model add[-h][-n NAME] -l LAUNCHER [--copy][--no-check][-p PROJECT_DIR][-- EXTRA_ARGS]
Parameters:
-l, --launcher (string) - Model launcher name
--copy - Copy model data into project. By default, only the link is saved.
--no-check - Don’t check the model can be loaded
-n, --name (string) - Name of the new model (default: generate
automatically)
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
<extra args> - Additional arguments for the model launcher
(use -- -h for help). Must be specified after the main command arguments.
Example: register an OpenVINO model
A model consists of a graph description and weights. There is also a script
used to convert model outputs to internal data structures.
datum create
datum model add\
-n <model_name> -l openvino -- \
-d <path_to_xml> -w <path_to_bin> -i <path_to_interpretation_script>
Interpretation script for an OpenVINO detection model (convert.py):
You can find OpenVINO model interpreter samples in
datumaro/plugins/openvino/samples (instruction).
from datumaro.components.extractor import*
max_det =10
conf_thresh =0.1defprocess_outputs(inputs, outputs):# inputs = model input, array or images, shape = (N, C, H, W)# outputs = model output, shape = (N, 1, K, 7)# results = conversion result, [ [ Annotation, ... ], ... ]
results =[]forinput, output inzip(inputs, outputs):
input_height, input_width =input.shape[:2]
detections = output[0]
image_results =[]for i, det inenumerate(detections):
label =int(det[1])
conf =float(det[2])if conf <= conf_thresh:continue
x =max(int(det[3]* input_width),0)
y =max(int(det[4]* input_height),0)
w =min(int(det[5]* input_width - x), input_width)
h =min(int(det[6]* input_height - y), input_height)
image_results.append(Bbox(x, y, w, h,
label=label, attributes={'score': conf}))
results.append(image_results[:max_det])return results
defget_categories():# Optionally, provide output categories - label map etc.# Example:
label_categories = LabelCategories()
label_categories.add('person')
label_categories.add('car')return{ AnnotationType.label: label_categories }
Remove Models
To remove a model from a project, use the datum model remove command.
Usage:
datum model remove [-h][-p PROJECT_DIR] name
Parameters:
<name> (string) - The name of the model to be removed
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
Example:
datum create
datum model add<...> -n model1
datum remove model1
Run Model
This command applies model to dataset images and produces a new dataset.
Usage:
datum model run
Parameters:
<target> (string) - A project build target to be used.
By default, uses the combined project target.
-m, --model (string) - Model name
-o, --output-dir (string) - Output directory. By default, results will
be stored in an auto-generated directory in the current directory.
--overwrite - Allows to overwrite existing files in the output directory,
when it is specified and is not empty.
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
Example: launch inference on a dataset
datum create
datum import<...>
datum model add mymodel <...>
datum model run -m mymodel -o inference
3.5.19 - Sources
These commands are specific for Data Sources. Read more about them here.
Import Dataset
Datasets can be added to a Datumaro project with the import command,
which adds a dataset link into the project and downloads (or copies)
the dataset. If you need to add a dataset already copied into the project,
use the add command.
Dataset format readers can provide some additional import options. To pass
such options, use the -- separator after the main command arguments.
The usage information can be printed with datum import -f <format> -- --help.
The list of currently available formats is listed in the command help output.
A dataset is imported by its URL. Currently, only local filesystem
paths are supported. The URL can be a file or a directory path
to a dataset. When the dataset is read, it is read as a whole.
However, many formats can have multiple subsets like train, val, test
etc. If you want to limit reading only to a specific subset, use
the -r/--path parameter. It can also be useful when subset files have
non-standard placement or names.
When a dataset is imported, the following things are done:
URL is saved in the project config
data in copied into the project
Each data source has a name assigned, which can be used in other commands. To
set a specific name, use the -n/--name parameter.
The dataset is added into the working tree of the project. A new commit
is not done automatically.
Usage:
datum import[-h][-n NAME] -f FORMAT [-r PATH][--no-check][-p PROJECT_DIR] url [-- EXTRA_FORMAT_ARGS]
Parameters:
<url> (string) - A file of directory path to the dataset.
-f, --format (string) - Dataset format
-r, --path (string) - A path relative to the source URL the data source.
Useful to specify a path to a subset, subtask, or a specific file in URL.
--no-check - Don’t try to read the source after importing
-n, --name (string) - Name of the new source (default: generate
automatically)
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
-- <extra format args> - Additional arguments for the format reader
(use -- -h for help). Must be specified after the main command arguments.
Example: create a project from images and annotations in different formats,
export as TFrecord for TF Detection API for model training
# 'default' is the name of the subset below
datum create
datum import -f coco_instances -r annotations/instances_default.json path/to/coco
datum import -f cvat <path/to/cvat/default.xml>
datum import -f voc_detection -r custom_subset_dir/default.txt <path/to/voc>
datum import -f datumaro <path/to/datumaro/default.json>
datum import -f image_dir <path/to/images/dir>
datum export -f tf_detection_api -- --save-images
Add Dataset
Existing datasets can be added to a Datumaro project with the add command.
The command adds a project-local directory as a data source in the project.
Unlike the import
command, it does not copy datasets and only works with local directories.
The source name is defined by the directory name.
Dataset format readers can provide some additional import options. To pass
such options, use the -- separator after the main command arguments.
The usage information can be printed with datum add -f <format> -- --help.
The list of currently available formats is listed in the command help output.
A dataset is imported as a directory. When the dataset is read, it is read
as a whole. However, many formats can have multiple subsets like train,
val, test etc. If you want to limit reading only to a specific subset,
use the -r/--path parameter. It can also be useful when subset files have
non-standard placement or names.
The dataset is added into the working tree of the project. A new commit
is not done automatically.
Usage:
datum add[-h] -f FORMAT [-r PATH][--no-check][-p PROJECT_DIR] path [-- EXTRA_FORMAT_ARGS]
Parameters:
<url> (string) - A file of directory path to the dataset.
-f, --format (string) - Dataset format
-r, --path (string) - A path relative to the source URL the data source.
Useful to specify a path to a subset, subtask, or a specific file in URL.
--no-check - Don’t try to read the source after importing
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
-- <extra format args> - Additional arguments for the format reader
(use -- -h for help). Must be specified after the main command arguments.
Example: create a project from images and annotations in different formats,
export in YOLO for model training
datum create
datum add -f coco -r annotations/instances_train.json dataset1/
datum add -f cvat dataset2/train.xml
datum export -f yolo -- --save-images
Example: add an existing dataset into a project, avoid data copying
To add a dataset, we need to have it inside the project directory:
datum create -o proj/
mv ~/my_coco/ proj/my_coco/ # move the dataset into the project directory
datum add -p proj/ -f coco proj/my_coco/
Remove Datasets
To remove a data source from a project, use the remove command.
Usage:
datum remove [-h][--force][--keep-data][-p PROJECT_DIR] name [name ...]
Parameters:
<name> (string) - The name of the source to be removed (repeatable)
-f, --force - Do not fail and stop on errors during removal
--keep-data - Do not remove source data from the working directory, remove
only project metainfo.
-p, --project (string) - Directory of the project to operate on
(default: current directory).
-h, --help - Print the help message and exit.
Example:
datum create
datum import -f voc -n src1 <path/to/dataset/>
datum remove src1
3.5.20 - Projects
Migrate project
Updates the project from an old version to the current one and saves the
resulting project in the output directory. Projects cannot be updated
inplace.
The command tries to map the old source configuration to the new one.
This can fail in some cases, so the command will exit with an error,
unless -f/--force is specified. With this flag, the command will
skip these errors an continue its work.
Usage:
datum project migrate [-h] -o DST_DIR [-f][-p PROJECT_DIR][--overwrite]
Parameters:
-o, --output-dir (string) - Output directory for the updated project
There are few ways to extend and customize Datumaro behavior, which is
supported by plugins. Check our contribution guide for
details on plugin implementation. In general, a plugin is a Python module.
It must be put into a plugin directory:
<project_dir>/.datumaro/plugins for project-specific plugins
<datumaro_dir>/plugins for global plugins
Built-in plugins
Datumaro provides several builtin plugins. Plugins can have dependencies,
which need to be installed separately.
TensorFlow
The plugin provides support of TensorFlow Detection API format, which includes
boxes and masks.
Dependencies
The plugin depends on TensorFlow, which can be installed with pip:
pip install tensorflow
# or
pip install tensorflow-gpu
# or
pip install datumaro[tf]# or
pip install datumaro[tf-gpu]
Accuracy Checker
This plugin allows to use Accuracy Checker
to launch deep learning models from various frameworks
(Caffe, MxNet, PyTorch, OpenVINO, …) through Accuracy Checker’s API.
Dependencies
The plugin depends on Accuracy Checker, which can be installed with pip:
To execute models with deep learning frameworks, they need to be installed too.
OpenVINO™
This plugin provides support for model inference with OpenVINO™.
Dependencies
The plugin depends on the OpenVINO™ Toolkit, which can be installed by
following these instructions
Dataset Formats
Dataset reading is supported by Extractors and Importers.
An Extractor produces a list of dataset items corresponding
to the dataset. An Importer creates a project from the data source location.
It is possible to add custom Extractors and Importers. To do this, you need
to put an Extractor and Importer implementation scripts to a plugin directory.
Dataset writing is supported by Converters.
A Converter produces a dataset of a specific format from dataset items.
It is possible to add custom Converters. To do this, you need to put a Converter
implementation script to a plugin directory.
Dataset Conversions (“Transforms”)
A Transform is a function for altering a dataset and producing a new one.
It can update dataset items, annotations, classes, and other properties.
A list of available transforms for dataset conversions can be extended by
adding a Transform implementation script into a plugin directory.
Model launchers
A list of available launchers for model execution can be extended by adding
a Launcher implementation script into a plugin directory.
* All sensitive arguments, such as filesystem paths or names, are sanitized
To enable the collection of telemetry data, the ISIP consent file
must exist and contain 1, otherwise telemetry will be disabled.
The ISIP file can be created/modified by an OpenVINO installer
or manually and used by other OpenVINO™ tools.
The location of the ISIP consent file depends on the OS:
Windows: %localappdata%\Intel Corporation\isip,
Linux, MacOS: $HOME/intel/isip.
4 - Dataset Management Framework (Datumaro) API and developer manual
Basics
The central part of the library is the Dataset class, which represents
a dataset and allows to iterate over its elements.
DatasetItem, an element of a dataset, represents a single
dataset entry with annotations - an image, video sequence, audio track etc.
It can contain only annotated data or meta information, only annotations, or
all of this.
Data is read (or produced) by one or many Extractors and
merged into a Dataset
The dataset is processed in some way
The dataset is saved with a Converter
Datumaro has a number of dataset and annotation features:
iteration over dataset elements
filtering of datasets and annotations by a custom criteria
working with subsets (e.g. train, val, test)
computing of dataset statistics
comparison and merging of datasets
various annotation operations
from datumaro.components.annotation import Bbox, Polygon
from datumaro.components.dataset import Dataset
from datumaro.components.extractor import DatasetItem
# Import and export a dataset
dataset = Dataset.import_from('src/dir','voc')
dataset.export('dst/dir','coco')# Create a dataset, convert polygons to masks, save in PASCAL VOC format
dataset = Dataset.from_iterable([
DatasetItem(id='image1', annotations=[
Bbox(x=1, y=2, w=3, h=4, label=1),
Polygon([1,2,3,2,4,4], label=2, attributes={'occluded':True}),]),], categories=['cat','dog','person'])
dataset.transform('polygons_to_masks')
dataset.export('dst/dir','voc')
The Dataset class
The Dataset class from the datumaro.components.dataset module represents
a dataset, consisting of multiple DatasetItems. Annotations are
represented by members of the datumaro.components.extractor module,
such as Label, Mask or Polygon. A dataset can contain items from one or
multiple subsets (e.g. train, test, val etc.), the list of dataset
subsets is available in dataset.subsets().
A DatasetItem is an element of a dataset. Its id is the name of the
corresponding image, video frame, or other media being annotated.
An item can have some attributes, associated media info and annotations.
Datasets typically have annotations, and these annotations can
require additional information to be interpreted correctly. For instance, it
can be class names, class hierarchy, keypoint connections,
class colors for masks, class attributes.
Such information is stored in dataset.categories(), which is a mapping from
AnnotationType to a corresponding ...Categories class. Each annotation type
can have its Categories. Typically, there will be at least LabelCategories;
if there are instance masks, the dataset will contain MaskCategories etc.
The “main” type of categories is LabelCategories - annotations and other
categories use label indices from this object.
The main operation for a dataset is iteration over its elements
(DatasetItems). An item corresponds to a single image, a video sequence,
etc. There are also many other operations available, such as filtration
(dataset.select()), transformation (dataset.transform()),
exporting (dataset.export()) and others. A Dataset is an Iterable and
Extractor by itself.
A Dataset can be created from scratch by its class constructor.
Categories can be set immediately or later with the
define_categories() method, but only once. You can create a dataset filled
with initial DatasetItems with Dataset.from_iterable().
If you need to create a dataset from one or many other extractors
(or datasets), it can be done with Dataset.from_extractors().
If a dataset is created from multiple extractors with
Dataset.from_extractors(), the source datasets will be joined,
so their categories must match. If datasets have mismatching categories,
use the more complex IntersectMerge class from datumaro.components.operations,
which will merge all the labels and remap the shifted indices in annotations.
A Dataset can be loaded from an existing dataset on disk with
Dataset.import_from() (for arbitrary formats) and
Dataset.load() (for the Datumaro data format).
By default, Dataset works lazily, which means all the operations requiring
iteration over inputs will be deferred as much as possible. If you don’t want
such behavior, use the init_cache() method or wrap the code in
eager_mode (from datumaro.components.dataset), which will load all
the annotations into memory. The media won’t be loaded unless the data
is required, because it can quickly waste all the available memory.
You can check if the dataset is cached with the is_cache_initialized
attribute.
Once created, a dataset can be modified in batch mode with transforms or
directly with the put() and remove() methods. Dataset instances
record information about changes done, which can be obtained by get_patch().
The patch information is used automatically on saving and exporting to
reduce the amount of disk writes. Changes can be flushed with
flush_changes().
from datumaro.components.annotation import Bbox, Polygon
from datumaro.components.dataset import Dataset
from datumaro.components.extractor import DatasetItem
# create a dataset directly from items
dataset1 = Dataset.from_iterable([
DatasetItem(id='image1', annotations=[
Bbox(x=1, y=2, w=3, h=4, label=1),
Polygon([1,2,3,2,4,4], label=2),]),], categories=['cat','dog','person','truck'])
dataset2 = Dataset(categories=dataset1.categories())
dataset2.put(DatasetItem(id='image2', annotations=[
Label(label=3),
Bbox(x=2, y=0, w=3, h=1, label=2)]))# create a dataset from other datasets
dataset = Dataset.from_extractors(dataset1, dataset2)# keep only annotated images
dataset.select(lambda item:len(item.annotations)!=0)# change dataset labels
dataset.transform('remap_labels',{'cat':'dog',# rename cat to dog'truck':'car',# rename truck to car'person':'',# remove this label}, default='delete')# iterate over elementsfor item in dataset:print(item.id, item.annotations)# iterate over subsets as Datasetsfor subset_name, subset in dataset.subsets().items():for item in subset:print(item.id, item.annotations)
Dataset merging
There are 2 methods of merging datasets in Datumaro:
simple merging (“joining”)
complex merging
The simple merging (“joining”)
This approach finds the corresponding DatasetItems in inputs,
finds equal annotations and leaves only the unique set of annotations.
This approach requires all the inputs to have categories with the same
labels (or no labels) in the same order.
This algorithm is applied automatically in Dataset.from_extractors()
and when the build targets are merged in the ProjectTree.make_dataset().
The complex merging
If datasets have mismatching categories, they can’t be
merged by the simple approach, because it can lead to errors in the
resulting annotations. For complex cases Datumaro provides a more
sophisticated algorithm, which finds matching annotations by computing
distances between them. Labels and attributes are deduced by voting,
spatial annotations use the corresponding metrics like
Intersection-over-Union (IoU), OKS, PDJ and others.
The categories of the input datasets are compared, the matching ones
complement missing information in each other, the mismatching ones are
appended after next. Label indices in annotations are shifted to the
new values.
The complex algorithm is available in the IntersectMerge class
from datumaro.components.operations. It must be used explicitly.
This class also allows to check the inputs and the output dataset
for errors and problems.
Projects
Projects are intended for complex use of Datumaro. They provide means of
persistence, versioning, high-level operations for datasets and also
allow to extend Datumaro via plugins. A project provides
access to build trees and revisions, data sources, models, configuration,
plugins and cache. Projects can have multiple data sources, which are
joined on dataset creation. Project configuration is available
in project.config. To add a data source into a Project, use
the import_source() method. The build tree of the current working
directory can be converted to a Dataset with
project.working_tree.make_dataset().
The Environment class is responsible for accessing built-in and
project-specific plugins. For a Project object, there is an instance of
related Environment in project.env.
The framework provides functions to read and write datasets in specific formats.
It is supported by Extractors, Importers, and Converters.
Dataset reading is supported by Extractors and Importers:
An Extractor produces a list of DatasetItems corresponding to the
dataset. Annotations are available in the DatasetItem.annotations list.
The SourceExtractor class is designed for loading simple, single-subset
datasets. It should be used by default. The Extractor base class should
be used when SourceExtractor’s functionality is not enough.
An Importer detects dataset files and generates dataset loading parameters
for the corresponding Extractors. Importers are optional, they
only extend the Extractor functionality and make them more flexible and
simple. They are mostly used to locate dataset subsets, but they also can
do some data compatibility checks and have other required logic.
It is possible to add custom Extractors and Importers. To do this, you need
to put an Extractor and Importer implementations to a plugin directory.
Dataset writing is supported by Converters.
A Converter produces a dataset of a specific format from dataset items.
It is possible to add custom Converters. To do this, you need to put a
Converter implementation script to a plugin directory.
Dataset Conversions (“Transforms”)
A Transform is a function for altering a dataset and producing a new one.
It can update dataset items, annotations, classes, and other properties.
A list of available transforms for dataset conversions can be extended by
adding a Transform implementation script into a plugin directory.
Model launchers
A list of available launchers for model execution can be extended by
adding a Launcher implementation script into a plugin directory.
Plugins
Datumaro comes with a number of built-in formats and other tools,
but it also can be extended by plugins. Plugins are optional components,
which dependencies are not installed by default.
In Datumaro there are several types of plugins, which include:
extractor - produces dataset items from data source
importer - recognizes dataset type and creates project
converter - exports dataset to a specific format
transformation - modifies dataset items or other properties
launcher - executes models
A plugin is a regular Python module. It must be present in a plugin directory:
<project_dir>/.datumaro/plugins for project-specific plugins
<datumaro_dir>/plugins for global plugins
A plugin can be used either via the Environment class instance,
or by regular module importing:
from datumaro.components.project import Environment, Project
from datumaro.plugins.yolo_format.converter import YoloConverter
# Import a dataset
dataset = Environment().make_importer('voc')(src_dir).make_dataset()# Load an existing project, save the dataset in some project-specific format
project = Project.load('project/dir')
project.env.converters.get('custom_format').convert(dataset, save_dir=dst_dir)# Save the dataset in some built-in format
Environment().converters.get('yolo').convert(dataset, save_dir=dst_dir)
YoloConverter.convert(dataset, save_dir=dst_dir)
Writing a plugin
A plugin is a Python module with any name, which exports some symbols. Symbols,
starting with _ are not exported by default. To export a symbol,
inherit it from one of the special classes:
from datumaro.components.extractor import Importer, Extractor, Transform
from datumaro.components.launcher import Launcher
from datumaro.components.converter import Converter
The exports list of the module can be used to override default behaviour:
classMyComponent1:...classMyComponent2:...
exports =[MyComponent2]# exports only MyComponent2
There is also an additional class to modify plugin appearance in command line:
from datumaro.components.cli_plugin import CliPlugin
classMyPlugin(Converter, CliPlugin):"""
Optional documentation text, which will appear in command-line help
"""
NAME ='optional_custom_plugin_name'defbuild_cmdline_parser(self,**kwargs):
parser =super().build_cmdline_parser(**kwargs)# set up argparse.ArgumentParser instance# the parsed args are supposed to be used as invocation optionsreturn parser
from datumaro.components.extractor import Transform, CliPlugin
from.file1 import something, useful
classMyTransform(Transform, CliPlugin):
NAME ="custom_name"# could be generated automatically"""
Some description. The text will be displayed in the command line output.
"""@classmethoddefbuild_cmdline_parser(cls,**kwargs):
parser =super().build_cmdline_parser(**kwargs)
parser.add_argument('-q',help="Very useful parameter")return parser
def__init__(self, extractor, q):super().__init__(extractor)
self.q = q
deftransform_item(self, item):return item
my_plugin2.py contents:
from datumaro.components.extractor import Extractor
classMyFormat:...class_MyFormatConverter(Converter):...classMyFormatExtractor(Extractor):...
exports =[MyFormat]# explicit exports declaration# MyFormatExtractor and _MyFormatConverter won't be exported
Command-line
Basically, the interface is divided on contexts and single commands.
Contexts are semantically grouped commands, related to a single topic or target.
Single commands are handy shorter alternatives for the most used commands
and also special commands, which are hard to be put into any specific context.
Docker is an example of similar approach.
Model-View-ViewModel (MVVM) UI pattern is used.
5 - Formats
List of dataset formats supported by Datumaro
5.1 - ADE20k (v2017)
Format specification
The original ADE20K 2017 dataset is available
here.
The consistency set (for checking the annotation consistency)
is available here.
Supported annotation types:
Masks
Supported annotation attributes:
occluded (boolean): whether the object is occluded by another object
other arbitrary boolean attributes, which can be specified
in the annotation file <image_name>_atr.txt
Import ADE20K 2017 dataset
A Datumaro project with an ADE20k source can be created in the following way:
datum create
datum import --format ade20k2017 <path/to/dataset>
It is also possible to import the dataset using Python API:
from datumaro.components.dataset import Dataset
ade20k_dataset = Dataset.import_from('<path/to/dataset>','ade20k2017')
ADE20K dataset directory should have the following structure:
The mask images <image_name>_seg.png contain information about the object
class segmentation masks and also separate each class into instances.
The channels R and G encode the objects class masks.
The channel B encodes the instance object masks.
The mask images <image_name>_parts_N.png contain segmentation masks for parts
of objects, where N is a number indicating the level in the part hierarchy.
The annotation files <image_name>_atr.txt describe the content of each
image. Each line in the text file contains:
column 1: instance number,
column 2: part level (0 for objects),
column 3: occluded (1 for true),
column 4: original raw name (might provide a more detailed categorization),
column 5: class name (parsed using wordnet),
column 6: double-quoted list of attributes, separated by commas.
Each column is separated by a #. See example of dataset
here.
Datumaro can convert an ADE20K dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports segmentation masks.
There are several ways to convert an ADE20k 2017 dataset to other dataset
formats using CLI:
datum create
datum import -f ade20k2017 <path/to/dataset>
datum export -f coco -o <output/dir> -- --save-images
# or
datum convert -if ade20k2017 -i <path/to/dataset>\
-f coco -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','ade202017')
dataset.export('save_dir','coco')
Examples
Examples of using this format from the code can be found in
the format tests
5.2 - ADE20k (v2020)
Format specification
The original ADE20K 2020 dataset is available
here.
The consistency set (for checking the annotation consistency)
is available here.
Supported annotation types:
Masks
Supported annotation attributes:
occluded (boolean): whether the object is occluded by another object
other arbitrary boolean attributes, which can be specified
in the annotation file <image_name>.json
Import ADE20K dataset
A Datumaro project with an ADE20k source can be created in the following way:
datum create
datum import --format ade20k2020 <path/to/dataset>
It is also possible to import the dataset using Python API:
from datumaro.components.dataset import Dataset
ade20k_dataset = Dataset.import_from('<path/to/dataset>','ade20k2020')
ADE20K dataset directory should have the following structure:
The mask images <image_name>_seg.png contain information about the object
class segmentation masks and also separate each class into instances.
The channels R and G encode the objects class masks.
The channel B encodes the instance object masks.
The mask images <image_name>_parts_N.png contain segmentation masks for
parts of objects, where N is a number indicating the level in the part
hierarchy.
The <image_name> directory contains instance masks for each
object in the image, these masks represent one-channel images,
each pixel of which indicates an affinity to a specific object.
The annotation files <image_name>.json describe the content of each image.
See our tests asset
for example of this file,
or check ADE20K toolkit for it.
Datumaro can convert an ADE20K dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports segmentation masks.
There are several ways to convert an ADE20k dataset to other dataset
formats using CLI:
datum create
datum import -f ade20k2020 <path/to/dataset>
datum export -f coco -o ./save_dir -- --save-images
# or
datum convert -if ade20k2020 -i <path/to/dataset>\
-f coco -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','ade20k2020')
dataset.export('save_dir','voc')
Examples
Examples of using this format from the code can be found in
the format tests
Datumaro can convert an align CelebA dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports labels or landmarks.
There are several ways to convert an align CelebA dataset to other dataset
formats using CLI:
datum create
datum import -f align_celeba <path/to/dataset>
datum export -f imagenet_txt -o ./save_dir -- --save-images
# or
datum convert -if align_celeba -i <path/to/dataset>\
-f imagenet_txt -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','align_celeba')
dataset.export('save_dir','voc')
Examples
Examples of using this format from the code can be found in
the format tests
Datumaro can convert a CelebA dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports labels, bounding boxes or landmarks.
There are several ways to convert a CelebA dataset to other dataset
formats using CLI:
datum create
datum import -f celeba <path/to/dataset>
datum export -f imagenet_txt -o ./save_dir -- --save-images
# or
datum convert -if celeba -i <path/to/dataset>\
-f imagenet_txt -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','celeba')
dataset.export('save_dir','voc')
Examples
Examples of using this format from the code can be found in
the format tests
Datumaro supports Python version CIFAR-10/100.
The difference between CIFAR-10 and CIFAR-100 is how labels are stored
in the meta files (batches.meta or meta) and in the annotation files.
The 100 classes in the CIFAR-100 are grouped into 20 superclasses. Each image
comes with a “fine” label (the class to which it belongs) and a “coarse” label
(the superclass to which it belongs). In CIFAR-10 there are no superclasses.
CIFAR formats contain 32 x 32 images. As an extension, Datumaro supports
reading and writing of arbitrary-sized images.
CIFAR-10:
num_cases_per_batch: 1000
label_names: list of strings (['airplane', 'automobile', 'bird', ...])
num_vis: 3072
CIFAR-100:
fine_label_names: list of strings (['apple', 'aquarium_fish', ...])
coarse_label_names: list of strings (['aquatic_mammals', 'fish', ...])
Annotation files:
Common:
'batch_label': 'training batch 1 of <N>'
'data': numpy.ndarray of uint8, layout N x C x H x W
'filenames': list of strings
If images have non-default size (32x32) (Datumaro extension):
'image_sizes': list of (H, W) tuples
CIFAR-10:
'labels': list of strings
CIFAR-100:
'fine_labels': list of integers
'coarse_labels': list of integers
Datumaro can convert a CIFAR dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports the classification task (e.g. MNIST, ImageNet, PascalVOC, etc.)
There are several ways to convert a CIFAR dataset to other dataset
formats using CLI:
datum create
datum import -f cifar <path/to/cifar>
datum export -f imagenet -o <output/dir># or
datum convert -if cifar -i <path/to/dataset>\
-f imagenet -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','cifar')
dataset.export('save_dir','imagenet', save_images=True)
Export to CIFAR
There are several ways to convert a dataset to CIFAR format:
# export dataset into CIFAR format from existing project
datum export -p <path/to/project> -f cifar -o <output/dir>\
-- --save-images
# converting to CIFAR format from other format
datum convert -if imagenet -i <path/to/dataset>\
-f cifar -o <output/dir> -- --save-images
Extra options for exporting to CIFAR format:
--save-images allow to export dataset with saving images
(by default False)
--image-ext <IMAGE_EXT> allow to specify image extension
for exporting the dataset (by default .png)
--save-dataset-meta - allow to export dataset with saving dataset meta
file (by default False)
The format (CIFAR-10 or CIFAR-100) in which the dataset will be
exported depends on the presence of superclasses in the LabelCategories.
Examples
Datumaro supports filtering, transformation, merging etc. for all formats
and for the CIFAR format in particular. Follow the user manual
to get more information about these operations.
There are several examples of using Datumaro operations to solve
particular problems with CIFAR dataset:
Example 1. How to create a custom CIFAR-like dataset
from datumaro.components.annotation import Label
from datumaro.components.dataset import Dataset
from datumaro.components.extractor import DatasetItem
dataset = Dataset.from_iterable([
DatasetItem(id=0, image=np.ones((32,32,3)),
annotations=[Label(3)]),
DatasetItem(id=1, image=np.ones((32,32,3)),
annotations=[Label(8)])], categories=['airplane','automobile','bird','cat','deer','dog','frog','horse','ship','truck'])
dataset.export('./dataset',format='cifar')
Example 2. How to filter and convert a CIFAR dataset to ImageNet
Convert a CIFAR dataset to ImageNet format, keep only images with the
dog class present:
Datumaro can convert an MPII Human Pose Dataset into
any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports bounding boxes or points.
There are several ways to convert an MPII Human Pose Dataset
to other dataset formats using CLI:
datum create
datum import -f mpii <path/to/dataset>
datum export -f voc -o ./save_dir -- --save-images
# or
datum convert -if mpii -i <path/to/dataset>\
-f voc -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','mpii')
dataset.export('save_dir','voc')
Examples
Examples of using this format from the code can be found in
the format tests
5.7 - MPII Human Pose Dataset (JSON)
Format specification
The original MPII Human Pose Dataset is available
here.
Supported annotation types:
Bbox
Points
Supported attributes:
center (a list with two coordinates of the center point
of the object)
scale (float)
Import MPII Human Pose Dataset (JSON)
A Datumaro project with an MPII Human Pose Dataset (JSON) source can be
created in the following way:
datum create
datum import --format mpii_json <path/to/dataset>
It is also possible to import the dataset using Python API:
from datumaro.components.dataset import Dataset
mpii_dataset = Dataset.import_from('<path/to/dataset>','mpii_json')
MPII Human Pose Dataset (JSON) directory should have the following structure:
Datumaro can convert an MPII Human Pose Dataset (JSON) into
any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports bounding boxes or points.
There are several ways to convert an MPII Human Pose Dataset (JSON)
to other dataset formats using CLI:
datum create
datum import -f mpii_json <path/to/dataset>
datum export -f voc -o ./save_dir -- --save-images
# or
datum convert -if mpii_json -i <path/to/dataset>\
-f voc -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','mpii_json')
dataset.export('save_dir','voc')
Examples
Examples of using this format from the code can be found in
the format tests
RGB folder containing standard RGB images used for training.
GT/LABELS folder containing containing PNG files (one per image).
Annotations are given in three channels. The red channel contains
the class of that pixel. The green channel contains the class only
for those objects that are dynamic (cars, pedestrians, etc.),
otherwise it contains 0.
GT/COLOR folder containing png files (one per image).
Annotations are given using a color representation.
When importing a dataset, only GT/LABELS folder will be used.
If it is missing, GT/COLOR folder will be used.
The original dataset also contains depth information, but Datumaro
does not currently support it.
Datumaro can convert a SYNTHIA dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports segmentation masks.
There are several ways to convert a SYNTHIA dataset to other dataset
formats using CLI:
datum create
datum import -f synthia <path/to/dataset>
datum export -f voc -o <output/dir> -- --save-images
# or
datum convert -if synthia -i <path/to/dataset>\
-f voc -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','synthia')
dataset.export('save_dir','voc')
Examples
Examples of using this format from the code can be found in
the format tests
5.9 - VoTT CSV
Format specification
VoTT (Visual Object Tagging Tool) is
an open source annotation tool released by Microsoft.
VoTT CSV is the format used by VoTT
when the user exports a project and selects “CSV” as the export format.
Supported annotation types:
Bbox
Import VoTT dataset
A Datumaro project with a VoTT CSV source can be created in the following way:
datum create
datum import --format vott_csv <path/to/dataset>
It is also possible to import the dataset using Python API:
from datumaro.components.dataset import Dataset
vott_csv_dataset = Dataset.import_from('<path/to/dataset>','vott_csv')
VoTT CSV dataset directory should have the following structure:
dataset/
├── dataset_meta.json # a list of custom labels (optional)
├── img0001.jpg
├── img0002.jpg
├── img0003.jpg
├── img0004.jpg
├── ...
├── test-export.csv
├── train-export.csv
└── ...
Datumaro can convert a VoTT CSV dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports bounding boxes.
There are several ways to convert a VoTT CSV dataset to other dataset
formats using CLI:
datum create
datum import -f vott_csv <path/to/dataset>
datum export -f voc -o ./save_dir -- --save-images
# or
datum convert -if vott_csv -i <path/to/dataset>\
-f voc -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','vott_csv')
dataset.export('save_dir','voc')
Examples
Examples of using this format from the code can be found in
VoTT CSV tests.
5.10 - VoTT JSON
Format specification
VoTT (Visual Object Tagging Tool) is
an open source annotation tool released by Microsoft.
VoTT JSON is the format used by VoTT
when the user exports a project and selects “VoTT JSON” as the export format.
Supported annotation types:
Bbox
Import VoTT dataset
A Datumaro project with a VoTT JSON source can be created in the following way:
datum create
datum import --format vott_json <path/to/dataset>
It is also possible to import the dataset using Python API:
from datumaro.components.dataset import Dataset
vott_json_dataset = Dataset.import_from('<path/to/dataset>','vott_json')
VoTT JSON dataset directory should have the following structure:
dataset/
├── dataset_meta.json # a list of custom labels (optional)
├── img0001.jpg
├── img0002.jpg
├── img0003.jpg
├── img0004.jpg
├── ...
├── test-export.json
├── train-export.json
└── ...
Datumaro can convert a VoTT JSON dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports bounding boxes.
There are several ways to convert a VoTT JSON dataset to other dataset
formats using CLI:
datum create
datum import -f vott_json <path/to/dataset>
datum export -f voc -o ./save_dir -- --save-images
# or
datum convert -if vott_json -i <path/to/dataset>\
-f voc -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','vott_json')
dataset.export('save_dir','voc')
Examples
Examples of using this format from the code can be found in
VoTT JSON tests.
Cityscapes format specification is available here.
Supported annotation types:
Masks
Supported annotation attributes:
is_crowd (boolean). Specifies if the annotation label can
distinguish between different instances.
If False, the annotation id field encodes the instance id.
Import Cityscapes dataset
The Cityscapes dataset is available for free download.
A Datumaro project with a Cityscapes source can be created in the following way:
datum create
datum import --format cityscapes <path/to/dataset>
Cityscapes dataset directory should have the following structure:
*_leftImg8bit.png - left images in 8-bit LDR format
*_color.png - class labels encoded by its color
*_labelIds.png - class labels are encoded by its index
*_instanceIds.png - class and instance labels encoded by an instance ID.
The pixel values encode class and the individual instance: the integer part
of a division by 1000 of each ID provides class ID, the remainder
is the instance ID. If a certain annotation describes multiple instances,
then the pixels have the regular ID of that class
To add custom classes, you can use dataset_meta.json
and label_colors.txt.
If the dataset_meta.json is not represented in the dataset, then
label_colors.txt will be imported if possible.
In label_colors.txt you can define custom color map and non-cityscapes labels,
for example:
To make sure that the selected dataset has been added to the project, you can
run datum project info, which will display the project information.
Export to other formats
Datumaro can convert a Cityscapes dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to formats
that support the segmentation task (e.g. PascalVOC, CamVID, etc.)
There are several ways to convert a Cityscapes dataset to other dataset
formats using CLI:
datum create
datum import -f cityscapes <path/to/cityscapes>
datum export -f voc -o <output/dir># or
datum convert -if cityscapes -i <path/to/cityscapes>\
-f voc -o <output/dir> -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','cityscapes')
dataset.export('save_dir','voc', save_images=True)
Export to Cityscapes
There are several ways to convert a dataset to Cityscapes format:
# export dataset into Cityscapes format from existing project
datum export -p <path/to/project> -f cityscapes -o <output/dir>\
-- --save-images
# converting to Cityscapes format from other format
datum convert -if voc -i <path/to/dataset>\
-f cityscapes -o <output/dir> -- --save-images
Extra options for exporting to Cityscapes format:
--save-images allow to export dataset with saving images
(by default False)
--image-ext IMAGE_EXT allow to specify image extension
for exporting dataset (by default - keep original or use .png, if none)
--save-dataset-meta - allow to export dataset with saving dataset meta
file (by default False)
--label_map allow to define a custom colormap. Example:
# mycolormap.txt :# 0 0 255 sky# 255 0 0 person#...
datum export -f cityscapes -- --label-map mycolormap.txt
# or you can use original cityscapes colomap:
datum export -f cityscapes -- --label-map cityscapes
Examples
Datumaro supports filtering, transformation, merging etc. for all formats
and for the Cityscapes format in particular. Follow the
user manual
to get more information about these operations.
There are several examples of using Datumaro operations to solve
particular problems with a Cityscapes dataset:
Example 1. Load the original Cityscapes dataset and convert to Pascal VOC
datum create -o project
datum import -p project -f cityscapes ./Cityscapes/
datum stats -p project
datum export -p project -o dataset/ -f voc -- --save-images
Example 2. Create a custom Cityscapes-like dataset
import numpy as np
from datumaro.components.annotation import Mask
from datumaro.components.dataset import Dataset
from datumaro.components.extractor import DatasetItem
import datumaro.plugins.cityscapes_format as Cityscapes
label_map = OrderedDict()
label_map['background']=(0,0,0)
label_map['label_1']=(1,2,3)
label_map['label_2']=(3,2,1)
categories = Cityscapes.make_cityscapes_categories(label_map)
dataset = Dataset.from_iterable([
DatasetItem(id=1,
image=np.ones((1,5,3)),
annotations=[
Mask(image=np.array([[1,0,0,1,1]]), label=1),
Mask(image=np.array([[0,1,1,0,0]]), label=2,id=2,
attributes={'is_crowd':False}),]),], categories=categories)
dataset.export('./dataset',format='cityscapes')
Examples of using this format from the code can be found in
the format tests
The dataset has annotations for multiple tasks. Each task has its own format
in Datumaro, and there is also a combined coco format, which includes all
the available tasks. The sub-formats have the same options as the “main”
format and only limit the set of annotation files they work with. To work with
multiple formats, use the corresponding option of the coco format.
A Datumaro project with a COCO source can be created in the following way:
datum create
datum import --format coco <path/to/dataset>
It is possible to specify project name and project directory. Run
datum create --help for more information.
Extra options for adding a source in the COCO format:
--keep-original-category-ids: Add dummy label categories so that
category indexes in the imported data source correspond to the category IDs
in the original annotation file.
A COCO dataset directory should have the following structure:
Annotation files must have the names like <task_name>_<subset_name>.json.
The year is treated as a part of the subset name.
If the annotation file name does’t match this pattern, use one of the
task-specific formats instead of plain coco: coco_captions,
coco_image_info, coco_instances, coco_labels, coco_panoptic,
coco_person_keypoints, coco_stuff. In this case all items of the
dataset will be added to the default subset.
You can import a dataset for one or several tasks
instead of the whole dataset. This option also allows to import annotation
files with non-default names. For example:
datum create
datum import --format coco_stuff -r <relpath/to/stuff.json><path/to/dataset>
To make sure that the selected dataset has been added to the project, you can
run datum project info, which will display the project information.
Notes:
COCO categories can have any integer ids, however, Datumaro will count
annotation category id 0 as “not specified”. This does not contradict
the original annotations, because they have category indices starting from 1.
Export to other formats
Datumaro can convert COCO dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to formats
that support the specified task (e.g. for panoptic segmentation - VOC, CamVID)
There are several ways to convert a COCO dataset to other dataset formats
using CLI:
datum create
datum import -f coco <path/to/coco>
datum export -f voc -o <output/dir># or
datum convert -if coco -i <path/to/coco> -f voc -o <output/dir>
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','coco')
dataset.export('save_dir','voc', save_images=True)
Export to COCO
There are several ways to convert a dataset to COCO format:
# export dataset into COCO format from existing project
datum export -p <path/to/project> -f coco -o <output/dir>\
-- --save-images
# converting to COCO format from other format
datum convert -if voc -i <path/to/dataset>\
-f coco -o <output/dir> -- --save-images
Extra options for exporting to COCO format:
--save-images allow to export dataset with saving images
(by default False)
--image-ext IMAGE_EXT allow to specify image extension
for exporting dataset (by default - keep original or use .jpg, if none)
--save-dataset-meta - allow to export dataset with saving dataset meta
file (by default False)
--segmentation-mode MODE allow to specify save mode for instance
segmentation:
‘guess’: guess the mode for each instance
(using ‘is_crowd’ attribute as hint)
‘polygons’: save polygons (merge and convert masks, prefer polygons)
‘mask’: save masks (merge and convert polygons, prefer masks)
(by default guess)
--crop-covered allow to crop covered segments so that background objects
segmentation was more accurate (by default False)
--allow-attributes ALLOW_ATTRIBUTES allow export of attributes
(by default True). The parameter enables or disables writing
the custom annotation attributes to the “attributes” annotation
field. This field is an extension to the original COCO format
--reindex REINDEX allow to assign new indices to images and annotations,
useful to avoid merge conflicts (by default False).
This option allows to control if the images and
annotations must be given new indices. It can be useful, when
you want to preserve the original indices in the produced dataset.
Consider having this option enabled when converting from other formats
or merging datasets to avoid conflicts
--merge-images allow to save all images into a single directory
(by default False). The parameter controls the output directory for
images. When enabled, the dataset images are saved into a single
directory, otherwise they are saved in separate directories by subsets.
--tasks TASKS allow to specify tasks for export dataset,
by default Datumaro uses all tasks. Example:
datum create
datum import -f coco <path/to/dataset>
datum export -f coco -- --tasks instances,stuff
Examples
Datumaro supports filtering, transformation, merging etc. for all formats
and for the COCO format in particular. Follow the
user manual
to get more information about these operations.
There are several examples of using Datumaro operations to solve
particular problems with a COCO dataset:
Example 1. How to load an original panoptic COCO dataset and convert to Pascal VOC
datum create -o project
datum import -p project -f coco_panoptic ./COCO/annotations/panoptic_val2017.json
datum stats -p project
datum export -p project -f voc -- --save-images
Example 2. How to create custom COCO-like dataset
import numpy as np
from datumaro.components.annotation import Mask
from datumaro.components.dataset import Dataset
from datumaro.components.extractor import DatasetItem
dataset = Dataset.from_iterable([
DatasetItem(id='000000000001',
image=np.ones((1,5,3)),
subset='val',
attributes={'id':40},
annotations=[
Mask(image=np.array([[0,0,1,1,0]]), label=3,id=7, group=7, attributes={'is_crowd':False}),
Mask(image=np.array([[0,1,0,0,1]]), label=1,id=20, group=20, attributes={'is_crowd':True}),]),], categories=['a','b','c','d'])
dataset.export('./dataset',format='coco_panoptic')
Examples of using this format from the code can be found in
the format tests
5.13 - Image zip
Format specification
The image zip format allows to export/import unannotated datasets
with images to/from a zip archive. The format doesn’t support any
annotations or attributes.
Import Image zip dataset
There are several ways to import unannotated datasets to your Datumaro project:
From an existing archive:
datum create
datum import -f image_zip ./images.zip
From a directory with zip archives. Datumaro will import images from
all zip files in the directory:
datum create
datum import -f image_zip ./foo
The directory with zip archives must have the following structure:
Images in the archives must have a supported extension,
follow the user manual to see the supported
extensions.
Export to other formats
Datumaro can convert image zip dataset into any other format Datumaro supports.
For example:
datum create -o project
datum import -p project -f image_zip ./images.zip
datum export -p project -f coco -o ./new_dir -- --save-images
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','image_zip')
dataset.export('save_dir','coco', save_images=True)
Export an unannotated dataset to a zip archive
Example: exporting images from a VOC dataset to zip archives:
datum create -o project
datum import -p project -f voc ./VOC2012
datum export -p project -f image_zip -- --name voc_images.zip
Extra options for exporting to image_zip format:
--save-images allow to export dataset with saving images
(default: False)
--image-ext <IMAGE_EXT> allow to specify image extension
for exporting dataset (default: use original or .jpg, if none)
--name name of output zipfile (default: default.zip)
--compression allow to specify archive compression method.
Available methods:
ZIP_STORED, ZIP_DEFLATED, ZIP_BZIP2, ZIP_LZMA (default: ZIP_STORED).
Follow zip documentation
for more information.
Examples
Examples of using this format from the code can be found in
the format tests
5.14 - Velodyne Points / KITTI Raw 3D
Format specification
Velodyne Points / KITTI Raw 3D data format homepage is
available here.
Velodyne Points / KITTI Raw 3D data format specification
is available here.
A Datumaro project with a KITTI source can be created in the following way:
datum create
datum import --format kitti_raw <path/to/dataset>
To make sure that the selected dataset has been added to the project,
you can run datum project info, which will display the project and dataset
information.
Export to other formats
Datumaro can convert a KITTI Raw dataset into any other
format Datumaro supports.
Such conversion will only be successful if the output
format can represent the type of dataset you want to convert,
e.g. 3D point clouds can be saved in Supervisely Point Clouds format,
but not in COCO keypoints.
There are several ways to convert a KITTI Raw dataset to other dataset formats:
datum create
datum import -f kitti_raw <path/to/kitti_raw>
datum export -f sly_pointcloud -o <output/dir># or
datum convert -if kitti_raw -i <path/to/kitti_raw> -f sly_pointcloud
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','kitti_raw')
dataset.export('save_dir','sly_pointcloud', save_images=True)
Export to KITTI Raw
There are several ways to convert a dataset to KITTI Raw format:
# export dataset into KITTI Raw format from existing project
datum export -p <path/to/project> -f kitti_raw -o <output/dir>\
-- --save-images
# converting to KITTI Raw format from other format
datum convert -if sly_pointcloud -i <path/to/dataset>\
-f kitti_raw -o <output/dir> -- --save-images --reindex
Extra options for exporting to KITTI Raw format:
--save-images allow to export dataset with saving images. This will
include point clouds and related images (by default False)
--image-ext IMAGE_EXT allow to specify image extension
for exporting dataset (by default - keep original or use .png, if none)
--reindex assigns new indices to frames and tracks. Allows annotations
without track_id attribute (they will be exported as single-frame tracks).
--allow-attrs allows writing arbitrary annotation attributes. They will
be written in <annotations> section of <poses><item>
(disabled by default)
Examples
Example 1. Import dataset, compute statistics
datum create -o project
datum import -p project -f kitti_raw ../kitti_raw/
datum stats -p project
Example 2. Convert Supervisely Pointclouds to KITTI Raw
truncated (boolean) - indicates that the bounding box specified for
the object does not correspond to the full extent of the object
occluded (boolean) - indicates that a significant portion of the object
within the bounding box is occluded by another object
score (float) - indicates confidence in detection
Import KITTI dataset
The KITTI left color images for object detection are available here.
The KITTI object detection labels are available here.
The KITTI segmentation dataset is available here.
A Datumaro project with a KITTI source can be created in the following way:
datum create
datum import --format kitti <path/to/dataset>
It is possible to specify project name and project directory. Run
datum create --help for more information.
KITTI detection dataset directory should have the following structure:
└─ Dataset/
├── testing/
│ └── image_2/
│ ├── <name_1>.<img_ext>
│ ├── <name_2>.<img_ext>
│ └── ...
└── training/
├── image_2/ # left color camera images
│ ├── <name_1>.<img_ext>
│ ├── <name_2>.<img_ext>
│ └── ...
└─── label_2/ # left color camera label files
├── <name_1>.txt
├── <name_2>.txt
└── ...
KITTI segmentation dataset directory should have the following structure:
└─ Dataset/
├── dataset_meta.json # a list of non-format labels (optional)
├── label_colors.txt # optional, color map for non-original segmentation labels
├── testing/
│ └── image_2/
│ ├── <name_1>.<img_ext>
│ ├── <name_2>.<img_ext>
│ └── ...
└── training/
├── image_2/ # left color camera images
│ ├── <name_1>.<img_ext>
│ ├── <name_2>.<img_ext>
│ └── ...
├── label_2/ # left color camera label files
│ ├── <name_1>.txt
│ ├── <name_2>.txt
│ └── ...
├── instance/ # instance segmentation masks
│ ├── <name_1>.png
│ ├── <name_2>.png
│ └── ...
├── semantic/ # semantic segmentation masks (labels are encoded by its id)
│ ├── <name_1>.png
│ ├── <name_2>.png
│ └── ...
└── semantic_rgb/ # semantic segmentation masks (labels are encoded by its color)
├── <name_1>.png
├── <name_2>.png
└── ...
To add custom classes, you can use dataset_meta.json
and label_colors.txt.
If the dataset_meta.json is not represented in the dataset, then
label_colors.txt will be imported if possible.
You can import a dataset for specific tasks
of KITTI dataset instead of the whole dataset,
for example:
datum import --format kitti_detection <path/to/dataset>
To make sure that the selected dataset has been added to the project, you can
run datum project info, which will display the project information.
Export to other formats
Datumaro can convert a KITTI dataset into any other format Datumaro supports.
Such conversion will only be successful if the output
format can represent the type of dataset you want to convert,
e.g. segmentation annotations can be
saved in Cityscapes format, but not as COCO keypoints.
There are several ways to convert a KITTI dataset to other dataset formats:
datum create
datum import -f kitti <path/to/kitti>
datum export -f cityscapes -o <output/dir># or
datum convert -if kitti -i <path/to/kitti> -f cityscapes -o <output/dir>
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','kitti')
dataset.export('save_dir','cityscapes', save_images=True)
Export to KITTI
There are several ways to convert a dataset to KITTI format:
# export dataset into KITTI format from existing project
datum export -p <path/to/project> -f kitti -o <output/dir>\
-- --save-images
# converting to KITTI format from other format
datum convert -if cityscapes -i <path/to/dataset>\
-f kitti -o <output/dir> -- --save-images
Extra options for exporting to KITTI format:
--save-images allow to export dataset with saving images
(by default False)
--image-ext IMAGE_EXT allow to specify image extension
for exporting dataset (by default - keep original or use .png, if none)
--save-dataset-meta - allow to export dataset with saving dataset meta
file (by default False)
--apply-colormap APPLY_COLORMAP allow to use colormap for class masks
(in folder semantic_rgb, by default True)
--label_map allow to define a custom colormap. Example:
# mycolormap.txt :# 0 0 255 sky# 255 0 0 person#...
datum export -f kitti -- --label-map mycolormap.txt
# or you can use original kitti colomap:
datum export -f kitti -- --label-map kitti
--tasks TASKS allow to specify tasks for export dataset,
by default Datumaro uses all tasks. Example:
datum export -f kitti -- --tasks detection
--allow-attributes ALLOW_ATTRIBUTES allow export of attributes
(by default True).
Examples
Datumaro supports filtering, transformation, merging etc. for all formats
and for the KITTI format in particular. Follow the
user manual
to get more information about these operations.
There are several examples of using Datumaro operations to solve
particular problems with KITTI dataset:
Example 1. How to load an original KITTI dataset and convert to Cityscapes
datum create -o project
datum import -p project -f kitti ./KITTI/
datum stats -p project
datum export -p project -f cityscapes -- --save-images
Example 2. How to create a custom KITTI-like dataset
import numpy as np
from datumaro.components.annotation import Mask
from datumaro.components.dataset import Dataset
from datumaro.components.extractor import DatasetItem
import datumaro.plugins.kitti_format as KITTI
label_map ={}
label_map['background']=(0,0,0)
label_map['label_1']=(1,2,3)
label_map['label_2']=(3,2,1)
categories = KITTI.make_kitti_categories(label_map)
dataset = Dataset.from_iterable([
DatasetItem(id=1,
image=np.ones((1,5,3)),
annotations=[
Mask(image=np.array([[1,0,0,1,1]]), label=1,id=0,
attributes={'is_crowd':False}),
Mask(image=np.array([[0,1,1,0,0]]), label=2,id=0,
attributes={'is_crowd':False}),]),], categories=categories)
dataset.export('./dataset',format='kitti')
Examples of using this format from the code can be found in
the format tests
5.16 - LFW
Format specification
LFW (Labeled Faces in the Wild Home)
it’s dataset for face identification task,
specification for this format is available
here.
You can also download original LFW dataset
here.
Original dataset contains images with people faces.
For each image contains information about person’s name, as well as
information about images that matched with this person
and mismatched with this person.
Also LFW contains additional information about landmark points on the face.
Supported annotation types:
Label
Points (face landmark points)
Supported attributes:
negative_pairs: list with names of mismatched persons;
positive_pairs: list with names of matched persons;
Import LFW dataset
Importing LFW dataset into the Datumaro project:
datum create
datum import -f lfw <path_to_lfw_dataset>
See more information about adding datasets to the project in the
docs.
Also you can import LFW dataset from Python API:
from datumaro.components.dataset import Dataset
lfw_dataset = Dataset.import_from('<path_to_lfw_dataset>','lfw')
For successful importing the LFW dataset, the directory with it
should has the following structure:
<path_to_lfw_dataset>/
├── subset_1
│ ├── annotations
│ │ ├── landmarks.txt # list with landmark points for each image
│ │ ├── pairs.txt # list of matched and mismatched pairs of person
│ │ └── people.txt # optional file with a list of persons name
│ └── images
│ ├── name0
│ │ ├── name0_0001.jpg
│ │ ├── name0_0002.jpg
│ │ ├── ...
│ ├── name1
│ │ ├── name1_0001.jpg
│ │ ├── name1_0002.jpg
│ │ ├── ...
├── subset_2
│ ├── ...
├── ...
Full description of annotation *.txt files available
here.
Export LFW dataset
With Datumaro you can convert LFW dataset into any other
format Datumaro supports.
Pay attention that this format should also support Label and/or Points
annotation types.
There is few ways to convert LFW dataset into other format:
# Converting to ImageNet with `convert` command:
datum convert -if lfw -i ./lfw_dataset \
-f imagenet -o ./output_dir -- --save-images
# Converting to VggFace2 through the Datumaro project:
datum create
datum add -f lfw ./lfw_dataset
datum export -f vgg_face2 -o ./output_dir2
Note: some formats have extra export options. For particular format see the
docs to get information about it.
Export dataset to the LFW format
With Datumaro you can export dataset that has Label or/and Points
annotations to the LFW format, example:
# Converting VGG Face2 dataset into the LFW format
datum convert -if vgg_face2 -i ./voc_dataset \
-f lfw -o ./output_dir
# Export dataaset to the LFW format through the Datumaro project:
datum create
datum import -f voc_classification ../vgg_dataset
datum export -f lfw -o ./output_dir -- --save-images --image-ext png
Available extra export options for LFW dataset format:
--save-images allow to export dataset with saving images.
(by default False)
--image-ext IMAGE_EXT allow to specify image extension
for exporting dataset (by default - keep original)
5.17 - Mapillary Vistas
Format specification
Mapillary Vistas dataset homepage is available
here.
After registration the dataset will be available for downloading.
The specification for this format contains in the root directory of original
dataset.
Supported atttibutes:
- is_crowd(boolean; on panoptic mask): Indicates that the annotation
covers multiple instances of the same class.
Import Mapillary Vistas dataset
Use these instructions to import Mapillary Vistas dataset into Datumaro project:
datum create
datum add -f mapillary_vistas ./dataset
Note: the directory with dataset should be subdirectory of the
project directory.
Note: there is no opportunity to import both instance and panoptic
masks for one dataset.
Use one of subformats (mapillary_vistas_instances, mapillary_vistas_panoptic),
if your dataset contains both panoptic and instance masks:
datum add -f mapillary_vistas_instances ./dataset
# or
datum add -f mapillary_vistas_panoptic ./dataset
Extra options for adding a source in the Mapillary Vistas format:
--use-original-config: Use original config_*.json file for your version of
Mapillary Vistas dataset. This options can helps to import dataset, in case
when you don’t have config_*.json file, but your dataset is using original
categories of Mapillary Vistas dataset. The version of dataset will be detect
by the name of annotation directory in your dataset (v1.2 or v2.0).
--keep-original-category-ids: Add dummy label categories so that
category indexes in the imported data source correspond to the category IDs
in the original annotation file.
Example of using extra options:
datum add -f mapillary_vistas ./dataset -- --use-original-config
Mapillary Vistas dataset has two versions: v1.2, v2.0.
They differ in the number of classes, the name of the classes, supported types
of annotations, and the names of the directory with annotations.
So, the directory with dataset should have one of these structures:
Fashion MNIST format specification is available here.
MNIST in CSV format specification is available here.
The dataset has several data formats available. Datumaro supports the
binary (Python pickle) format and the CSV variant. Each data format is covered
by a separate Datumaro format.
Supported formats:
Binary (Python pickle) - mnist
CSV - mnist_csv
Supported annotation types:
Label
The format only supports single channel 28 x 28 images.
A Datumaro project with a MNIST source can be created in the following way:
datum create
datum import --format mnist <path/to/dataset>
datum import --format mnist_csv <path/to/dataset>
MNIST dataset directory should have the following structure:
└─ Dataset/
├── dataset_meta.json # a list of non-format labels (optional)
├── labels.txt # a list of non-digit labels in other format (optional)
├── t10k-images-idx3-ubyte.gz
├── t10k-labels-idx1-ubyte.gz
├── train-images-idx3-ubyte.gz
└── train-labels-idx1-ubyte.gz
MNIST in CSV dataset directory should have the following structure:
└─ Dataset/
├── dataset_meta.json # a list of non-format labels (optional)
├── labels.txt # a list of non-digit labels in other format (optional)
├── mnist_test.csv
└── mnist_train.csv
To add custom classes, you can use dataset_meta.json
and labels.txt.
If the dataset_meta.json is not represented in the dataset, then
labels.txt will be imported if possible.
For example, labels.txt for Fashion MNIST the following contents:
Datumaro can convert a MNIST dataset into any other format Datumaro supports.
To get the expected result, convert the dataset to formats
that support the classification task (e.g. CIFAR-10/100, ImageNet, PascalVOC,
etc.)
There are several ways to convert a MNIST dataset to other dataset formats:
datum create
datum import -f mnist <path/to/mnist>
datum export -f imagenet -o <output/dir># or
datum convert -if mnist -i <path/to/mnist> -f imagenet -o <output/dir>
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','mnist')
dataset.export('save_dir','imagenet', save_images=True)
These steps also will work for MNIST in CSV, if you use mnist_csv
instead of mnist.
Export to MNIST
There are several ways to convert a dataset to MNIST format:
# export dataset into MNIST format from existing project
datum export -p <path/to/project> -f mnist -o <output/dir>\
-- --save-images
# converting to MNIST format from other format
datum convert -if imagenet -i <path/to/dataset>\
-f mnist -o <output/dir> -- --save-images
Extra options for exporting to MNIST format:
--save-images allow to export dataset with saving images
(by default False)
--image-ext <IMAGE_EXT> allow to specify image extension
for exporting dataset (by default .png)
--save-dataset-meta - allow to export dataset with saving dataset meta
file (by default False)
These commands also work for MNIST in CSV if you use mnist_csv instead of mnist.
Examples
Datumaro supports filtering, transformation, merging etc. for all formats
and for the MNIST format in particular. Follow the user manual
to get more information about these operations.
There are several examples of using Datumaro operations to solve
particular problems with MNIST dataset:
Example 1. How to create a custom MNIST-like dataset
from datumaro.components.annotation import Label
from datumaro.components.dataset import Dataset
from datumaro.components.extractor import DatasetItem
dataset = Dataset.from_iterable([
DatasetItem(id=0, image=np.ones((28,28)),
annotations=[Label(2)]),
DatasetItem(id=1, image=np.ones((28,28)),
annotations=[Label(7)])], categories=[str(label)for label inrange(10)])
dataset.export('./dataset',format='mnist')
Example 2. How to filter and convert a MNIST dataset to ImageNet
Convert MNIST dataset to ImageNet format, keep only images with 3 class
presented:
ICDAR is a dataset for text recognition task, it’s available
for download here. There is exists
two most popular version of this dataset: ICDAR13 and ICDAR15,
Datumaro supports both of them.
Original dataset contains the following subformats:
ICDAR word recognition;
ICDAR text localization;
ICDAR text segmentation.
Supported types of annotations:
ICDAR word recognition
Caption
ICDAR text localization
Polygon, Bbox
ICDAR text segmentation
Mask
Supported attributes:
ICDAR text localization
text: transcription of text is inside a Polygon/Bbox.
ICDAR text segmentation
index: identifier of the annotation object, which is encoded in the mask
and coincides with the line number in which the description
of this object is written;
text: transcription of text is inside a Mask;
color: RGB values of the color corresponding text in the mask image
(three numbers separated by space);
center: coordinates of the center of text
(two numbers separated by space).
Import ICDAR dataset
There is few ways to import ICDAR dataset with Datumaro:
Through the Datumaro project
datum create
datum import -f icdar_text_localization <text_localization_dataset>
datum import -f icdar_text_segmentation <text_segmentation_dataset>
datum import -f icdar_word_recognition <word_recognition_dataset>
See more information about adding datasets to the project in the
docs.
Export to other formats
Datumaro can convert ICDAR dataset into any other format
Datumaro supports. Examples:
# converting ICDAR text segmentation dataset into the VOC with `convert` command
datum convert -if icdar_text_segmentation -i source_dataset \
-f voc -o export_dir -- --save-images
# converting ICDAR text localization into the LabelMe through Datumaro project
datum create
datum import -f icdar_text_localization source_dataset
datum export -f label_me -o ./export_dir -- --save-images
Note: some formats have extra export options. For particular format see the
docs to get information about it.
With Datumaro you can also convert your dataset to one of the ICDAR formats,
but to get expected result, the source dataset should contain required
attributes, described in previous section.
Note: in case with icdar_text_segmentation format, if your dataset contains
masks without attribute color then it will be generated automatically.
Available extra export options for ICDAR dataset formats:
--save-images allow to export dataset with saving images.
(by default False)
--image-ext IMAGE_EXT allow to specify image extension
for exporting dataset (by default - keep original)
5.20 - Open Images
Format specification
A description of the Open Images Dataset (OID) format is available
here.
Datumaro supports versions 4, 5 and 6.
Supported annotation types:
Label (human-verified image-level labels)
Bbox (bounding boxes)
Mask (segmentation masks)
Supported annotation attributes:
Labels
score (read/write, float).
The confidence level from 0 to 1.
A score of 0 indicates that
the image does not contain objects of the corresponding class.
Bounding boxes
score (read/write, float).
The confidence level from 0 to 1.
In the original dataset this is always equal to 1,
but custom datasets may be created with arbitrary values.
occluded (read/write, boolean).
Whether the object is occluded by another object.
truncated (read/write, boolean).
Whether the object extends beyond the boundary of the image.
is_group_of (read/write, boolean).
Whether the object represents a group of objects of the same class.
is_depiction (read/write, boolean).
Whether the object is a depiction (such as a drawing)
rather than a real object.
is_inside (read/write, boolean).
Whether the object is seen from the inside.
Masks
box_id (read/write, string).
An identifier for the bounding box associated with the mask.
predicted_iou (read/write, float).
Predicted IoU value with respect to the ground truth.
Import Open Images dataset
The Open Images dataset is available for free download.
All annotation files are optional,
except that if the mask metadata files for a given subset are downloaded,
all corresponding images must be downloaded as well, and vice versa.
A Datumaro project with an OID source can be created in the following way:
datum create
datum import --format open_images <path/to/dataset>
It is possible to specify project name and project directory. Run
datum create --help for more information.
Open Images dataset directory should have the following structure:
To load bounding box and segmentation mask annotations,
Datumaro needs to know the sizes of the corresponding images.
By default, it will determine these sizes by loading each image from disk,
which requires the images to be present and makes the loading process slow.
If you want to load the aforementioned annotations on a machine where
the images are not available,
or just to speed up the dataset loading process,
you can extract the image size information in advance
and record it in an image metadata file.
This file must be placed at annotations/images.meta,
and must contain one line per image, with the following structure:
<ID> <height> <width>
Where <ID> is the file name of the image without the extension,
and <height> and <width> are the dimensions of that image.
<ID> may be quoted with either single or double quotes.
The image metadata file, if present, will be used to determine the image
sizes without loading the images themselves.
Here’s one way to create the images.meta file using ImageMagick,
assuming that the images are present on the current machine:
# run this from the dataset directoryfind images -name '*.jpg' -exec \
identify -format '"%[basename]" %[height] %[width]\n'{} + \> annotations/images.meta
Export to other formats
Datumaro can convert OID into any other format Datumaro supports.
To get the expected result, convert the dataset to a format
that supports image-level labels.
There are several ways to convert OID to other dataset formats:
datum create
datum import -f open_images <path/to/open_images>
datum export -f cvat -o <output/dir># or
datum convert -if open_images -i <path/to/open_images> -f cvat -o <output/dir>
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','open_images')
dataset.export('save_dir','cvat', save_images=True)
Export to Open Images
There are several ways to convert an existing dataset to the Open Images format:
# export dataset into Open Images format from existing project
datum export -p <path/to/project> -f open_images -o <output/dir>\
-- --save_images
# convert a dataset in another format to the Open Images format
datum convert -if imagenet -i <path/to/dataset>\
-f open_images -o <output/dir>\
-- --save-images
Extra options for exporting to the Open Images format:
--save-images - save image files when exporting the dataset
(by default, False)
--image-ext IMAGE_EXT - save image files with the specified extension
when exporting the dataset (by default, uses the original extension
or .jpg if there isn’t one)
--save-dataset-meta - allow to export dataset with saving dataset meta
file (by default False)
Examples
Datumaro supports filtering, transformation, merging etc. for all formats
and for the Open Images format in particular. Follow the
user manual
to get more information about these operations.
Here are a few examples of using Datumaro operations to solve
particular problems with the Open Images dataset:
Example 1. Load the Open Images dataset and convert to the CVAT format
datum create -o project
datum import -p project -f open_images ./open-images-dataset/
datum stats -p project
datum export -p project -f cvat -- --save-images
Example 2. Create a custom OID-like dataset
import numpy as np
from datumaro.components.dataset import Dataset
from datumaro.components.annotation import(
AnnotationType, Label, LabelCategories,)from datumaro.components.extractor import DatasetItem
dataset = Dataset.from_iterable([
DatasetItem(id='0000000000000001',
image=np.ones((1,5,3)),
subset='validation',
annotations=[
Label(0, attributes={'score':1}),
Label(1, attributes={'score':0}),],),], categories=['/m/0','/m/1'])
dataset.export('./dataset',format='open_images')
Examples of using this format from the code can be found in
the format tests.
5.21 - ImageNet
Format specification
ImageNet is one of the most popular datasets for image classification task,
this dataset is available for downloading
here
Supported types of annotations:
Label
Format doesn’t support any attributes for annotations objects.
The original ImageNet dataset contains about 1.2M images and information
about class name for each image. Datumaro supports two versions of ImageNet
format: imagenet and imagenet_txt. The imagenet_txt format assumes storing
information about the class of the image in *.txt files. And imagenet format
assumes storing information about the class of the image in the name of
directory where is this image stored.
Import ImageNet dataset
A Datumaro project with a ImageNet dataset can be created
in the following way:
datum create
datum import -f imagenet <path_to_dataset>
# or
datum import -f imagenet_txt <path_to_dataset>
Note: if you use datum import then <path_to_dataset> should not be a
subdirectory of directory with Datumaro project, see more information about
it in the docs.
Load ImageNet dataset through the Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path_to_dataset>',format='imagenet_txt')
For successful importing of ImageNet dataset the input directory with dataset
should has the following structure:
imagenet_txt_dataset/
├── images # directory with images
│ ├── <image_name_0>.jpg
│ ├── <image_name_1>.jpg
│ ├── <image_name_2>.jpg
│ ├── ...
├── synsets.txt # optional, list of labels
└── train.txt # list of pairs (image_name, label)
Note: if you don’t have synsets file then Datumaro will automatically generate
classes with a name pattern class-<i>.
Datumaro has few import options for imagenet_txt format, to apply them
use the -- after the main command argument.
imagenet_txt import options:
--labels {file, generate}: allow to specify where to get label
descriptions from (use file to load from the file specified
by --labels-file; generate to create generic ones)
--labels-file allow to specify path to the file with label descriptions
(“synsets.txt”)
Export ImageNet dataset
Datumaro can convert ImageNet into any other format
Datumaro supports.
To get the expected result, convert the dataset to a format
that supports Label annotation objects.
# Using `convert` command
datum convert -if imagenet -i <path_to_imagenet> \
-f voc -o <output_dir> -- --save-images
# Using Datumaro project
datum create
datum import -f imagenet_txt <path_to_imagenet> -- --labels generate
datum export -f open_images -o <output_dir>
And also you can convert your ImageNet dataset using Python API
from datumaro.components.dataset import Dataset
imagenet_dataset = Dataset.import_from('<path_to_dataset',format='imagenet')
imagenet_dataset.export('<output_dir>',format='vgg_face2', save_images=True)
Note: some formats have extra export options. For particular format see the
docs to get information about it.
Export dataset to the ImageNet format
If your dataset contains Label for images and you want to convert this
dataset into the ImagetNet format, you can use Datumaro for it:
# Using convert command
datum convert -if open_images -i <path_to_oid> \
-f imagenet_txt -o <output_dir> -- --save-images --save-dataset-meta
# Using Datumaro project
datum create
datum import -f open_images <path_to_oid>
datum export -f imagenet -o <output_dir>
Extra options for exporting to ImageNet formats:
--save-images allow to export dataset with saving images
(by default False)
--image-ext <IMAGE_EXT> allow to specify image extension
for exporting the dataset (by default .png)
--save-dataset-meta - allow to export dataset with saving dataset meta
file (by default False)
5.22 - Pascal VOC
Format specification
Pascal VOC format specification is available
here.
The dataset has annotations for multiple tasks. Each task has its own format
in Datumaro, and there is also a combined voc format, which includes all
the available tasks. The sub-formats have the same options as the “main”
format and only limit the set of annotation files they work with. To work with
multiple formats, use the corresponding option of the voc format.
Supported tasks / formats:
The combined format - voc
Image classification - voc_classification
Object detection - voc_detection
Action classification - voc_action
Class and instance segmentation - voc_segmentation
Person layout detection - voc_layout
Supported annotation types:
Label (classification)
Bbox (detection, action detection and person layout)
Mask (segmentation)
Supported annotation attributes:
occluded (boolean) - indicates that a significant portion of the
object within the bounding box is occluded by another object
truncated (boolean) - indicates that the bounding box specified for
the object does not correspond to the full extent of the object
difficult (boolean) - indicates that the object is considered difficult
to recognize
action attributes (boolean) - jumping, reading and
others.
Indicate that the object does the corresponding action.
arbitrary attributes (string/number) - A Datumaro extension. Stored
in the attributes section of the annotation xml file. Available for
bbox annotations only.
Import Pascal VOC dataset
The Pascal VOC dataset is available for free download
here
A Datumaro project with a Pascal VOC source can be created in the following way:
datum create
datum import --format voc <path/to/dataset>
It is possible to specify project name and project directory. Run
datum create --help for more information.
Pascal VOC dataset directory should have the following structure:
└─ Dataset/
├── dataset_meta.json # a list of non-Pascal labels (optional)
├── labelmap.txt # or a list of non-Pascal labels in other format (optional)
│
├── Annotations/
│ ├── ann1.xml # Pascal VOC format annotation file
│ ├── ann2.xml
│ └── ...
├── JPEGImages/
│ ├── img1.jpg
│ ├── img2.jpg
│ └── ...
├── SegmentationClass/ # directory with semantic segmentation masks
│ ├── img1.png
│ ├── img2.png
│ └── ...
├── SegmentationObject/ # directory with instance segmentation masks
│ ├── img1.png
│ ├── img2.png
│ └── ...
│
└── ImageSets/
├── Main/ # directory with list of images for detection and classification task
│ ├── test.txt # list of image names in test subset (without extension)
| ├── train.txt # list of image names in train subset (without extension)
| └── ...
├── Layout/ # directory with list of images for person layout task
│ ├── test.txt
| ├── train.txt
| └── ...
├── Action/ # directory with list of images for action classification task
│ ├── test.txt
| ├── train.txt
| └── ...
└── Segmentation/ # directory with list of images for segmentation task
├── test.txt
├── train.txt
└── ...
The ImageSets directory should contain at least one of the directories:
Main, Layout, Action, Segmentation.
These directories contain .txt files with a list of images in a subset,
the subset name is the same as the .txt file name. Subset names can be
arbitrary.
To add custom classes, you can use dataset_meta.json
and labelmap.txt.
If the dataset_meta.json is not represented in the dataset, then
labelmap.txt will be imported if possible.
In labelmap.txt you can define custom color map and non-pascal labels,
for example:
It is also possible to import grayscale (1-channel) PNG masks.
For grayscale masks provide a list of labels with the number of lines
equal to the maximum color index on images. The lines must be in the
right order so that line index is equal to the color index. Lines can
have arbitrary, but different, colors. If there are gaps in the used
color indices in the annotations, they must be filled with arbitrary
dummy labels. Example:
car:0,128,0:: # color index 0
aeroplane:10,10,128:: # color index 1
_dummy2:2,2,2:: # filler for color index 2
_dummy3:3,3,3:: # filler for color index 3
boat:108,0,100:: # color index 3
...
_dummy198:198,198,198:: # filler for color index 198
_dummy199:199,199,199:: # filler for color index 199
the_last_label:12,28,0:: # color index 200
You can import dataset for specific tasks
of Pascal VOC dataset instead of the whole dataset,
for example:
datum import -f voc_detection -r ImageSets/Main/train.txt <path/to/dataset>
To make sure that the selected dataset has been added to the project, you
can run datum project info, which will display the project information.
Export to other formats
Datumaro can convert a Pascal VOC dataset into any other format
Datumaro supports.
Such conversion will only be successful if the output
format can represent the type of dataset you want to convert,
e.g. image classification annotations can be
saved in ImageNet format, but not as COCO keypoints.
There are several ways to convert a Pascal VOC dataset to other dataset formats:
datum create
datum import -f voc <path/to/voc>
datum export -f coco -o <output/dir># or
datum convert -if voc -i <path/to/voc> -f coco -o <output/dir>
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','voc')
dataset.export('save_dir','coco', save_images=True)
Export to Pascal VOC
There are several ways to convert an existing dataset to Pascal VOC format:
# export dataset into Pascal VOC format (classification) from existing project
datum export -p <path/to/project> -f voc -o <output/dir> -- --tasks classification
# converting to Pascal VOC format from other format
datum convert -if imagenet -i <path/to/dataset>\
-f voc -o <output/dir>\
-- --label_map voc --save-images
Extra options for exporting to Pascal VOC format:
--save-images - allow to export dataset with saving images
(by default False)
--image-ext IMAGE_EXT - allow to specify image extension
for exporting dataset (by default use original or .jpg if none)
--save-dataset-meta - allow to export dataset with saving dataset meta
file (by default False)
--apply-colormap APPLY_COLORMAP - allow to use colormap for class
and instance masks (by default True)
--allow-attributes ALLOW_ATTRIBUTES - allow export of attributes
(by default True)
--keep-empty KEEP_EMPTY - write subset lists even if they are empty
(by default False)
--tasks TASKS - allow to specify tasks for export dataset,
by default Datumaro uses all tasks. Example:
datum export -f voc -- --tasks detection,classification
--label_map PATH - allows to define a custom colormap. Example:
# mycolormap.txt [label : color_rgb : parts : actions]:# cat:0,0,255::# person:255,0,0:head:
datum export -f voc_segmentation -- --label-map mycolormap.txt
# or you can use original voc colomap:
datum export -f voc_segmentation -- --label-map voc
Examples
Datumaro supports filtering, transformation, merging etc. for all formats
and for the Pascal VOC format in particular. Follow
user manual
to get more information about these operations.
There are few examples of using Datumaro operations to solve
particular problems with Pascal VOC dataset:
Example 1. How to prepare an original dataset for training.
In this example, preparing the original dataset to train the semantic
segmentation model includes:
loading,
checking duplicate images,
setting the number of images,
splitting into subsets,
export the result to Pascal VOC format.
datum create -o project
datum import -p project -f voc_segmentation ./VOC2012/ImageSets/Segmentation/trainval.txt
datum stats -p project # check statisctics.json -> repeated images
datum transform -p project -t ndr -- -w trainval -k 2500
datum filter -p project -e '/item[subset="trainval"]'
datum transform -p project -t random_split -- -s train:.8 -s val:.2
datum export -p project -f voc -- --label-map voc --save-images
Load Pascal VOC dataset, and export train subset with items
which has jumping attribute:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('./VOC2012',format='voc')
train_dataset = dataset.get_subset('train').as_dataset()defonly_jumping(item):for ann in item.annotations:if ann.attributes.get('jumping'):returnTruereturnFalse
train_dataset.select(only_jumping)
train_dataset.export('./jumping_label_me',format='label_me', save_images=True)
Example 4. Get information about items in Pascal VOC 2012 dataset for segmentation task:
from datumaro.components.annotation import AnnotationType
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('./VOC2012',format='voc')defhas_mask(item):for ann in item.annotations:if ann.type== AnnotationType.mask:returnTruereturnFalse
dataset.select(has_mask)print("Pascal VOC 2012 has %s images for segmentation task:"%len(dataset))for subset_name, subset in dataset.subsets().items():for item in subset:print(item.id, subset_name, end=";")
After executing this code, we can see that there are 5826 images
in Pascal VOC 2012 has for segmentation task and this result is the same as the
official documentation
Examples of using this format from the code can be found in
tests
5.23 - Supervisely Point Cloud
Format specification
Specification for the Point Cloud data format is available
here.
You can also find examples of working with the dataset
here.
Supported annotation types:
cuboid_3d
Supported annotation attributes:
track_id (read/write, integer), responsible for object field
createdAt (write, string),
updatedAt (write, string),
labelerLogin (write, string), responsible for the corresponding fields
in the annotation file.
arbitrary attributes
Supported image attributes:
description (read/write, string),
createdAt (write, string),
updatedAt (write, string),
labelerLogin (write, string), responsible for the corresponding fields
in the annotation file.
frame (read/write, integer). Indicates frame number of the image.
arbitrary attributes
Import Supervisely Point Cloud dataset
An example dataset in Supervisely Point Cloud format is available for download:
There are two ways to import a Supervisely Point Cloud dataset:
datum create
datum import --format sly_pointcloud --input-path <path/to/dataset># or
datum create
datum import -f sly_pointcloud <path/to/dataset>
To make sure that the selected dataset has been added to the project,
you can run datum project info, which will display the project and dataset
information.
Export to other formats
Datumaro can convert Supervisely Point Cloud dataset into any other
format Datumaro supports.
Such conversion will only be successful if the output
format can represent the type of dataset you want to convert,
e.g. 3D point clouds can be saved in KITTI Raw format,
but not in COCO keypoints.
There are several ways to convert a Supervisely Point Cloud dataset
to other dataset formats:
datum create
datum import -f sly_pointcloud <path/to/sly_pcd/>
datum export -f kitti_raw -o <output/dir># or
datum convert -if sly_pointcloud -i <path/to/sly_pcd/> -f kitti_raw
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','sly_pointcloud')
dataset.export('save_dir','kitti_raw', save_images=True)
Export to Supervisely Point Cloud
There are several ways to convert a dataset to Supervisely Point Cloud format:
# export dataset into Supervisely Point Cloud format from existing project
datum export -p <path/to/project> -f sly_pointcloud -o <output/dir>\
-- --save-images
# converting to Supervisely Point Cloud format from other format
datum convert -if kitti_raw -i <path/to/dataset>\
-f sly_pointcloud -o <output/dir> -- --save-images
Extra options for exporting in Supervisely Point Cloud format:
--save-images allow to export dataset with saving images. This will
include point clouds and related images (by default False)
--image-ext IMAGE_EXT allow to specify image extension
for exporting dataset (by default - keep original or use .png, if none)
--reindex assigns new indices to frames and annotations.
--allow-undeclared-attrs allows writing arbitrary annotation attributes.
By default, only attributes specified in the input dataset metainfo
will be written.
Examples
Example 1. Import dataset, compute statistics
datum create -o project
datum import -p project -f sly_pointcloud ../sly_dataset/
datum stats -p project
Example 2. Convert Supervisely Point Clouds to KITTI Raw
Examples of using this format from the code can be found in
the format tests
5.24 - YOLO
Format specification
The YOLO dataset format is for training and validating object detection
models. Specification for this format is available
here.
You can also find official examples of working with YOLO dataset here.
Supported annotation types:
Bounding boxes
YOLO format doesn’t support attributes for annotations.
The format only supports subsets named train or valid.
Import YOLO dataset
A Datumaro project with a YOLO source can be created in the following way:
datum create
datum import --format yolo <path/to/dataset>
YOLO dataset directory should have the following structure:
└─ yolo_dataset/
│
├── dataset_meta.json # a list of non-format labels (optional)
├── obj.names # file with list of classes
├── obj.data # file with dataset information
├── train.txt # list of image paths in train subset
├── valid.txt # list of image paths in valid subset
│
├── obj_train_data/ # directory with annotations and images for train subset
│ ├── image1.txt # list of labeled bounding boxes for image1
│ ├── image1.jpg
│ ├── image2.txt
│ ├── image2.jpg
│ └── ...
│
└── obj_valid_data/ # directory with annotations and images for valid subset
├── image101.txt
├── image101.jpg
├── image102.txt
├── image102.jpg
└── ...
YOLO dataset cannot contain a subset with a name other than train or valid.
If an imported dataset contains such subsets, they will be ignored.
If you are exporting a project into YOLO format,
all subsets different from train and valid will be skipped.
If there is no subset separation in a project, the data
will be saved in train subset.
obj.data should have the following content, it is not necessary to have both
subsets, but necessary to have one of them:
Here x_center, y_center, width, and height are relative to the image’s
width and height. The x_center and y_center are center of rectangle
(are not top-left corner).
Datumaro can convert YOLO dataset into any other format
Datumaro supports.
For successful conversion the output format should support
object detection task (e.g. Pascal VOC, COCO, TF Detection API etc.)
There are several ways to convert a YOLO dataset to other dataset formats:
datum create
datum add -f yolo <path/to/yolo/>
datum export -f voc -o <output/dir># or
datum convert -if yolo -i <path/to/dataset>\
-f coco_instances -o <path/to/dataset>
Or, using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path/to/dataset>','yolo')
dataset.export('save_dir','coco_instances', save_images=True)
Export to YOLO format
Datumaro can convert an existing dataset to YOLO format,
if the dataset supports object detection task.
Example:
datum create
datum import -f coco_instances <path/to/dataset>
datum export -f yolo -o <path/to/dataset> -- --save-images
Extra options for exporting to YOLO format:
--save-images allow to export dataset with saving images
(default: False)
--image-ext <IMAGE_EXT> allow to specify image extension
for exporting dataset (default: use original or .jpg, if none)
Examples
Example 1. Prepare PASCAL VOC dataset for exporting to YOLO format dataset
datum create -o project
datum import -p project -f voc ./VOC2012
datum filter -p project -e '/item[subset="train" or subset="val"]'
datum transform -p project -t map_subsets -- -s train:train -s val:valid
datum export -p project -f yolo -- --save-images
Example 2. Remove a class from YOLO dataset
Delete all items, which contain cat objects and remove
cat from list of classes:
datum create -o project
datum import -p project -f yolo ./yolo_dataset
datum filter -p project -m i+a -e '/item/annotation[label!="cat"]'
datum transform -p project -t remap_labels -- -l cat:
datum export -p project -f yolo -o ./yolo_without_cats
Example 3. Create a custom dataset in YOLO format
import numpy as np
from datumaro.components.annotation import Bbox
from datumaro.components.dataset import Dataset
from datumaro.components.extractor import DatasetItem
dataset = Dataset.from_iterable([
DatasetItem(id='image_001', subset='train',
image=np.ones((20,20,3)),
annotations=[
Bbox(3.0,1.0,8.0,5.0, label=1),
Bbox(1.0,1.0,10.0,1.0, label=2)]),
DatasetItem(id='image_002', subset='train',
image=np.ones((15,10,3)),
annotations=[
Bbox(4.0,4.0,4.0,4.0, label=3)])], categories=['house','bridge','crosswalk','traffic_light'])
dataset.export('../yolo_dataset',format='yolo', save_images=True)
Example 4. Get information about objects on each image
If you only want information about label names for each
image, then you can get it from code:
from datumaro.components.annotation import AnnotationType
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('./yolo_dataset',format='yolo')
cats = dataset.categories()[AnnotationType.label]for item in dataset:for ann in item.annotations:print(item.id, cats[ann.label].name)
And If you want complete information about each item you can run:
datum create -o project
datum import -p project -f yolo ./yolo_dataset
datum filter -p project --dry-run -e '/item'
5.25 - Market-1501
Format specification
Market-1501 is a dataset for person re-identification task, link
for downloading this dataset is available
here.
Supported items attributes:
person_id (str): four-digit number that represent ID of pedestrian;
camera_id (int): one-digit number that represent ID of camera that took
the image (original dataset has totally 6 cameras);
track_id (int): one-digit number that represent ID of the track with
the particular pedestrian, this attribute matches with sequence_id
in the original dataset;
frame_id (int): six-digit number, that mean number of
frame within this track. For the tracks, their names are accumulated
for each ID, but for frames, they start from “0001” in each track;
bbox_id (int): two-digit number, that mean number of
bounding bbox that was selected for that image
(see the
original docs
for more info).
These item attributes decodes into the image name with such convention:
0000_c1s1_000000_00.jpg
first four digits indicate the person_id;
digit after c indicates the camera_id;
digit after s indicate the track_id;
six digits after s1_ indicate the frame_id;
the last two digits before .jpg indicate the bbox_id.
Import Market-1501 dataset
Importing of Market-1501 dataset into the Datumaro project:
datum create
datum import -f market1501 <path_to_market1501>
See more information about adding datasets to the project in the
docs.
Or you can import Market-1501 using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path_to_dataset>','market1501')
For successful importing the Market-1501 dataset, the directory with it
should has the following structure:
With Datumaro you can export dataset, that has person_id item attribute,
to the Market-1501 format, example:
# Converting MARS dataset into the Market-1501
datum convert -if mars -i ./mars_dataset \
-f market1501 -o ./output_dir
# Export dataaset to the Market-1501 format through the Datumaro project:
datum create
datum add -f mars ../mars
datum export -f market1501 -o ./output_dir -- --save-images --image-ext png
Note: if your dataset contains only person_id attributes Datumaro
will assign default values for other attributes (camera_id, track_id, bbox_id)
and increment frame_id for collisions.
Available extra export options for Market-1501 dataset format:
--save-images allow to export dataset with saving images.
(by default False)
--image-ext IMAGE_EXT allow to specify image extension
for exporting dataset (by default - keep original)
5.26 - MARS
Format specification
MARS is a dataset for the motion analysis and person identification task.
MARS dataset is available for downloading
here
Supported types of annotations:
Bbox
Required attributes:
person_id (str): four-digit number that represent ID of pedestrian;
camera_id (int): one-digit number that represent ID of camera that took
the image (original dataset has totally 6 cameras);
track_id (int): four-digit number that represent ID of the track with
the particular pedestrian;
frame_id (int): three-digit number, that mean number of
frame within this track. For the tracks, their names are accumulated
for each ID, but for frames, they start from “0001” in each track.
Import MARS dataset
Use these instructions to import MARS dataset into Datumaro project:
datum create
datum add -f mars ./dataset
Note: the directory with dataset should be subdirectory of the
project directory.
mars_dataset
├── <bbox_subset_name1>
│ ├── 0001 # directory with images of pedestrian with id 0001
│ │ ├── 0001C1T0001F001.jpg
│ │ ├── 0001C1T0001F002.jpg
│ │ ├── ...
│ ├── 0002 # directory with images of pedestrian with id 0002
│ │ ├── 0002C1T0001F001.jpg
│ │ ├── 0002C1T0001F001.jpg
│ │ ├── ...
│ ├── 0000 # distractors images, which negatively affect retrieval accuracy.
│ │ ├── 0000C1T0001F001.jpg
│ │ ├── 0000C1T0001F001.jpg
│ │ ├── ...
│ ├── 00-1 # junk images which do not affect retrieval accuracy
│ │ ├── 00-1C1T0001F001.jpg
│ │ ├── 00-1C1T0001F001.jpg
│ │ ├── ...
├── <bbox_subset_name2>
│ ├── ...
├── ...
All images in MARS dataset follow a strict convention of naming:
xxxxCxTxxxxFxxx.jpg
the first four digits indicate the pedestrian’s number;
digit after C indicates the camera id;
four digits after T indicate the track id for this pedestrian;
three digits after F indicate the frame id with this track.
Note: there are two specific pedestrian IDs 0000 and 00-1
which indicate distracting images and unwanted images respectively.
5.27 - WIDER Face
Format specification
WIDER Face dataset is a face detection benchmark dataset,
that available for download here.
Supported types of annotation:
Bbox
Label
Supported attributes for bboxes:
blur:
0 face without blur;
1 face with normal blur;
2 face with heavy blur.
expression:
0 face with typical expression;
1 face with exaggerate expression.
illumination:
0 image contains normal illumination;
1 image contains extreme illumination.
pose:
0 pose is typical;
1 pose is atypical.
invalid:
0 image is valid;
1 image is invalid.
occluded:
0 face without occlusion;
1 face with partial occlusion;
2 face with heavy occlusion.
Import WIDER Face dataset
Importing of WIDER Face dataset into the Datumaro project:
datum create
datum import -f wider_face <path_to_wider_face>
Directory with WIDER Face dataset should has the following structure:
<path_to_wider_face>
├── labels.txt # optional file with list of classes
├── wider_face_split # directory with description of bboxes for each image
│ ├── wider_face_subset1_bbx_gt.txt
│ ├── wider_face_subset2_bbx_gt.txt
│ ├── ...
├── WIDER_subset1 # instead of 'subset1' you can use any other subset name
│ └── images
│ ├── 0--label_0 # instead of 'label_<n>' you can use any other class name
│ │ ├── 0_label_0_image_01.jpg
│ │ ├── 0_label_0_image_02.jpg
│ │ ├── ...
│ ├── 1--label_1
│ │ ├── 1_label_1_image_01.jpg
│ │ ├── 1_label_1_image_02.jpg
│ │ ├── ...
│ ├── ...
├── WIDER_subset2
│ └── images
│ ├── ...
├── ...
Check README
file of the original WIDER Face dataset to get more information
about structure of .txt annotation files.
Also example of WIDER Face dataset available in our
test assets.
Export WIDER Face dataset
With Datumaro you can convert WIDER Face dataset into any other
format Datumaro supports.
Pay attention that this format should also support Label and/or Bbox
annotation types.
Few ways to export WIDER Face dataset using CLI:
# Using `convert` command
datum convert -if wider_face -i <path_to_wider_face> \
-f voc -o <output_dir> -- --save-images
# Through the Datumaro project
datum create
datum import -f wider_face <path_to_wider_face>
datum export -f voc -o <output_dir> -- -save-images
Export WIDER Face dataset using Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path_to_wider_face','wider_face')# Here you can perform some transformation using dataset.transform or# dataset.filter
dataset.export('output_dir','open_images', save_images=True)
Note: some formats have extra export options. For particular format see the
docs to get information about it.
Export to WIDER Face dataset
Using Datumaro you can convert your dataset into the WIDER Face format,
but for succseful exporting your dataset should contain Label and/or Bbox.
Here example of exporting VOC dataset (object detection task)
into the WIDER Face format:
datum create
datum import -f voc_detection <path_to_voc>
datum export -f wider_face -o <output_dir> -- --save-images --image-ext='.png'
Available extra export options for WIDER Face dataset format:
--save-images allow to export dataset with saving images.
(by default False)
--image-ext IMAGE_EXT allow to specify image extension
for exporting dataset (by default - keep original)
5.28 - Vgg Face2 CSV
Format specification
Vgg Face 2 is a dataset for face-recognition task,
the repository with some information and sample data of Vgg Face 2 is available
here
Supported types of annotations:
Bbox
Points
Label
Format doesn’t support any attributes for annotations objects.
Import Vgg Face2 dataset
A Datumaro project with a Vgg Face 2 dataset can be created
in the following way:
datum create
datum import -f vgg_face2 <path_to_dataset>
Note: if you use datum import then <path_to_dataset> should not be a
subdirectory of directory with Datumaro project, see more information about
it in the docs.
And you can also load Vgg Face 2 through the Python API:
from datumaro.components.dataset import Dataset
dataset = Dataset.import_from('<path_to_dataset>',format='vgg_face2')
For successful importing of Vgg Face2 face the input directory with dataset
should has the following structure:
vgg_face2_dataset/
├── labels.txt # labels mapping
├── bb_landmark
│ ├── loose_bb_test.csv # information about bounding boxes for test subset
│ ├── loose_bb_train.csv
│ ├── loose_bb_<any_other_subset_name>.csv
│ ├── loose_landmark_test.csv # landmark points information for test subset
│ ├── loose_landmark_train.csv
│ └── loose_landmark_<any_other_subset_name>.csv
├── test
│ ├── n000001 # directory with images for n000001 label
│ │ ├── 0001_01.jpg
│ │ ├── 0001_02.jpg
│ │ ├── ...
│ ├── n000002 # directory with images for n000002 label
│ │ ├── 0002_01.jpg
│ │ ├── 0003_01.jpg
│ │ ├── ...
│ ├── ...
├── train
│ ├── n000004
│ │ ├── 0004_01.jpg
│ │ ├── 0004_02.jpg
│ │ ├── ...
│ ├── ...
└── <any_other_subset_name>
├── ...
Export Vgg Face2 dataset
Datumaro can convert a Vgg Face2 dataset into any other format
Datumaro supports.
There is few examples how to do it:
# Using `convert` command
datum convert -if vgg_face2 -i <path_to_vgg_face2> \
-f voc -o <output_dir> -- --save-images
# Using Datumaro project
datum create
datum import -f vgg_face2 <path_to_vgg_face2>
datum export -f yolo -o <output_dir>
Note: to get the expected result from the conversion, the output format
should support the same types of annotations (one or more) as Vgg Face2
(Bbox, Points, Label)
And also you can convert your Vgg Face2 dataset using Python API
from datumaro.components.dataset import Dataset
vgg_face2_dataset = Dataset.import_from('<path_to_dataset',format='vgg_face2')
vgg_face2_dataset.export('<output_dir>',format='open_images', save_images=True)
Note: some formats have extra export options. For particular format see the
docs to get information about it.
Export dataset to the Vgg Face2 format
If you have dataset in some format and want to convert this dataset
into the Vgg Face2, ensure that this dataset contains Bbox or/and Points
or/and Label and use Datumaro to perform conversion.
There is few examples:
# Using convert command
datum convert -if wider_face -i <path_to_wider> \
-f vgg_face2 -o <output_dir>
# Using Datumaro project
datum create
datum import -f wider_face <path_to_wider>
datum export -f vgg_face2 -o <output_dir> -- --save-images --image-ext '.png'
Note: vgg_face2 format supports only one Bbox per image
Extra options for exporting to Vgg Face2 format:
--save-images allow to export dataset with saving images
(by default False)
--image-ext <IMAGE_EXT> allow to specify image extension
for exporting the dataset (by default .png)
--save-dataset-meta - allow to export dataset with saving dataset meta
file (by default False)
6 - Plugins
6.1 - OpenVINO™ Inference Interpreter
Interpreter samples to parse OpenVINO™ inference outputs. This section on GitHub
Models supported from interpreter samples
There are detection and image classification examples.
You can find more OpenVINO™ Trained Models
here
To run the inference with OpenVINO™, the model format should be Intermediate
Representation(IR).
For the Caffe/TensorFlow/MXNet/Kaldi/ONNX models, please see the
Model Conversion Instruction
You need to implement your own interpreter samples to support the other
OpenVINO™ Trained Models.
To run the inference with OpenVINO™ models and the interpreter samples,
please follow the instructions below.
source<openvino_dir>/bin/setupvars.sh
datum create -o <proj_dir>
datum model add -l <launcher> -p <proj_dir> --copy -- \
-d <path/to/xml> -w <path/to/bin> -i <path/to/interpreter/script>
datum import -p <proj_dir> -f <format><path_to_dataset>
datum model run -p <proj_dir> -m model-0
Detection: ssd_mobilenet_v2_coco
source /opt/intel/openvino/bin/setupvars.sh
cd datumaro/plugins/openvino_plugin
datum create -o proj
datum model add -l openvino -p proj --copy -- \
--output-layers=do_ExpandDims_conf/sigmoid \
-d model/ssd_mobilenet_v2_coco.xml \
-w model/ssd_mobilenet_v2_coco.bin \
-i samples/ssd_mobilenet_coco_detection_interp.py
datum import -p proj -f voc VOCdevkit/
datum model run -p proj -m model-0
Classification: mobilenet-v2-pytorch
source /opt/intel/openvino/bin/setupvars.sh
cd datumaro/plugins/openvino_plugin
datum create -o proj
datum model add -l openvino -p proj --copy -- \
-d model/mobilenet-v2-pytorch.xml \
-w model/mobilenet-v2-pytorch.bin \
-i samples/mobilenet_v2_pytorch_interp.py
datum import -p proj -f voc VOCdevkit/
datum model run -p proj -m model-0
Try to be readable and consistent with the existing codebase.
The project mostly follows PEP8 with little differences.
Continuation lines have a standard indentation step by default,
or any other, if it improves readability. For long conditionals use 2 steps.
No trailing whitespaces, 80 characters per line.
Example:
defdo_important_work(parameter1, parameter2, parameter3,
option1=None, option2=None, option3=None)->str:"""
Optional description. Mandatory for API.
Use comments for implementation specific information, use docstrings
to give information to user / developer.
Returns: status (str) - Possible values: 'done', 'failed'
"""... do stuff ...# Use +1 level of indentation for continuation lines
variable_with_a_long_but_meaningful_name = \
function_with_a_long_but_meaningful_name(arg1, arg2, arg3,
kwarg1=value_with_a_long_name, kwarg2=value_with_a_long_name)# long conditions, loops, with etc. also use +1 level of indentationif condition1 and long_condition2 or \
not condition3 and condition4 and condition5 or \
condition6 and condition7:... do other stuff ...elif other_conditions:... some other things ...# in some cases special formatting can improve code readability
specific_case_formatting = np.array([[0,1,1,0],[1,1,0,0],[1,1,0,1],], dtype=np.int32)return status
Environment
The recommended editor is VS Code with the Python language plugin.
Testing
It is expected that all Datumaro functionality is covered and checked by
unit tests. Tests are placed in the tests/ directory. Additional
pre-generated files for tests can be stored in the tests/assets/ directory.
CLI tests are separated from the core tests, they are stored in the
tests/cli/ directory.
DATUM_244 ="Add Snyk integration"
DATUM_BUG_219 ="Return format is not uniform"
# Fully defined in GitHub issues:@pytest.mark.reqids(Requirements.DATUM_244, Requirements.DATUM_333)# And defined any other way:@pytest.mark.reqids(Requirements.DATUM_GENERAL_REQ)
Tests are documented with docs strings. Test descriptions must contain
the following: sections: Description, Expected results and Steps.
deftest_can_convert_polygons_to_mask(self):"""
<b>Description:</b>
Ensure that the dataset polygon annotation can be properly converted
into dataset segmentation mask.
<b>Expected results:</b>
Dataset segmentation mask converted from dataset polygon annotation
is equal to an expected mask.
<b>Steps:</b>
1. Prepare dataset with polygon annotation
2. Prepare dataset with expected mask segmentation mode
3. Convert source dataset to target, with conversion of annotation
from polygon to mask.
4. Verify that resulting segmentation mask is equal to the expected mask.
"""