Note
Open Model Zoo is in maintenance mode as a source of models. Check out model tutorials in Jupyter notebooks.
We appreciate your intention to contribute model to the OpenVINO™ Open Model Zoo (OMZ). OMZ is licensed under the Apache* License, Version 2.0. By contributing to the project, you agree to the license and copyright terms therein and release your contribution under these terms. Note that we accept models under permissive licenses, such as MIT, Apache 2.0, and BSD-3-Clause. Otherwise, it might take longer time to get your model approved.
Frameworks supported by the Open Model Zoo:
- TensorFlow*
- PyTorch* (via conversion to ONNX*)
Open Model Zoo also supports models already in the ONNX format.
To contribute to OMZ, create a pull request (PR) in this repository using the master
branch.
Pull requests are strictly formalized and are reviewed by the OMZ maintainers for consistence and legal compliance.
Each PR contributing a model must contain:
- configuration file
model.yml
- documentation of model in markdown format
- accuracy validation configuration file
- (optional) demo
Follow the rules in the sections below before submitting a pull request.
Name your model in OMZ according to the following rules:
- Use a name that is consistent with an original name, but complete match is not necessary
- Use lowercase
- Use
-
(preferable) or_
as delimiters, for spaces are not allowed - Add a suffix according to framework identifier (see
framework
description in the configuration file section for examples), if the model is a reimplementation of an existing model from another framework
This name will be used for downloading, converting, and other operations.
An example of model name is resnet-50-pytorch
Place your files as shown in the table below:
File | Destination |
---|---|
configuration file | models/public/<model_name>/model.yml |
documentation file | models/public/<model_name>/README.md |
validation configuration file | models/public/<model_name>/accuracy-check.yml |
demo | demos/<demo_name>/<language>/ |
Your PR must pass next tests:
- Model is downloadable by the
tools/model_tools/downloader.py
script. See Configuration file for details. - Model is convertible by the
tools/model_tools/converter.py
script. See Model conversion for details. - Model is usable by demo or sample and provides adequate results. See Demo for details.
- Model passes accuracy validation. See Accuracy validation for details.
Your PR may be rejected in some cases, for example:
- If a license is inappropriate (such as GPL-like licenses).
- If a dataset is inaccessible.
- If the PR fails one of the tests above.
The model configuration file contains information about model: what it is, how to download it, and how to convert it to the IR format. This information must be specified in the model.yml
file that must be located in the model subfolder.
The detailed descriptions of file entries provided below.
description
Description of the model. Must match with the description from the model documentation. Use this script for easy update.
task_type
Model task type. If there is no task type of your model, add a new one to the list KNOWN_TASK_TYPES
of the omz_tools._common
module.
model_info
(optional)
Dict for ModelAPI model wrappers. Used for Python Model API class creation.
files
NOTE: Before filling this section, make sure that the model can be downloaded either via a direct HTTP(S) link or from Google Drive*.
Downloadable files. Each file is described by:
name
- sets a file name after downloadingsize
- sets a file sizechecksum
- sets a file hash sumsource
- sets a direct link to a file OR describes a file access parameters
TIP: You can obtain a hash sum using the
sha384sum <file_name>
command on Linux*.
If file is located on Google Drive*, the source
section must contain:
$type: google_drive
id
file ID on Google Drive*
NOTE: If file is on GitHub*, use the specific file version.
postprocessing
(optional)
Post processing of the downloaded files.
For unpacking archive:
$type: unpack_archive
file
— Archive file nameformat
— Archive format (zip | tar | gztar | bztar | xztar)
For replacement operation:
$type: regex_replace
file
— Name of file to run replacement inpattern
— Regular expressionreplacement
— Replacement stringcount
(optional) — Exact number of replacements (if number ofpattern
occurrences less then this number, downloading will be aborted)
input_info
List of inputs containing the information about input name, shape and layout. For example:
input_info:
- name: Placeholder
shape: [1, 224, 224, 3]
layout: NHWC
conversion_to_onnx_args
(only for PyTorch* models)
List of ONNX* conversion parameters, see model_optimizer_args
for details.
model_optimizer_args
Conversion parameters (learn more in the Model conversion section). For example:
- --mean_values=data[127.5]
- --scale_values=data[127.5]
- --reverse_input_channels
- --output=prob
- --input_model=$conv_dir/googlenet-v3.onnx
NOTE: Do not specify
framework
,data_type
,model_name
andoutput_dir
, since they are deduced automatically.
NOTE:
$dl_dir
used to substitute subdirectory of downloaded model and$conv_dir
used to substitute subdirectory of converted model.
framework
Framework of the original model (see here for details).
license
URL of the model license.
This example shows how to download the classification model efficientdet-d1-tf* pretrained in TensorFlow*.
description: >-
The "efficientdet-d1-tf" model is one of the EfficientDet <https://arxiv.org/abs/1911.09070>
models designed to perform object detection. This model was pre-trained in TensorFlow*.
All the EfficientDet models have been pre-trained on the Common Objects in Context
(COCO) <https://cocodataset.org/#home> image database. For details about this family
of models, check out the Google AutoML repository <https://github.com/google/automl/tree/master/efficientdet>.
task_type: detection
files:
- name: efficientdet-d1_frozen.pb
size: 28613804
checksum: f89a4fe6072e5dddc71c0532261e1cd89a69472d92a85445437e8b57129708adbad35cf09a8e7603a25b05ee0ebdf4cb
source: https://storage.openvinotoolkit.org/repositories/open_model_zoo/public/2023.0/efficientdet-d1-tf/efficientdet-d1_frozen.pb
input_info:
- name: image_arrays
shape: [1, 640, 640, 3]
layout: NHWC
model_optimizer_args:
- --reverse_input_channels
- --input_model=$dl_dir/efficientdet-d1_frozen.pb
- --transformations_config=$mo_ext_dir/front/tf/automl_efficientdet.json
framework: tf
license: https://raw.githubusercontent.com/google/automl/master/LICENSE
OpenVINO™ Runtime supports models in the Intermediate Representation (IR) format. A model from any supported framework can be converted to IR using the Model Optimizer tool included in the OpenVINO™ toolkit. Find more information about conversion in the [Model Optimizer Developer Guide](@ref openvino_docs_MO_DG_Deep_Learning_Model_Optimizer_DevGuide). After a successful conversion, you get a model in the IR format, with the *.xml
file representing the net graph and the *.bin
file containing the net parameters.
NOTE: Image preprocessing parameters (mean and scale) must be built into a converted model to simplify model usage.
NOTE 2: If a model input is a color image, color channel order should be
BGR
.
A demo shows the main idea of how to infer a model using OpenVINO™. If your model solves one of the tasks supported by the Open Model Zoo, try to find an appropriate option from demos or samples. Otherwise, you must provide your own demo (C++ or Python).
The demo's name should end with _demo
suffix to follow the convention of the project.
Demos are required to support the following args:
-h, --help
show this help message and exit-m <MODEL FILE>
path to an .xml file with a trained model. If the demo uses several models an extended syntax can be used, like--mdet
-i <INPUT>
input to process. For vision tasks the input might be a path to single image or video file, a path to folder of images, or numeric camera id. For vision tasks the default value must be0
. For speech/audio tasks input is path to WAV file. For NLP tasks input might be a path to text file or just quoted sentence of text.-d <DEVICE>
specify a device to infer on (the list of available devices is shown below). Default is CPU-o <FILE PATTERN>
pattern for output file(s) to save
Add README.md
file, which describes demo usage. Update demos' README.md by adding your demo to the list.
Accuracy validation can be performed by the Accuracy Checker tool. This tool can use either IE to run a converted model, or an original framework to run an original model. Accuracy Checker supports lots of datasets, metrics and preprocessing options, which simplifies validation if a task is supported by the tool. You only need to create a configuration file that contains necessary parameters for accuracy validation (specify a dataset and annotation, pre- and post-processing parameters, accuracy metrics to compute and so on) of converted model. For details, refer to Testing new models.
If a model uses a dataset which is not supported by the Accuracy Checker, you also must provide the license and the link to it and mention it in the PR description.
When the configuration file is ready, you must run the Accuracy Checker to obtain metric results. If they match your results, that means conversion was successful and the Accuracy Checker fully supports your model, metric and dataset. Otherwise, recheck the conversion parameters or the validation configuration file.
This example uses validation configuration file for efficientdet-d1-tf* from TensorFlow*:
models:
- name: efficientdet-d1-tf
launchers:
- framework: openvino
adapter: ssd
datasets:
- name: ms_coco_detection_90_class_without_background
preprocessing:
- type: resize
aspect_ratio_scale: fit_to_window
size: 640
- type: padding
size: 640
pad_type: right_bottom
postprocessing:
- type: faster_rcnn_postprocessing_resize
size: 640
- type: shift_labels
offset: 1
metrics:
- type: coco_precision
reference: 0.3754
Documentation is a very important part of model contribution as it helps to better understand the possible usage of the model. It must be located in a README.md
file in the model subdirectory.
The documentation should contain:
- description of a model
- main purpose
- features
- references to a paper or/and a source
- model specification
- type
- framework
- GFLOPs (if available)
- number of parameters (if available)
- validation dataset description and/or a link
- main accuracy values (also description of a metric)
- detailed description of input and output for original and converted models
- the model's licensing terms
Learn the detailed structure and headers naming convention from any model documentation (for example, densenet-121-tf).
[*] Other names and brands may be claimed as the property of others.
OpenVINO is a trademark of Intel Corporation or its subsidiaries in the U.S. and/or other countries.
Copyright © 2018-2024 Intel Corporation
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.