We provide a collection of detection models pre-trained on the COCO dataset, the Kitti dataset, the Open Images dataset, the AVA v2.1 dataset and the iNaturalist Species Detection Dataset. These models can be useful for out-of-the-box inference if you are interested in categories already in those datasets. They are also useful for initializing your models when training on novel datasets.
In the table below, we list each such pre-trained model including:
samples/configs
directory,Boxes
, and Masks
if applicable )You can un-tar each tar.gz file via, e.g.,:
tar -xzvf ssd_mobilenet_v1_coco.tar.gz
Inside the un-tar'ed directory, you will find:
graph.pbtxt
)model.ckpt.data-00000-of-00001
, model.ckpt.index
, model.ckpt.meta
)frozen_inference_graph.pb
) to be used for out of the box inference
(try this out in the Jupyter notebook!)pipeline.config
) which was used to generate the graph. These
directly correspond to a config file in the
samples/configs) directory but often with a modified score threshold. In the case
of the heavier Faster R-CNN models, we also provide a version of the model
that uses a highly reduced number of proposals for speed.Some remarks on frozen inference graphs:
Note: The asterisk (☆) at the end of model name indicates that this model supports TPU training.
Note: If you download the tar.gz file of quantized models and un-tar, you will get different set of files - a checkpoint, a config file and tflite frozen graphs (txt/binary).
Model name | Speed (ms) | Pascal mAP@0.5 | Outputs |
---|---|---|---|
faster_rcnn_resnet101_kitti | 79 | 87 | Boxes |
Model name | Speed (ms) | Open Images mAP@0.52 | Outputs |
---|---|---|---|
faster_rcnn_inception_resnet_v2_atrous_oidv2 | 727 | 37 | Boxes |
faster_rcnn_inception_resnet_v2_atrous_lowproposals_oidv2 | 347 | Boxes | |
facessd_mobilenet_v2_quantized_open_image_v4 3 | 20 | 73 (faces) | Boxes |
Model name | Speed (ms) | Open Images mAP@0.54 | Outputs |
---|---|---|---|
faster_rcnn_inception_resnet_v2_atrous_oidv4 | 425 | 54 | Boxes |
ssd_mobilenetv2_oidv4 | 89 | 36 | Boxes |
ssd_resnet_101_fpn_oidv4 | 237 | 38 | Boxes |
Model name | Speed (ms) | Pascal mAP@0.5 | Outputs |
---|---|---|---|
faster_rcnn_resnet101_fgvc | 395 | 58 | Boxes |
faster_rcnn_resnet50_fgvc | 366 | 55 | Boxes |
Model name | Speed (ms) | Pascal mAP@0.5 | Outputs |
---|---|---|---|
faster_rcnn_resnet101_ava_v2.1 | 93 | 11 | Boxes |
See MSCOCO evaluation protocol. The COCO mAP numbers here are evaluated on COCO 14 minival set (note that our split is different from COCO 17 Val). A full list of image ids used in our split could be fould here. ↩︎
This is PASCAL mAP with a slightly different way of true positives computation: see Open Images evaluation protocols, oid_V2_detection_metrics. ↩︎
Non-face boxes are dropped during training and non-face groundtruth boxes are ignored when evaluating. ↩︎
This is Open Images Challenge metric: see Open Images evaluation protocols, oid_challenge_detection_metrics. ↩︎