Skip to content

Commit fe3f809

Browse files
authored
Bump version to v3.2.0 (#11030)
2 parents f78af77 + dfad04c commit fe3f809

File tree

333 files changed

+30994
-3378
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

333 files changed

+30994
-3378
lines changed

.circleci/test.yml

Lines changed: 9 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -94,21 +94,27 @@ jobs:
9494
type: string
9595
cuda:
9696
type: enum
97-
enum: ["11.1", "11.7"]
97+
enum: ["11.1", "11.7", "11.8"]
9898
cudnn:
9999
type: integer
100100
default: 8
101101
machine:
102-
image: ubuntu-2004-cuda-11.4:202110-01
102+
image: linux-cuda-11:default
103103
# docker_layer_caching: true
104-
resource_class: gpu.nvidia.small
104+
resource_class: gpu.nvidia.small.multi
105105
steps:
106106
- checkout
107107
- run:
108108
# CLoning repos in VM since Docker doesn't have access to the private key
109109
name: Clone Repos
110110
command: |
111111
git clone -b main --depth 1 ssh://[email protected]/open-mmlab/mmengine.git /home/circleci/mmengine
112+
- run:
113+
name: Install nvidia-container-toolkit and Restart Docker
114+
command: |
115+
sudo apt-get update
116+
sudo apt-get install -y nvidia-container-toolkit
117+
sudo systemctl restart docker
112118
- run:
113119
name: Build Docker image
114120
command: |

.dev_scripts/benchmark_train_models.txt

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,11 +2,13 @@ atss/atss_r50_fpn_1x_coco.py
22
faster_rcnn/faster-rcnn_r50_fpn_1x_coco.py
33
mask_rcnn/mask-rcnn_r50_fpn_1x_coco.py
44
cascade_rcnn/cascade-mask-rcnn_r50_fpn_1x_coco.py
5+
configs/grounding_dino/grounding_dino_swin-t_finetune_16xb2_1x_coco.py
6+
configs/glip/glip_atss_swin-t_a_fpn_dyhead_16xb2_ms-2x_funtune_coco.py
7+
configs/ddq/ddq-detr-4scale_r50_8xb2-12e_coco.py
58
panoptic_fpn/panoptic-fpn_r50_fpn_1x_coco.py
69
retinanet/retinanet_r50_fpn_1x_coco.py
710
rtmdet/rtmdet_s_8xb32-300e_coco.py
811
rtmdet/rtmdet-ins_s_8xb32-300e_coco.py
9-
deformable_detr/deformable-detr_r50_16xb2-50e_coco.py
1012
fcos/fcos_r50-caffe_fpn_gn-head-center-normbbox-centeronreg-giou_1x_coco.py
1113
centernet/centernet-update_r50-caffe_fpn_ms-1x_coco.py
1214
dino/dino-4scale_r50_8xb2-12e_coco.py

.dev_scripts/gather_models.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
import yaml
1313
from mmengine.config import Config
1414
from mmengine.fileio import dump
15-
from mmengine.utils import mkdir_or_exist, scandir
15+
from mmengine.utils import digit_version, mkdir_or_exist, scandir
1616

1717

1818
def ordered_yaml_dump(data, stream=None, Dumper=yaml.SafeDumper, **kwds):
@@ -45,7 +45,7 @@ def process_checkpoint(in_file, out_file):
4545

4646
# if it is necessary to remove some sensitive data in checkpoint['meta'],
4747
# add the code here.
48-
if torch.__version__ >= '1.6':
48+
if digit_version(torch.__version__) >= digit_version('1.6'):
4949
torch.save(checkpoint, out_file, _use_new_zipfile_serialization=False)
5050
else:
5151
torch.save(checkpoint, out_file)

README.md

Lines changed: 52 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,7 @@
2525
[![license](https://img.shields.io/github/license/open-mmlab/mmdetection.svg)](https://github.com/open-mmlab/mmdetection/blob/main/LICENSE)
2626
[![open issues](https://isitmaintained.com/badge/open/open-mmlab/mmdetection.svg)](https://github.com/open-mmlab/mmdetection/issues)
2727
[![issue resolution](https://isitmaintained.com/badge/resolution/open-mmlab/mmdetection.svg)](https://github.com/open-mmlab/mmdetection/issues)
28+
[![Open in OpenXLab](https://cdn-static.openxlab.org.cn/app-center/openxlab_demo.svg)](https://openxlab.org.cn/apps?search=mmdet)
2829

2930
[📘Documentation](https://mmdetection.readthedocs.io/en/latest/) |
3031
[🛠️Installation](https://mmdetection.readthedocs.io/en/latest/get_started.html) |
@@ -102,6 +103,52 @@ Apart from MMDetection, we also released [MMEngine](https://github.com/open-mmla
102103

103104
### Highlight
104105

106+
**v3.2.0** was released in 12/10/2023:
107+
108+
**1. Detection Transformer SOTA Model Collection**
109+
(1) Supported four updated and stronger SOTA Transformer models: [DDQ](configs/ddq/README.md), [CO-DETR](projects/CO-DETR/README.md), [AlignDETR](projects/AlignDETR/README.md), and [H-DINO](projects/HDINO/README.md).
110+
(2) Based on CO-DETR, MMDet released a model with a COCO performance of 64.1 mAP.
111+
(3) Algorithms such as DINO support `AMP/Checkpoint/FrozenBN`, which can effectively reduce memory usage.
112+
113+
**2. [Comprehensive Performance Comparison between CNN and Transformer](<(projects/RF100-Benchmark/README.md)>)**
114+
RF100 consists of a dataset collection of 100 real-world datasets, including 7 domains. It can be used to assess the performance differences of Transformer models like DINO and CNN-based algorithms under different scenarios and data volumes. Users can utilize this benchmark to quickly evaluate the robustness of their algorithms in various scenarios.
115+
116+
<div align=center>
117+
<img src="https://github.com/open-mmlab/mmdetection/assets/17425982/86420903-36a8-410d-9251-4304b9704f7d"/>
118+
</div>
119+
120+
**3. Support for [GLIP](configs/glip/README.md) and [Grounding DINO](configs/grounding_dino/README.md) fine-tuning, the only algorithm library that supports Grounding DINO fine-tuning**
121+
The Grounding DINO algorithm in MMDet is the only library that supports fine-tuning. Its performance is one point higher than the official version, and of course, GLIP also outperforms the official version.
122+
We also provide a detailed process for training and evaluating Grounding DINO on custom datasets. Everyone is welcome to give it a try.
123+
124+
| Model | Backbone | Style | COCO mAP | Official COCO mAP |
125+
| :----------------: | :------: | :-------: | :--------: | :---------------: |
126+
| Grounding DINO-T | Swin-T | Zero-shot | 48.5 | 48.4 |
127+
| Grounding DINO-T | Swin-T | Finetune | 58.1(+0.9) | 57.2 |
128+
| Grounding DINO-B | Swin-B | Zero-shot | 56.9 | 56.7 |
129+
| Grounding DINO-B | Swin-B | Finetune | 59.7 | |
130+
| Grounding DINO-R50 | R50 | Scratch | 48.9(+0.8) | 48.1 |
131+
132+
**4. Support for the open-vocabulary detection algorithm [Detic](projects/Detic_new/README.md) and multi-dataset joint training.**
133+
**5. Training detection models using [FSDP and DeepSpeed](<(projects/example_largemodel/README.md)>).**
134+
135+
| ID | AMP | GC of Backbone | GC of Encoder | FSDP | Peak Mem (GB) | Iter Time (s) |
136+
| :-: | :-: | :------------: | :-----------: | :--: | :-----------: | :-----------: |
137+
| 1 | | | | | 49 (A100) | 0.9 |
138+
| 2 || | | | 39 (A100) | 1.2 |
139+
| 3 | || | | 33 (A100) | 1.1 |
140+
| 4 ||| | | 25 (A100) | 1.3 |
141+
| 5 | ||| | 18 | 2.2 |
142+
| 6 |||| | 13 | 1.6 |
143+
| 7 | |||| 14 | 2.9 |
144+
| 8 ||||| 8.5 | 2.4 |
145+
146+
**6. Support for the [V3Det](configs/v3det/README.md) dataset, a large-scale detection dataset with over 13,000 categories.**
147+
148+
<div align=center>
149+
<img width=960 src="https://github.com/open-mmlab/mmdetection/assets/17425982/9c216387-02be-46e6-b0f2-b856f80f6d84"/>
150+
</div>
151+
105152
We are excited to announce our latest work on real-time object recognition tasks, **RTMDet**, a family of fully convolutional single-stage detectors. RTMDet not only achieves the best parameter-accuracy trade-off on object detection from tiny to extra-large model sizes but also obtains new state-of-the-art performance on instance segmentation and rotated object detection tasks. Details can be found in the [technical report](https://arxiv.org/abs/2212.07784). Pre-trained models are [here](configs/rtmdet).
106153

107154
[![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/rtmdet-an-empirical-study-of-designing-real/real-time-instance-segmentation-on-mscoco)](https://paperswithcode.com/sota/real-time-instance-segmentation-on-mscoco?p=rtmdet-an-empirical-study-of-designing-real)
@@ -118,13 +165,6 @@ We are excited to announce our latest work on real-time object recognition tasks
118165
<img src="https://user-images.githubusercontent.com/12907710/208044554-1e8de6b5-48d8-44e4-a7b5-75076c7ebb71.png"/>
119166
</div>
120167

121-
**v3.1.0** was released in 30/6/2023:
122-
123-
- Supports tracking algorithms including multi-object tracking (MOT) algorithms SORT, DeepSORT, StrongSORT, OCSORT, ByteTrack, QDTrack, and video instance segmentation (VIS) algorithm MaskTrackRCNN, Mask2Former-VIS.
124-
- Support [ViTDet](projects/ViTDet)
125-
- Supports inference and evaluation of multimodal algorithms [GLIP](configs/glip) and [XDecoder](projects/XDecoder), and also supports datasets such as COCO semantic segmentation, COCO Caption, ADE20k general segmentation, and RefCOCO. GLIP fine-tuning will be supported in the future.
126-
- Provides a [gradio demo](https://github.com/open-mmlab/mmdetection/blob/dev-3.x/projects/gradio_demo/README.md) for image type tasks of MMDetection, making it easy for users to experience.
127-
128168
## Installation
129169

130170
Please refer to [Installation](https://mmdetection.readthedocs.io/en/latest/get_started.html) for installation instructions.
@@ -236,9 +276,12 @@ Results and models are available in the [model zoo](docs/en/model_zoo.md).
236276
<li><a href="configs/dab_detr">DAB-DETR (ICLR'2022)</a></li>
237277
<li><a href="configs/dino">DINO (ICLR'2023)</a></li>
238278
<li><a href="configs/glip">GLIP (CVPR'2022)</a></li>
279+
<li><a href="configs/ddq">DDQ (CVPR'2023)</a></li>
239280
<li><a href="projects/DiffusionDet">DiffusionDet (ArXiv'2023)</a></li>
240281
<li><a href="projects/EfficientDet">EfficientDet (CVPR'2020)</a></li>
282+
<li><a href="projects/ViTDet">ViTDet (ECCV'2022)</a></li>
241283
<li><a href="projects/Detic">Detic (ECCV'2022)</a></li>
284+
<li><a href="projects/CO-DETR">CO-DETR (ICCV'2023)</a></li>
242285
</ul>
243286
</td>
244287
<td>
@@ -260,13 +303,15 @@ Results and models are available in the [model zoo](docs/en/model_zoo.md).
260303
<li><a href="projects/SparseInst">SparseInst (CVPR'2022)</a></li>
261304
<li><a href="configs/rtmdet">RTMDet (ArXiv'2022)</a></li>
262305
<li><a href="configs/boxinst">BoxInst (CVPR'2021)</a></li>
306+
<li><a href="projects/ConvNeXt-V2">ConvNeXt-V2 (Arxiv'2023)</a></li>
263307
</ul>
264308
</td>
265309
<td>
266310
<ul>
267311
<li><a href="configs/panoptic_fpn">Panoptic FPN (CVPR'2019)</a></li>
268312
<li><a href="configs/maskformer">MaskFormer (NeurIPS'2021)</a></li>
269313
<li><a href="configs/mask2former">Mask2Former (ArXiv'2021)</a></li>
314+
<li><a href="configs/XDecoder">XDecoder (CVPR'2023)</a></li>
270315
</ul>
271316
</td>
272317
<td>

README_zh-CN.md

Lines changed: 55 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,7 @@
2525
[![license](https://img.shields.io/github/license/open-mmlab/mmdetection.svg)](https://github.com/open-mmlab/mmdetection/blob/main/LICENSE)
2626
[![open issues](https://isitmaintained.com/badge/open/open-mmlab/mmdetection.svg)](https://github.com/open-mmlab/mmdetection/issues)
2727
[![issue resolution](https://isitmaintained.com/badge/resolution/open-mmlab/mmdetection.svg)](https://github.com/open-mmlab/mmdetection/issues)
28+
[![Open in OpenXLab](https://cdn-static.openxlab.org.cn/app-center/openxlab_demo.svg)](https://openxlab.org.cn/apps?search=mmdet)
2829

2930
[📘使用文档](https://mmdetection.readthedocs.io/zh_CN/latest/) |
3031
[🛠️安装教程](https://mmdetection.readthedocs.io/zh_CN/latest/get_started.html) |
@@ -101,6 +102,53 @@ MMDetection 是一个基于 PyTorch 的目标检测开源工具箱。它是 [Ope
101102

102103
### 亮点
103104

105+
**v3.2.0** 版本已经在 2023.10.12 发布:
106+
107+
**1. 检测 Transformer SOTA 模型大合集**
108+
(1) 支持了 [DDQ](configs/ddq/README.md)[CO-DETR](projects/CO-DETR/README.md)[AlignDETR](projects/AlignDETR/README.md)[H-DINO](projects/HDINO/README.md) 4 个更新更强的 SOTA Transformer 模型
109+
(2) 基于 CO-DETR, MMDet 中发布了 COCO 性能为 64.1 mAP 的模型
110+
(3) DINO 等算法支持 AMP/Checkpoint/FrozenBN,可以有效降低显存
111+
112+
**2. [提供了全面的 CNN 和 Transformer 的性能对比](projects/RF100-Benchmark/README_zh-CN.md)**
113+
RF100 是由 100 个现实收集的数据集组成,包括 7 个域,可以验证 DINO 等 Transformer 模型和 CNN 类算法在不同场景不同数据量下的性能差异。用户可以用这个 Benchmark 快速验证自己的算法在不同场景下的鲁棒性。
114+
115+
<div align=center>
116+
<img src="https://github.com/open-mmlab/mmdetection/assets/17425982/86420903-36a8-410d-9251-4304b9704f7d"/>
117+
</div>
118+
119+
**3. 支持了 [GLIP](configs/glip/README.md)[Grounding DINO](configs/grounding_dino/README.md) 微调,全网唯一支持 Grounding DINO 微调**
120+
MMDet 中的 Grounding DINO 是全网唯一支持微调的算法库,且性能高于官方 1 个点,当然 GLIP 也比官方高。
121+
我们还提供了详细的 Grounding DINO 在自定义数据集上训练评估的流程,欢迎大家试用。
122+
123+
| Model | Backbone | Style | COCO mAP | Official COCO mAP |
124+
| :----------------: | :------: | :-------: | :--------: | :---------------: |
125+
| Grounding DINO-T | Swin-T | Zero-shot | 48.5 | 48.4 |
126+
| Grounding DINO-T | Swin-T | Finetune | 58.1(+0.9) | 57.2 |
127+
| Grounding DINO-B | Swin-B | Zero-shot | 56.9 | 56.7 |
128+
| Grounding DINO-B | Swin-B | Finetune | 59.7 | |
129+
| Grounding DINO-R50 | R50 | Scratch | 48.9(+0.8) | 48.1 |
130+
131+
**4. 支持开放词汇检测算法 [Detic](projects/Detic_new/README.md) 并提供多数据集联合训练可能**
132+
133+
**5. 轻松使用 [FSDP 和 DeepSpeed 训练检测模型](projects/example_largemodel/README_zh-CN.md)**
134+
135+
| ID | AMP | GC of Backbone | GC of Encoder | FSDP | Peak Mem (GB) | Iter Time (s) |
136+
| :-: | :-: | :------------: | :-----------: | :--: | :-----------: | :-----------: |
137+
| 1 | | | | | 49 (A100) | 0.9 |
138+
| 2 || | | | 39 (A100) | 1.2 |
139+
| 3 | || | | 33 (A100) | 1.1 |
140+
| 4 ||| | | 25 (A100) | 1.3 |
141+
| 5 | ||| | 18 | 2.2 |
142+
| 6 |||| | 13 | 1.6 |
143+
| 7 | |||| 14 | 2.9 |
144+
| 8 ||||| 8.5 | 2.4 |
145+
146+
**6. 支持了 [V3Det](configs/v3det/README.md) 1.3w+ 类别的超大词汇检测数据集**
147+
148+
<div align=center>
149+
<img width=960 src="https://github.com/open-mmlab/mmdetection/assets/17425982/9c216387-02be-46e6-b0f2-b856f80f6d84"/>
150+
</div>
151+
104152
我们很高兴向大家介绍我们在实时目标识别任务方面的最新成果 RTMDet,包含了一系列的全卷积单阶段检测模型。 RTMDet 不仅在从 tiny 到 extra-large 尺寸的目标检测模型上实现了最佳的参数量和精度的平衡,而且在实时实例分割和旋转目标检测任务上取得了最先进的成果。 更多细节请参阅[技术报告](https://arxiv.org/abs/2212.07784)。 预训练模型可以在[这里](configs/rtmdet)找到。
105153

106154
[![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/rtmdet-an-empirical-study-of-designing-real/real-time-instance-segmentation-on-mscoco)](https://paperswithcode.com/sota/real-time-instance-segmentation-on-mscoco?p=rtmdet-an-empirical-study-of-designing-real)
@@ -117,13 +165,6 @@ MMDetection 是一个基于 PyTorch 的目标检测开源工具箱。它是 [Ope
117165
<img src="https://user-images.githubusercontent.com/12907710/208044554-1e8de6b5-48d8-44e4-a7b5-75076c7ebb71.png"/>
118166
</div>
119167

120-
**v3.1.0** 版本已经在 2023.6.30 发布:
121-
122-
- 支持 Tracking 类算法,包括多目标跟踪 MOT 算法 SORT、DeepSORT、StrongSORT、OCSORT、ByteTrack、QDTrack 和视频实例分割 VIS 算法 MaskTrackRCNN、Mask2Former-VIS。
123-
- 支持 [ViTDet](projects/ViTDet)
124-
- 支持多模态开放检测算法 [GLIP](configs/glip)[XDecoder](projects/XDecoder) 推理和评估,并同时支持了 COCO 语义分割、COCO Caption、ADE20k 通用分割、RefCOCO 等数据集。后续将支持 GLIP 微调
125-
- 提供了包括 MMDetection 图片任务的 [gradio demo](https://github.com/open-mmlab/mmdetection/blob/dev-3.x/projects/gradio_demo/README.md),方便用户快速体验
126-
127168
## 安装
128169

129170
请参考[快速入门文档](https://mmdetection.readthedocs.io/zh_CN/latest/get_started.html)进行安装。
@@ -237,9 +278,12 @@ MMDetection 是一个基于 PyTorch 的目标检测开源工具箱。它是 [Ope
237278
<li><a href="configs/dab_detr">DAB-DETR (ICLR'2022)</a></li>
238279
<li><a href="configs/dino">DINO (ICLR'2023)</a></li>
239280
<li><a href="configs/glip">GLIP (CVPR'2022)</a></li>
281+
<li><a href="configs/ddq">DDQ (CVPR'2023)</a></li>
240282
<li><a href="projects/DiffusionDet">DiffusionDet (ArXiv'2023)</a></li>
241283
<li><a href="projects/EfficientDet">EfficientDet (CVPR'2020)</a></li>
284+
<li><a href="projects/ViTDet">ViTDet (ECCV'2022)</a></li>
242285
<li><a href="projects/Detic">Detic (ECCV'2022)</a></li>
286+
<li><a href="projects/CO-DETR">CO-DETR (ICCV'2023)</a></li>
243287
</ul>
244288
</td>
245289
<td>
@@ -261,13 +305,15 @@ MMDetection 是一个基于 PyTorch 的目标检测开源工具箱。它是 [Ope
261305
<li><a href="projects/SparseInst">SparseInst (CVPR'2022)</a></li>
262306
<li><a href="configs/rtmdet">RTMDet (ArXiv'2022)</a></li>
263307
<li><a href="configs/boxinst">BoxInst (CVPR'2021)</a></li>
308+
<li><a href="projects/ConvNeXt-V2">ConvNeXt-V2 (Arxiv'2023)</a></li>
264309
</ul>
265310
</td>
266311
<td>
267312
<ul>
268313
<li><a href="configs/panoptic_fpn">Panoptic FPN (CVPR'2019)</a></li>
269314
<li><a href="configs/maskformer">MaskFormer (NeurIPS'2021)</a></li>
270315
<li><a href="configs/mask2former">Mask2Former (ArXiv'2021)</a></li>
316+
<li><a href="configs/XDecoder">XDecoder (CVPR'2023)</a></li>
271317
</ul>
272318
</td>
273319
<td>
@@ -443,10 +489,10 @@ MMDetection 是一款由来自不同高校和企业的研发人员共同参与
443489

444490
## 欢迎加入 OpenMMLab 社区
445491

446-
扫描下方的二维码可关注 OpenMMLab 团队的 [知乎官方账号](https://www.zhihu.com/people/openmmlab)加入 OpenMMLab 团队的 [官方交流 QQ 群](https://jq.qq.com/?_wv=1027&k=aCvMxdr3)
492+
扫描下方的二维码可关注 OpenMMLab 团队的 [知乎官方账号](https://www.zhihu.com/people/openmmlab)扫描下方微信二维码添加喵喵好友,进入 MMDectection 微信交流社群。【加好友申请格式:研究方向+地区+学校/公司+姓名】
447493

448494
<div align="center">
449-
<img src="resources/zhihu_qrcode.jpg" height="400" /> <img src="resources/qq_group_qrcode.jpg" height="400" />
495+
<img src="resources/zhihu_qrcode.jpg" height="400" /> <img src="resources/miaomiao_qrcode.jpg" height="400" />
450496
</div>
451497

452498
我们会在 OpenMMLab 社区为大家

0 commit comments

Comments
 (0)