You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
UniverseNets are state-of-the-art detectors for universal-scale object detection.
Please refer to our paper for details.
https://arxiv.org/abs/2103.14027
4 GPUs x 16 samples_per_gpu for small test scales ((512, 512) and (224, 224)).
You will be able to use BatchNorm with norm_eval=False even on 1 GPU.
Test scale and test-dev AP
Method
Backbone
Max test scale
TTA
Inf time (fps)
box AP (val)
box AP (test-dev)
UniverseNet
R2-50
(1333, 800)
-
15.8
48.9
49.2
UniverseNet
R2-50
(1600, 960)
-
14.5
49.2
49.5
UniverseNet 20.08s
R-50-C
(1333, 800)
-
31.6
46.9
47.4
UniverseNet 20.08
R2-50
(1333, 800)
-
24.9
48.5
48.8
UniverseNet 20.08d
R2-101
(1333, 800)
-
11.7
50.9
51.3
UniverseNet 20.08d
R2-101
(2000, 1200)
5
-
53.1
53.8
UniverseNet 20.08d
R2-101
(3000, 1800)
13
-
53.5
54.1
TTA: test-time augmentation including horizontal flip and multi-scale testing (numbers denote scales).
Misc.
Other hyperparameters and details for reproduction
Method
warmup_iters
lcconv_padding
GPUs x samples_per_gpu
box AP
UniverseNet
500
0
4x4 -> 8x2
48.9
UniverseNet
1000
1
4x4
48.9
UniverseNet
3665
0
4x4
48.8
The checkpoints in release 20.06 were trained with a warmup_iters of 500.
To make training more stable, the current config sets warmup_iters to 1000. The difference will not affect the final accuracy so much.
In the official SEPC implementation, padding values in lconv and cconv (we call lcconv_padding) are set to 0.
Setting lcconv_padding to 1 doesn't affect accuracy.
To accelerate training for CVPR competitions, we used 8 GPUs for 9-24 epochs, after using 4 GPUs for 1-8 epochs.