Model Zoo¶
Mirror sites¶
We use AWS as the main site to host our model zoo, and maintain a mirror on aliyun.
You can replace https://s3.ap-northeast-2.amazonaws.com/open-mmlab
with https://open-mmlab.oss-cn-beijing.aliyuncs.com
in model urls.
Common settings¶
We use distributed training.
For fair comparison with other codebases, we report the GPU memory as the maximum value of
torch.cuda.max_memory_allocated()
for all 8 GPUs. Note that this value is usually less than whatnvidia-smi
shows.We report the inference time as the total time of network forwarding and post-processing, excluding the data loading time. Results are obtained with the script benchmark.py which computes the average time on 2000 images.
Baselines¶
PointPillars¶
Please refer to PointPillars for details.
Dynamic Voxelization¶
Please refer to Dynamic Voxelization for details.