BEVCorner, a novel framework that synergistically integrates monocular and multiview pipelines for enhanced 3D object detection in autonomous driving. By fusing depth maps from Bird’s-Eye View (BEV) with object-centric depth estimates from monocular detection, BEVCorner enhances both global context and local precision, addressing the limitations of existing methods in depth precision, occlusion robustness, and computational efficiency. For more details, please refer to our paper.
Step 0. Install pytorch(v1.9.0).
Step 1. Install MMDetection3D(v1.0.0rc4).
Step 2. Install requirements.
pip install -r requirements.txtStep 3. Install BEVDepth(gpu required).
cd bevdepth/layers/backbones/DCNv2
. make.sh
cd ../../../..
python setup.py developPersonal note
cp /root/.cache/torch/checkpoints/dla34-ba72cf86.pth /root/.cache/torch/hub/checkpoints/dla34-ba72cf86.pthStep 0. Download nuScenes official dataset.
Step 1. Symlink the dataset root to ./data/.
ln -s [nuscenes root] ./data/
The directory will be as follows.
BEVCorner
├── data
│ ├── nuScenes
│ │ ├── maps
│ │ ├── samples
│ │ ├── sweeps
│ │ ├── v1.0-test
| | ├── v1.0-trainval
Step 2. Prepare infos.
python scripts/gen_info.py
Run from VSCode.
All the commands are provided in .vscode/launch.json. Refer to Debug code with Visual Studio Code on how to run them.
Train.
python [EXP_PATH] --amp_backend native -b 8 --gpus 8
Eval.
python [EXP_PATH] --ckpt_path [CKPT_PATH] -e -b 8 --gpus 8
If you use BEVCorner in your research, please cite our work by using the following BibTeX entry:
(TBD)