Skip to content

The fusion between RGB and Lidar method for long-tailed 3d detection

Notifications You must be signed in to change notification settings

mayechi/lt3d-lf

Repository files navigation

LONG-TAILED 3D DETECTION VIA 2D LATE FUSION

Late Fusion for LiDAR and RGB detections.

Get started

  1. Clone the LT3D reposityory from https://github.com/neeharperi/LT3D and the LT3D nuscenes-devkit from https://github.com/neeharperi/nuscenes-lt3d.git

  2. You should mkdir a file folder in "./tools" (e.g. "fusion" folder) containing all the files from this project when you need to use mmdetection3d framework.

  3. You should get 3D detections and 2D detections results with .pkl and .txt format respectively. The .pkl results from lidar should be getted by "tools/test.py" in mmdetection3d codebase and the .txt results from camera should be getted by some 2D dectector methods (e.g. DINO). I will provide above two results in lidar_results (https://drive.google.com/file/d/1qbfghLnyBHl_OrY6h4qfKxQFifm2tutd/view?usp=sharing) and camera_results (https://drive.google.com/file/d/1Z8B_ljPVnZ9d2-k1_bjs-jq569hcsv2D/view?usp=share_link) getted by Centerpoint (Hierarchy) and DINO respectively.

  4. Calibrate the 18 classes for long-tailed derections, you can get the best metrics (AP) for each classes from the logs.

python tools/fusion/main_tune.py
  1. Add overlapping mechanism for evaluating the metrics (mAP).
python tools/fusion/main.py
  1. You can get metircs (mAP) for "Many, medium, few" classes.
python get_LCA0_mmf.py
  1. The 'cal.py' script contain precomputed calibrations by different 3D (e.g. Centerpoint) and 2D detections (e.g. DINO).

About

The fusion between RGB and Lidar method for long-tailed 3d detection

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages