๐ Welcome to the 3D Object Detection Hub
This site was born out of a Masterโs thesis effort to centralize and compare every major 3D-OD method using multiple sensor modalities. While many prior surveys focus on one sensor or era of work, here youโll find:
A unified, searchable, and maintainable catalog
covering camera-only, LiDAR-only, and multi-modal fusion approaches - all organized by sensor and representation.
If you use this site or data, please cite our publication:
Valverde, M., Moutinho, A., & Zacchi, J. V. (2025). A Survey of Deep Learning-Based 3D Object Detection Methods for Autonomous Driving Across Different Sensor Modalities. Sensors.
๐ Navigate
-
๐ Datasets
A concise table of the leading benchmarks (KITTI, nuScenes, Waymo, โฆ) and their key properties. -
๐ ๏ธ Models
A comprehensive database of published 3D-OD methods.
โข Filter by sensor โ representation
โข Search for a method by name
โข Sort by year, mAP, runtime, etc. -
๐ References
The full bibliography of every paper, library, and dataset used to create this work. -
๐ค About
Background on the site, the underlying thesis, and contact details.
๐ Why This Hub?
Although 3D object detection has exploded over the last decade, existing resources often:
- โ๏ธ Cover only one modality (e.g. camera-only or LiDAR-only).
- โ๏ธ Become outdated as new methods appear.
- โ๏ธ Lack a unified taxonomy across sensors & representations.
This Hub aggregates:
- Multi-sensor methods (monocular, stereo, multiview, point-cloud, voxels, fusionโฆ).
- Performance comparisons across KITTI, nuScenes & Waymo.
- Runtime metrics to inform real-world feasibility.
- Interactive filtering, searching, and sorting.
๐ What Youโll Find
๐ Datasets & Benchmarks
- KITTI โ The foundational benchmark (2 cameras + 1 LiDAR) with 15 000 annotated frames.
- nuScenes โ Multi-modal data from 6 cameras, 1 LiDAR, 5 radars over 1 000 scenes.
- Waymo Open โ High-density 5 ร LiDAR + 5 ร camera over 1 150 segments and 230 000+ frames.
โฆand more (ApolloScape, Argoverse, Lyft Level 5, H3D).
๐ ๏ธ Methods & Models
- Monocular (Mono3D, SMOKE, DD3D, MonoFlex, โฆ)
- Stereo & Multiview (LIGA-Stereo, Pseudo-BEV, โฆ)
- LiDAR-only (PointPillars, PV-RCNN, SECOND, โฆ)
- Multi-Modal Fusion (MVXNet, CLOCS, FUTR3D, โฆ)
Each entry shows accuracy, inference time, paper link, and code availability.
๐ How to Use
- Browse the Datasets page to choose your benchmark.
- Visit Models to filter by sensor & representation.
- Search or sort by mAP, runtime, year, or code availability.
- Click โLinkโ to read the original paperโcode links when available.
- Cite via the References section when you publish your own results! ๐
๐โโ๏ธ About & Contact
For background on the thesis, site construction, and to reach out, visit ๐ About or drop me a line at โ๏ธ miguel.heitor.valverde@tecnico.ulisboa.pt.
Keep Exploring!
Whether youโre benchmarking a new sensor, prototyping a fusion network, or writing the next SOTA paper, the 3D Object Detection Hub is here to accelerate your research. ๐ฌ๐
๐ฎ Future Work
- ๐ Continuously update with new methods & datasets.
- โ๏ธ Introduce a Sensor Guide: strengths & trade-offs of cameras, LiDARs, radars.
- ๐ Expand to 2D OD and Depth Estimation surveys.
- ๐ Include references for a paper to be published, detailing how this study was carried out and its methodology.
๐ Citation
If you use this site or data, please cite our publication:
Valverde, M., Moutinho, A., & Zacchi, J. V. (2025). A Survey of Deep Learning-Based 3D Object Detection Methods for Autonomous Driving Across Different Sensor Modalities. Sensors.