I’ve always believed that relying solely on a vision-only perception system is not a viable path forward — even though Tesla currently remains the most advanced player in autonomous driving.
The so-called advantages of vision-only perception — lower cost and algorithmic maturity — are becoming less compelling in today’s highly competitive landscape. TOF and FMCW LiDAR technologies are rapidly dropping in price; even a relatively ordinary project I’ve worked on recently has adopted a solid-state TOF solution. Meanwhile, sensor fusion algorithms are inevitably maturing as well.
The depth maps generated by vision-only systems are essentially pseudo-3D representations. When the technology is immature, depth estimation tends to suffer from errors and is easily deceived by environmental conditions. In contrast, LiDAR can offer millimeter-level — and with FMCW, even sub-millimeter — precision, along with significantly improved resistance to interference.
Moreover, using neural networks to compute depth and perform object detection demands substantial computational power. LiDAR, on the other hand, directly outputs 3D data, greatly reducing the processing burden.
Vision-only perception systems are also highly dependent on lighting conditions and scene texture complexity. LiDAR remains stable under backlight, low-light, and textureless scenarios, and continues to function reliably even in adverse weather such as rain and fog.
The lack of safety redundancy in vision-only approaches makes them unreliable for use in such challenging conditions — and realistically, we can’t expect users to simply avoid driving at night or in bad weather.
The future of autonomous driving will undoubtedly rely on multi-sensor fusion — to enable generalized perception that balances cost control and system robustness, leveraging the unique advantages of each modality.
The goal shouldn’t be to force a vision-only solution, but rather to build heterogeneous systems where vision serves as the backbone, complemented by LiDAR for critical redundancy and precision.
Looking ahead, for high-level autonomous driving (L4 and above), TOF — and especially FMCW LiDAR — will most likely become standard components.