Autonomous vehicles rely on 3D perception to understand their surrounding environment. Although there has been remarkable progress in enhancing the accuracy of perception models, their efficiency still lags behind real-time performance, impeding their use in real-world applications.
In this lecture, I will present our recent work, BEVFusion (ICRA 2023), which facilitates efficient multi-task multi-sensor fusion by unifying camera, LiDAR, and radar features in a shared bird's-eye view (BEV) space. We addressed an efficiency bottleneck by accelerating the key view transformation operator by 40x. BEVFusion achieved the leading solution on three popular 3D perception benchmarks, including nuScenes, Argoverse, and Waymo, across different tasks, such as object detection, object tracking, and map segmentation. It has received more than 1k GitHub stars since its release.
Subsequently, I will discuss two of our latest works, FlatFormer (CVPR 2023) and SparseViT (CVPR 2023), which aim to accelerate 2D image and 3D point cloud backbones for perception. FlatFormer is an efficient point cloud transformer that attains real-time performance on edge GPUs and is faster than sparse convolutional methods while retaining superior accuracy. SparseViT explores the idea of spatial sparsity in the context of 2D image transformers and delivers a 1.5 times measured speedup compared with its dense counterpart without compromising accuracy.
Zhijian Liu is a final-year Ph.D. student at MIT, advised by Professor Song Han. He received his B.Eng. degree from Shanghai Jiao Tong University in 2018. His research focuses on efficient algorithms and systems for deep learning, with applications in computer vision. His work on efficient 3D deep learning ranked first on multiple competitive 3D benchmarks, won the first place in the nuScenes LiDAR segmentation challenge, and was featured by media outlets, such as NVIDIA News and MIT News. He was selected as the recipient of Qualcomm Innovation Fellowship, NVIDIA Graduate Fellowship (declined) and MIT Ho-Ching and Han-Ching Fund Award.