LiDAR-Camera fusion is pivotal for perceiving and understanding complex traffic environments, particularly valuable in autonomous driving and traffic monitoring. Traditional calibration algorithms, primarily designed for onboard sensors, are inadequate for roadside setups where sensors are positioned higher and more dispersed. To address this challenge, we introduce the RLCFormer, a Transformer-based network specifically tailored for precise calibration of roadside sensors. This method innovatively integrates depth and RGB features, utilizing correlation layers and a Transformer decoder to accurately match features across modalities. Evaluated on the DAIR-V2X-I Roadside 3D detection dataset, the RLCFormer achieves an average translation error of 3.3187 cm and a rotation error of 0.0469°, surpassing existing methods. Our approach significantly enhances scene representation and calibration precision, offering a robust solution for roadside sensor calibration and advancing the state of the art in sensor fusion technology.
Keywords: Calibration and identification; Computer vision for transportation; LiDAR-camera systems; Roadside traffic monitoring; Sensor fusion.
© 2024 The Authors.