Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 90% Match Research Paper Autonomous driving researchers,Robotics engineers,Computer vision engineers,AI researchers in tracking and fusion 2 weeks ago

FutrTrack: A Camera-LiDAR Fusion Transformer for 3D Multiple Object Tracking

computer-vision › object-detection
📄 Abstract

Abstract: We propose FutrTrack, a modular camera-LiDAR multi-object tracking framework that builds on existing 3D detectors by introducing a transformer-based smoother and a fusion-driven tracker. Inspired by query-based tracking frameworks, FutrTrack employs a multimodal two-stage transformer refinement and tracking pipeline. Our fusion tracker integrates bounding boxes with multimodal bird's-eye-view (BEV) fusion features from multiple cameras and LiDAR without the need for an explicit motion model. The tracker assigns and propagates identities across frames, leveraging both geometric and semantic cues for robust re-identification under occlusion and viewpoint changes. Prior to tracking, we refine sequences of bounding boxes with a temporal smoother over a moving window to refine trajectories, reduce jitter, and improve spatial consistency. Evaluated on nuScenes and KITTI, FutrTrack demonstrates that query-based transformer tracking methods benefit significantly from multimodal sensor features compared with previous single-sensor approaches. With an aMOTA of 74.7 on the nuScenes test set, FutrTrack achieves strong performance on 3D MOT benchmarks, reducing identity switches while maintaining competitive accuracy. Our approach provides an efficient framework for improving transformer-based trackers to compete with other neural-network-based methods even with limited data and without pretraining.
Authors (3)
Martha Teiko Teye
Ori Maoz
Matthias Rottmann
Submitted
October 22, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

FutrTrack is a modular camera-LiDAR fusion framework for 3D Multiple Object Tracking (MOT) that uses a transformer-based smoother and a fusion-driven tracker. It employs a multimodal two-stage transformer pipeline, integrating BEV fusion features without an explicit motion model, and leverages geometric/semantic cues for robust re-identification, significantly improving tracking performance.

Business Value

Crucial for the development of safe and reliable autonomous driving systems. Accurate 3D object tracking is essential for perception, prediction, and planning modules, enabling vehicles to navigate complex environments safely.