Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: We propose FutrTrack, a modular camera-LiDAR multi-object tracking framework
that builds on existing 3D detectors by introducing a transformer-based
smoother and a fusion-driven tracker. Inspired by query-based tracking
frameworks, FutrTrack employs a multimodal two-stage transformer refinement and
tracking pipeline. Our fusion tracker integrates bounding boxes with multimodal
bird's-eye-view (BEV) fusion features from multiple cameras and LiDAR without
the need for an explicit motion model. The tracker assigns and propagates
identities across frames, leveraging both geometric and semantic cues for
robust re-identification under occlusion and viewpoint changes. Prior to
tracking, we refine sequences of bounding boxes with a temporal smoother over a
moving window to refine trajectories, reduce jitter, and improve spatial
consistency. Evaluated on nuScenes and KITTI, FutrTrack demonstrates that
query-based transformer tracking methods benefit significantly from multimodal
sensor features compared with previous single-sensor approaches. With an aMOTA
of 74.7 on the nuScenes test set, FutrTrack achieves strong performance on 3D
MOT benchmarks, reducing identity switches while maintaining competitive
accuracy. Our approach provides an efficient framework for improving
transformer-based trackers to compete with other neural-network-based methods
even with limited data and without pretraining.
Authors (3)
Martha Teiko Teye
Ori Maoz
Matthias Rottmann
Submitted
October 22, 2025
Key Contributions
FutrTrack is a modular camera-LiDAR fusion framework for 3D Multiple Object Tracking (MOT) that uses a transformer-based smoother and a fusion-driven tracker. It employs a multimodal two-stage transformer pipeline, integrating BEV fusion features without an explicit motion model, and leverages geometric/semantic cues for robust re-identification, significantly improving tracking performance.
Business Value
Crucial for the development of safe and reliable autonomous driving systems. Accurate 3D object tracking is essential for perception, prediction, and planning modules, enabling vehicles to navigate complex environments safely.