Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: We present the first sub-microsecond transformer implementation on an FPGA
achieving competitive performance for state-of-the-art high-energy physics
benchmarks. Transformers have shown exceptional performance on multiple tasks
in modern machine learning applications, including jet tagging at the CERN
Large Hadron Collider (LHC). However, their computational complexity prohibits
use in real-time applications, such as the hardware trigger system of the
collider experiments up until now. In this work, we demonstrate the first
application of transformers for jet tagging on FPGAs, achieving
$\mathcal{O}(100)$ nanosecond latency with superior performance compared to
alternative baseline models. We leverage high-granularity quantization and
distributed arithmetic optimization to fit the entire transformer model on a
single FPGA, achieving the required throughput and latency. Furthermore, we add
multi-head attention and linear attention support to hls4ml, making our work
accessible to the broader fast machine learning community. This work advances
the next-generation trigger systems for the High Luminosity LHC, enabling the
use of transformers for real-time applications in high-energy physics and
beyond.
Authors (10)
Lauri Laatu
Chang Sun
Arianna Cox
Abhijith Gandrakota
Benedikt Maier
Jennifer Ngadiuba
+4 more
Submitted
October 26, 2025
arXiv Category
physics.ins-det
Key Contributions
This paper presents the first sub-microsecond transformer implementation on an FPGA for jet tagging in high-energy physics. It achieves competitive performance and superior latency compared to baseline models by leveraging quantization and distributed arithmetic optimization, making real-time applications of transformers feasible.
Business Value
Enables real-time data analysis and decision-making in high-stakes scientific experiments, potentially leading to faster discoveries and more efficient data acquisition.