Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Physical symmetries provide a strong inductive bias for constructing
functions to analyze data. In particular, this bias may improve robustness,
data efficiency, and interpretability of machine learning models. However,
building machine learning models that explicitly respect symmetries can be
difficult due to the dedicated components required. Moreover, real-world
experiments may not exactly respect fundamental symmetries at the level of
finite granularities and energy thresholds. In this work, we explore an
alternative approach to create symmetry-aware machine learning models. We
introduce soft constraints that allow the model to decide the importance of
added symmetries during the learning process instead of enforcing exact
symmetries. We investigate two complementary approaches, one that penalizes the
model based on specific transformations of the inputs and one inspired by group
theory and infinitesimal transformations of the inputs. Using top quark jet
tagging and Lorentz equivariance as examples, we observe that the addition of
the soft constraints leads to more robust performance while requiring
negligible changes to current state-of-the-art models.
Key Contributions
SEAL (Symmetry EncourAging Loss) introduces a novel approach to build symmetry-aware machine learning models by using soft constraints instead of hard enforcement. This allows models to learn the importance of symmetries during training, improving robustness, data efficiency, and interpretability, demonstrated on top quark jet tagging.
Business Value
Enhances the reliability and efficiency of AI models used in scientific research and other domains where physical laws or known invariances are crucial, leading to more trustworthy and data-efficient AI.