Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Supervised dimensionality reduction maps labeled data into a low-dimensional
feature space while preserving class discriminability. A common approach is to
maximize a statistical measure of dissimilarity between classes in the feature
space. Information geometry provides an alternative framework for measuring
class dissimilarity, with the potential for improved insights and novel
applications. Information geometry, which is grounded in Riemannian geometry,
uses the Fisher information metric, a local measure of discriminability that
induces the Fisher-Rao distance. Here, we present Supervised Quadratic Feature
Analysis (SQFA), a linear dimensionality reduction method that maximizes
Fisher-Rao distances between class-conditional distributions, under Gaussian
assumptions. We motivate the Fisher-Rao distance as a good proxy for
discriminability. We show that SQFA features support good classification
performance with Quadratic Discriminant Analysis (QDA) on three real-world
datasets. SQFA provides a novel framework for supervised dimensionality
reduction, motivating future research in applying information geometry to
machine learning and neuroscience.
Authors (2)
Daniel Herrera-Esposito
Johannes Burge
Submitted
January 31, 2025
Key Contributions
Introduces Supervised Quadratic Feature Analysis (SQFA), a novel linear dimensionality reduction method that leverages information geometry to maximize Fisher-Rao distances between class-conditional distributions. This approach offers a new perspective on feature extraction by using a principled measure of dissimilarity derived from Riemannian geometry.
Business Value
Can lead to more effective feature representations for classification tasks, improving the performance of downstream applications in areas like image recognition and data analysis.