Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 92% Match Research Paper Machine learning theorists,Researchers in deep learning interpretability,Students of advanced ML topics 20 hours ago

Neural Collapse in Cumulative Link Models for Ordinal Regression: An Analysis with Unconstrained Feature Model

ai-safety › interpretability
📄 Abstract

Abstract: A phenomenon known as ''Neural Collapse (NC)'' in deep classification tasks, in which the penultimate-layer features and the final classifiers exhibit an extremely simple geometric structure, has recently attracted considerable attention, with the expectation that it can deepen our understanding of how deep neural networks behave. The Unconstrained Feature Model (UFM) has been proposed to explain NC theoretically, and there emerges a growing body of work that extends NC to tasks other than classification and leverages it for practical applications. In this study, we investigate whether a similar phenomenon arises in deep Ordinal Regression (OR) tasks, via combining the cumulative link model for OR and UFM. We show that a phenomenon we call Ordinal Neural Collapse (ONC) indeed emerges and is characterized by the following three properties: (ONC1) all optimal features in the same class collapse to their within-class mean when regularization is applied; (ONC2) these class means align with the classifier, meaning that they collapse onto a one-dimensional subspace; (ONC3) the optimal latent variables (corresponding to logits or preactivations in classification tasks) are aligned according to the class order, and in particular, in the zero-regularization limit, a highly local and simple geometric relationship emerges between the latent variables and the threshold values. We prove these properties analytically within the UFM framework with fixed threshold values and corroborate them empirically across a variety of datasets. We also discuss how these insights can be leveraged in OR, highlighting the use of fixed thresholds.

Key Contributions

This paper investigates the phenomenon of Neural Collapse (NC) in the context of deep Ordinal Regression (OR) tasks by combining cumulative link models with the Unconstrained Feature Model (UFM). It demonstrates that a similar phenomenon, termed Ordinal Neural Collapse (ONC), emerges, characterized by feature collapse to class means and specific geometric properties of these means. This work extends the understanding of NC beyond standard classification.

Business Value

Deeper understanding of how deep models learn can lead to more robust, interpretable, and reliable AI systems, particularly in applications where ordered categories are important (e.g., medical severity grading).