Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 85% Match Research Paper Theoretical ML Researchers,Mathematicians,Students of Deep Learning Theory 20 hours ago

The stability of shallow neural networks on spheres: A sharp spectral analysis

ai-safety › interpretability
📄 Abstract

Abstract: We present an estimation of the condition numbers of the \emph{mass} and \emph{stiffness} matrices arising from shallow ReLU$^k$ neural networks defined on the unit sphere~$\mathbb{S}^d$. In particular, when $\{\theta_j^*\}_{j=1}^n \subset \mathbb{S}^d$ is \emph{antipodally quasi-uniform}, the condition number is sharp. Indeed, in this case, we obtain sharp asymptotic estimates for the full spectrum of eigenvalues and characterize the structure of the corresponding eigenspaces, showing that the smallest eigenvalues are associated with an eigenbasis of low-degree polynomials while the largest eigenvalues are linked to high-degree polynomials. This spectral analysis establishes a precise correspondence between the approximation power of the network and its numerical stability.

Key Contributions

This paper provides a sharp spectral analysis of shallow ReLU^k neural networks defined on spheres, estimating condition numbers of mass and stiffness matrices. It establishes a precise correspondence between approximation power and numerical stability, showing how eigenvalues relate to polynomial degrees, particularly for antipodally quasi-uniform data.

Business Value

Contributes to the fundamental understanding of neural network behavior, which can lead to the design of more stable and reliable models, especially for applications involving spherical or manifold data.