Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 70% Match Research Paper Machine Learning Researchers,Deep Learning Engineers,Researchers focused on efficient AI 2 weeks ago

MetaCluster: Enabling Deep Compression of Kolmogorov-Arnold Network

generative-ai › vae
📄 Abstract

Abstract: Kolmogorov-Arnold Networks (KANs) replace scalar weights with per-edge vectors of basis coefficients, thereby boosting expressivity and accuracy but at the same time resulting in a multiplicative increase in parameters and memory. We propose MetaCluster, a framework that makes KANs highly compressible without sacrificing accuracy. Specifically, a lightweight meta-learner, trained jointly with the KAN, is used to map low-dimensional embedding to coefficient vectors, shaping them to lie on a low-dimensional manifold that is amenable to clustering. We then run K-means in coefficient space and replace per-edge vectors with shared centroids. Afterwards, the meta-learner can be discarded, and a brief fine-tuning of the centroid codebook recovers any residual accuracy loss. The resulting model stores only a small codebook and per-edge indices, exploiting the vector nature of KAN parameters to amortize storage across multiple coefficients. On MNIST, CIFAR-10, and CIFAR-100, across standard KANs and ConvKANs using multiple basis functions, MetaCluster achieves a reduction of up to 80$\times$ in parameter storage, with no loss in accuracy. Code will be released upon publication.
Authors (3)
Matthew Raffel
Adwaith Renjith
Lizhong Chen
Submitted
October 21, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

This paper proposes MetaCluster, a framework for highly compressing Kolmogorov-Arnold Networks (KANs) without sacrificing accuracy. It uses a meta-learner to map embeddings to a low-dimensional manifold for clustering coefficients, significantly reducing parameters by sharing centroids and using indices.

Business Value

Enables the deployment of highly expressive KAN models on resource-constrained devices or in applications requiring lower memory footprints, expanding their practical utility.