Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 85% Match Research Paper ML Theorists,Researchers in Probabilistic AI,Students of Machine Learning 1 week ago

On the Hardness of Approximating Distributions with Tractable Probabilistic Models

generative-ai › flow-models
📄 Abstract

Abstract: A fundamental challenge in probabilistic modeling is to balance expressivity and inference efficiency. Tractable probabilistic models (TPMs) aim to directly address this tradeoff by imposing constraints that guarantee efficient inference of certain queries while maintaining expressivity. In particular, probabilistic circuits (PCs) provide a unifying framework for many TPMs, by characterizing families of models as circuits satisfying different structural properties. Because the complexity of inference on PCs is a function of the circuit size, understanding the size requirements of different families of PCs is fundamental in mapping the trade-off between tractability and expressive efficiency. However, the study of expressive efficiency of circuits are often concerned with exact representations, which may not align with model learning, where we look to approximate the underlying data distribution closely by some distance measure. Moreover, due to hardness of inference tasks, exactly representing distributions while supporting tractable inference often incurs exponential size blow-ups. In this paper, we consider a natural, yet so far underexplored, question: can we avoid such size blow-up by allowing for some small approximation error? We study approximating distributions with probabilistic circuits with guarantees based on $f$-divergences, and analyze which inference queries remain well-approximated under this framework. We show that approximating an arbitrary distribution with bounded $f$-divergence is $\mathsf{NP}$-hard for any model that can tractably compute marginals. In addition, we prove an exponential size gap for approximation between the class of decomposable PCs and that of decomposable and deterministic PCs.
Authors (2)
John Leland
YooJung Choi
Submitted
June 2, 2025
arXiv Category
cs.AI
arXiv PDF

Key Contributions

This paper investigates the fundamental challenge of balancing expressivity and inference efficiency in probabilistic modeling, particularly for Tractable Probabilistic Models (TPMs) like Probabilistic Circuits (PCs). It analyzes the hardness of approximating distributions and the circuit size requirements for different families of PCs, mapping the trade-offs between tractability and expressive efficiency.

Business Value

Contributes to the theoretical understanding of probabilistic models, which can inform the design of more efficient and powerful AI systems for various applications.