Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: A fundamental challenge in probabilistic modeling is to balance expressivity
and inference efficiency. Tractable probabilistic models (TPMs) aim to directly
address this tradeoff by imposing constraints that guarantee efficient
inference of certain queries while maintaining expressivity. In particular,
probabilistic circuits (PCs) provide a unifying framework for many TPMs, by
characterizing families of models as circuits satisfying different structural
properties. Because the complexity of inference on PCs is a function of the
circuit size, understanding the size requirements of different families of PCs
is fundamental in mapping the trade-off between tractability and expressive
efficiency. However, the study of expressive efficiency of circuits are often
concerned with exact representations, which may not align with model learning,
where we look to approximate the underlying data distribution closely by some
distance measure. Moreover, due to hardness of inference tasks, exactly
representing distributions while supporting tractable inference often incurs
exponential size blow-ups. In this paper, we consider a natural, yet so far
underexplored, question: can we avoid such size blow-up by allowing for some
small approximation error? We study approximating distributions with
probabilistic circuits with guarantees based on $f$-divergences, and analyze
which inference queries remain well-approximated under this framework. We show
that approximating an arbitrary distribution with bounded $f$-divergence is
$\mathsf{NP}$-hard for any model that can tractably compute marginals. In
addition, we prove an exponential size gap for approximation between the class
of decomposable PCs and that of decomposable and deterministic PCs.
Key Contributions
This paper investigates the fundamental challenge of balancing expressivity and inference efficiency in probabilistic modeling, particularly for Tractable Probabilistic Models (TPMs) like Probabilistic Circuits (PCs). It analyzes the hardness of approximating distributions and the circuit size requirements for different families of PCs, mapping the trade-offs between tractability and expressive efficiency.
Business Value
Contributes to the theoretical understanding of probabilistic models, which can inform the design of more efficient and powerful AI systems for various applications.