Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Text-to-audio models are a type of generative model that produces audio
output in response to a given textual prompt. Although level generators and the
properties of the functional content that they create (e.g., playability)
dominate most discourse in procedurally generated content (PCG), games that
emotionally resonate with players tend to weave together a range of creative
and multimodal content (e.g., music, sounds, visuals, narrative tone), and
multimodal models have begun seeing at least experimental use for this purpose.
However, it remains unclear what exactly such models generate, and with what
degree of variability and fidelity: audio is an extremely broad class of output
for a generative system to target.
Within the PCG community, expressive range analysis (ERA) has been used as a
quantitative way to characterize generators' output space, especially for level
generators. This paper adapts ERA to text-to-audio models, making the analysis
tractable by looking at the expressive range of outputs for specific, fixed
prompts. Experiments are conducted by prompting the models with several
standardized prompts derived from the Environmental Sound Classification
(ESC-50) dataset. The resulting audio is analyzed along key acoustic dimensions
(e.g., pitch, loudness, and timbre). More broadly, this paper offers a
framework for ERA-based exploratory evaluation of generative audio models.
Authors (6)
Jonathan Morse
Azadeh Naderi
Swen Gaudl
Mark Cartwright
Amy K. Hoover
Mark J. Nelson
Submitted
October 31, 2025
Key Contributions
Adapts Expressive Range Analysis (ERA) from level generators to text-to-audio models, providing a quantitative method to characterize their output space. This allows for a better understanding of the variability and fidelity of generated audio, addressing the ambiguity in current text-to-audio capabilities.
Business Value
Enables developers and researchers to better understand and control the output of text-to-audio models, leading to more predictable and higher-quality audio content for various applications.