Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Large language models (LLMs) are increasingly used to simulate humans, with
applications ranging from survey prediction to decision-making. However, are
LLMs strictly necessary, or can smaller, domain-grounded models suffice? We
identify a large class of simulation problems in which individuals make choices
among discrete options, where a graph neural network (GNN) can match or surpass
strong LLM baselines despite being three orders of magnitude smaller. We
introduce Graph-basEd Models for human Simulation (GEMS), which casts discrete
choice simulation tasks as a link prediction problem on graphs, leveraging
relational knowledge while incorporating language representations only when
needed. Evaluations across three key settings on three simulation datasets show
that GEMS achieves comparable or better accuracy than LLMs, with far greater
efficiency, interpretability, and transparency, highlighting the promise of
graph-based modeling as a lightweight alternative to LLMs for human simulation.
Our code is available at https://github.com/schang-lab/gems.
Key Contributions
This paper proposes GEMS (Graph-basEd Models for human Simulation), a Graph Neural Network-based framework that matches or surpasses LLM performance in simulating human discrete choices, while being three orders of magnitude smaller. By framing simulation tasks as link prediction problems on graphs, GEMS leverages relational knowledge effectively, offering greater efficiency, interpretability, and transparency compared to LLMs for specific simulation problems.
Business Value
Enables cost-effective and understandable simulation of human behavior for applications like market analysis, policy testing, and personalized user experiences, without the high computational cost of LLMs.