Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 92% Match Research Paper AI Researchers,Social Scientists,Data Scientists,Simulation Modelers,Economists 17 hours ago

Rethinking LLM Human Simulation: When a Graph is What You Need

graph-neural-networks › graph-learning
📄 Abstract

Abstract: Large language models (LLMs) are increasingly used to simulate humans, with applications ranging from survey prediction to decision-making. However, are LLMs strictly necessary, or can smaller, domain-grounded models suffice? We identify a large class of simulation problems in which individuals make choices among discrete options, where a graph neural network (GNN) can match or surpass strong LLM baselines despite being three orders of magnitude smaller. We introduce Graph-basEd Models for human Simulation (GEMS), which casts discrete choice simulation tasks as a link prediction problem on graphs, leveraging relational knowledge while incorporating language representations only when needed. Evaluations across three key settings on three simulation datasets show that GEMS achieves comparable or better accuracy than LLMs, with far greater efficiency, interpretability, and transparency, highlighting the promise of graph-based modeling as a lightweight alternative to LLMs for human simulation. Our code is available at https://github.com/schang-lab/gems.

Key Contributions

This paper proposes GEMS (Graph-basEd Models for human Simulation), a Graph Neural Network-based framework that matches or surpasses LLM performance in simulating human discrete choices, while being three orders of magnitude smaller. By framing simulation tasks as link prediction problems on graphs, GEMS leverages relational knowledge effectively, offering greater efficiency, interpretability, and transparency compared to LLMs for specific simulation problems.

Business Value

Enables cost-effective and understandable simulation of human behavior for applications like market analysis, policy testing, and personalized user experiences, without the high computational cost of LLMs.