Predicting Effects, Missing Distributions: Evaluating LLMs as Human Behavior Simulators in Operations Management
Nov 25, 2025·

·
0 min read
Runze Zhang
Xiaowei Zhang
Mingyang Zhao

Abstract
LLMs are emerging tools for simulating human behavior in business, economics, and social science, offering a lower‑cost complement to laboratory experiments, field studies, and surveys. This paper evaluates how well LLMs replicate human behavior in operations management. Using nine published experiments in behavioral operations, we assess two criteria: replication of hypothesis‑test outcomes and distributional alignment via Wasserstein distance. LLMs reproduce most hypothesis‑level effects, capturing key decision biases, but their response distributions diverge from human data, including for strong commercial models. We also test two lightweight strategies—Chain‑of‑Thought prompting and hyperparameter tuning—which reduce misalignment and can sometimes let smaller or open‑source models match or surpass larger systems.
Type
Publication
INFORMS Journal on Data Science, Major Revision

Authors
PhD student in the Department of Industrial Engineering and Decision Analytics, HKUST.

Authors
I am an Associate Professor at HKUST, jointly appointed in the Department of Industrial Engineering and Decision Analytics and the Department of Economics, and the Academic Director of the MSc in FinTech program. I serve as an Associate Editor for several leading journals in the field, including Management Science, Operations Research, Navel Research Logistics, and Queueing Systems.

Authors
MPhil student in the Department of Industrial Engineering and Decision Analytics, HKUST.