Abstract
As large language models (LLMs) like GPT become increasingly prevalent, it is essential that we assess their capabilities beyond language processing. This paper examines the economic rationality of GPT by instructing it to make budgetary decisions in four domains: risk, time, social, and food preferences. We measure economic rationality by assessing the consistency of GPT’s decisions with utility maximization in classic revealed preference theory. We find that GPT’s decisions are largely rational in each domain and demonstrate higher rationality score than those of human subjects in a parallel experiment and in the literature. Moreover, the estimated preference parameters of GPT are slightly different from human subjects and exhibit a lower degree of heterogeneity. We also find that the rationality scores are robust to the degree of randomness and demographic settings such as age and gender but are sensitive to contexts based on the language frames of the choice situations. These results suggest the potential of LLMs to make good decisions and the need to further understand their capabilities, limitations, and underlying mechanisms.
Original language | English |
---|---|
Article number | e2316205120 |
Journal | Proceedings of the National Academy of Sciences of the United States of America |
Volume | 120 |
Issue number | 51 |
DOIs | |
Publication status | Published - 12 Dec 2023 |
Bibliographical note
Publisher Copyright:© 2023 the Author(s).
Funding
T.X.L. gratefully acknowledges financial support from National Natural Science Foundation of China (72222005) and Tsinghua University (No. 2022Z04W01032). S.Z. gratefully acknowledges financial support from the Singapore Ministry of Education (Academic Research Fund Tier 1) and the National University of Singapore (Dean’s Chair Grant).
Keywords
- decision-making
- economic rationality
- large language models
- revealed preference analysis