Testing Theory of Mind in GPT Models and Humans

James Strachan, Dalila Albergo, Giulia Borghini, Oriana Pansardi,Eugenio Scaliti, Alessandro Rufo, G. Manzi,Michael S. A. Graziano,Cristina Becchio

Research Square (Research Square)(2023)

引用 1|浏览3
暂无评分
摘要
Abstract Interacting with other people involves reasoning about and prediction of others' mental states, or Theory of Mind. This capacity is a distinguishing feature of human cognition but recent advances in Large Language Models (LLMs) such as ChatGPT suggest that they may possess some emergent capacity for human-like Theory of Mind. Such claims merit a systematic approach to explore the limits of GPT models' emergent Theory of Mind capacity and compare it against humans. We show that while GPT models show impressive Theory of Mind-like capacity in controlled tests, there are key deviations from human performance that call into question how human-like this capacity is. Specifically, across a battery of Theory of Mind tests, we found that GPT models performed at human levels when recognising indirect requests, false beliefs, and higher-order mental states like misdirection, but were specifically impaired at recognising faux pas. Follow-up studies revealed that this was due to GPT's conservatism in drawing conclusions that humans took to be self-evident. Our results suggest that while GPT may demonstrate the competence for sophisticated mentalistic inference, its lack of embodiment within an action-oriented environment make this capacity qualitatively different from human cognition.
更多
查看译文
关键词
gpt models,mind,testing,theory
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要