Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 90% Match Position Paper Cognitive Scientists,Linguists,AI Researchers,Psychologists 4 weeks ago

To model human linguistic prediction, make LLMs less superhuman

large-language-models › reasoning
📄 Abstract

Abstract: When people listen to or read a sentence, they actively make predictions about upcoming words: words that are less predictable are generally read more slowly than predictable ones. The success of large language models (LLMs), which, like humans, make predictions about upcoming words, has motivated exploring the use of these models as cognitive models of human linguistic prediction. Surprisingly, in the last few years, as language models have become better at predicting the next word, their ability to predict human reading behavior has declined. This is because LLMs are able to predict upcoming words much better than people can, leading them to predict lower processing difficulty in reading than observed in human experiments; in other words, mainstream LLMs are 'superhuman' as models of language comprehension. In this position paper, we argue that LLMs' superhumanness is primarily driven by two factors: compared to humans, LLMs have much stronger long-term memory for facts and training examples, and they have much better short-term memory for previous words in the text. We advocate for creating models that have human-like long-term and short-term memory, and outline some possible directions for achieving this goal. Finally, we argue that currently available human data is insufficient to measure progress towards this goal, and outline human experiments that can address this gap.

Key Contributions

This paper argues that current Large Language Models (LLMs) have become 'superhuman' predictors of human linguistic processing, leading to a decline in their ability to accurately model human reading behavior. It identifies stronger long-term memory and superior prediction capabilities as key factors contributing to this superhumanness, suggesting a need to adjust LLMs for better cognitive modeling.

Business Value

Understanding the cognitive processes behind language comprehension can lead to more intuitive and human-like AI interactions, improving user experience in applications that rely on natural language understanding.