Skip to main content
LLMs Special Issue

Comparing children and large language models in word sense disambiguation: Insights and challenges

Authors
  • Francesco Cabiddu (University College London, UK)
  • Mitja Nikolaus (CerCo, CNRS, France)
  • Abdellah Fourtassi (Aix-Marseille University, France)

Abstract

Understanding how children process ambiguous words is a challenge because sense disambiguation is a complex task that depends on both bottom-up and top-down cues. Here, we seek insight into this phenomenon by investigating how such a competence might arise in large distributional learners (Transformers) that purport to acquire sense representations from language input in a largely unsupervised fashion. We investigated how sense disambiguation might be achieved using model representations derived from naturalistic child-directed speech. We tested a large pool of Transformer models, varying in their pretraining input size/nature as well as the size of their parameter space. Tested across three behavioral experiments from the developmental literature, we found that these models capture some essential properties of child sense disambiguation, although most still struggle in the more challenging tasks with contrastive cues. We discuss implications for both theories of word learning and for using Transformers to capture child language processing.

Keywords: Child Word Sense Disambiguation, Transformers, Usage-Based Learning

Downloads:
Download PDF
View PDF

Published on
2024-08-28

Peer Reviewed