Publication Details
Overview
 
 
Zakaria Lemhaouri
 

Thesis

Abstract 

Recent advances in natural language processing—especially large-scale transformer models—have dramatically improved language generation and understanding. These systems, however, are not designed to explain how infants acquire language: they learn from massive static datasets, do not exhibit infant-like developmental trajectories, and overlook the cognitive, social, and psychological precursors that shape language learning.We propose a cognitive model for a robot{\textquoteright}s early language acquisition, inspired by how human babies learn language. The robot{\textquoteright}s learning process relies on sensorimotor development, social interactions with a caregiver, and real time learning, making it an active learner. This modular architecture includes three components: a motivational module, a perception module, and a communication/action module. The robot employs two associative learning methods to form meaningful symbols and acquire words with functional meaning: first, through trial and error to learn the correct word for each situation, and second, by using a neural network to ground each word in the goal it achieves, as well as in proprioceptive and exteroceptive signals. This yields a dual word–referent association.We implemented this architecture in a humanoid robot to study the development of its communicative skills. We aimed to follow major milestones in language learning, such as babbling, lexical development (learning nouns and verbs), syntax and early grammar development. The results indicate that the robot successfully acquired motivation-grounded language.Developmentally plausible models, such as the one we present here, can be valuable tools for investigating questions related to cognitive and psychological development in humans. We conducted several experiments to study how extralinguistic factors like motivation, sensory-motor development, caregiver responsiveness and social interaction influence the emergence and shaping of language. The results show that these factors, which are often overlooked by most language models, promote more efficient and faster language learning, a richer vocabulary, better retention of acquired words, and improved categorization learning.

Reference