Skip to main content

Modelling Multiple Language Learning in a Developmental Cognitive Architecture

Giorgi, I., Golosio, B., Esposito, M., Cangelosi, A., Masala, Giovanni Luca (2020) Modelling Multiple Language Learning in a Developmental Cognitive Architecture. IEEE Transactions on Cognitive and Developmental Systems, . ISSN 2379-8920. E-ISSN 2379-8939. (doi:10.1109/TCDS.2020.3033963) (The full text of this publication is not currently available from this repository. You may be able to access a copy if URLs are provided) (KAR id:91361)

The full text of this publication is not currently available from this repository. You may be able to access a copy if URLs are provided. (Contact us about this Publication)
Official URL
http://dx.doi.org/10.1109/TCDS.2020.3033963

Abstract

In this work, we model multiple natural language learning in a developmental neuroscience-inspired architecture. The ANNABELL model (Artificial Neural Network with Adaptive Behaviour Exploited for Language Learning), is a large-scale neural network, however, unlike most deep learning methods that solve natural language processing (NLP) tasks, it does not represent an empirical engineering solution for specific NLP problems; rather, its organisation complies with findings from cognitive neuroscience, particularly the multi-compartment working memory models. The system is appropriately trained to understand the level of cognitive development required for language acquisition and the robustness achieved in learning simultaneously four languages, using a corpus of text-based exchanges of developmental complexity. The selected languages, Greek, Italian and Albanian, besides English, differ significantly in structure and complexity. Initially, the system was validated in each language alone and was then compared with the open-ended cumulative training, in which languages are learned jointly, prior to querying with random language at random order. We aimed to assess if the model could learn the languages together to the same degree of skill as learning each apart. Moreover, we explored the generalisation skill in multilingual context questions and the ability to elaborate a short text of preschool literature. We verified if the system could follow a dialogue coherently and cohesively, keeping track of its previous answers and recalling them in subsequent queries. The results show that the architecture developed broad language processing functionalities, with satisfactory performances in each language trained singularly, maintaining high accuracies when they are acquired cumulatively.

Item Type: Article
DOI/Identification number: 10.1109/TCDS.2020.3033963
Additional information: cited By 1
Uncontrolled keywords: Computer architecture; Computational modeling; Brain modeling; Deep learning; Natural language processing; Task analysis; Neural network; cognitive system; natural language understanding; multilingual system
Subjects: Q Science > QA Mathematics (inc Computing science) > QA 75 Electronic computers. Computer science
Divisions: Divisions > Division of Computing, Engineering and Mathematical Sciences > School of Computing
Depositing User: Amy Boaler
Date Deposited: 05 Nov 2021 11:58 UTC
Last Modified: 08 Nov 2021 10:26 UTC
Resource URI: https://kar.kent.ac.uk/id/eprint/91361 (The current URI for this page, for reference purposes)
Masala, Giovanni Luca: https://orcid.org/0000-0001-6734-9424
  • Depositors only (login required):