Online perceptual learning and natural language acquisition for autonomous robots

作者:

摘要

In this work, the problem of bootstrapping knowledge in language and vision for autonomous robots is addressed through novel techniques in grammar induction and word grounding to the perceptual world. In particular, we demonstrate a system, called OLAV, which is able, for the first time, to (1) learn to form discrete concepts from sensory data; (2) ground language (n-grams) to these concepts; (3) induce a grammar for the language being used to describe the perceptual world; and moreover to do all this incrementally, without storing all previous data. The learning is achieved in a loosely-supervised manner from raw linguistic and visual data. Moreover, the learnt model is transparent, rather than a black-box model and is thus open to human inspection. The visual data is collected using three different robotic platforms deployed in real-world and simulated environments and equipped with different sensing modalities, while the linguistic data is collected using online crowdsourcing tools and volunteers. The analysis performed on these robots demonstrates the effectiveness of the framework in learning visual concepts, language groundings and grammatical structure in these three online settings.

论文关键词:Language and vision,Language acquisition,Language grounding,Grammar induction

论文评审过程:Received 31 October 2019, Revised 27 August 2021, Accepted 21 November 2021, Available online 26 November 2021, Version of Record 30 November 2021.

论文官网地址:https://doi.org/10.1016/j.artint.2021.103637