Abstract
Intelligent systems need to understand and respond to human words to enable them to interact with humans in a natural way. Several studies attempted to realize these abilities by investigating the symbol grounding problem. For example, we proposed multilayered multimodal latent Dirichlet allocation (mMLDA) to enable the formation of various concepts and inference using grounded concepts. We previously reported on the issue of connecting words to various hierarchical concepts and also proposed a simple preliminary algorithm for generating sentences. This paper proposes a novel method that enables a sensing system to verbalize an everyday scene it observes. The method uses mMLDA and Bayesian hidden Markov models (BHMM) and the proposed algorithm improves the word inference of our previous work. The advantage of our approach is that grammar learning based on BHMM not only boosts concept selection results but enables our method to process functional words. The proposed verbalization algorithm produces results that are far superior to those of previous methods. Finally, we developed a system to obtain multimodal data from human everyday activities. We evaluate language learning and sentence generation as a complete process under this realistic setting. The results demonstrate the effectiveness of our method.
Original language | English |
---|---|
Pages (from-to) | 806-824 |
Number of pages | 19 |
Journal | Advanced Robotics |
Volume | 30 |
Issue number | 11-12 |
DOIs | |
Publication status | Published - 17 Jun 2016 |
Externally published | Yes |
Keywords
- Language acquisition
- Multimodal categorization
- Sentence generation
- Symbol grounding
- Unsupervised learning