Dr Yanchao Yu Y.Yu@napier.ac.uk
Lecturer
We present a multi-modal dialogue system for interactive learning of perceptually grounded word meanings from a human tutor. The system integrates an incremental, semantic parsing/generation framework - Dynamic Syntax and Type Theory with Records (DS-TTR) - with a set of visual classifiers that are learned throughout the interaction and which ground the meaning representations that it produces. We use this system in interaction with a simulated human tutor to study the effect of different dialogue policies and capabilities on accuracy of learned meanings, learning rates, and efforts/costs to the tutor. We show that the overall performance of the learning agent is affected by (1) who takes initiative in the dialogues; (2) the ability to express/use their confidence level about visual attributes; and (3) the ability to process elliptical as well as incrementally constructed dialogue turns.
Yu, Y., Eshghi, A., & Lemon, O. (2016, August). Interactively learning visually grounded word meanings from a human tutor. Presented at 5th Workshop on Vision and Language, Berlin, Germany
Presentation Conference Type | Conference Paper (published) |
---|---|
Conference Name | 5th Workshop on Vision and Language |
Start Date | Aug 12, 2016 |
Publication Date | 2016 |
Deposit Date | Jun 28, 2023 |
Publicly Available Date | Jun 28, 2023 |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 48-53 |
Book Title | Proceedings of the 5th Workshop on Vision and Language |
Publisher URL | https://aclanthology.org/W16-3206/ |
Interactively learning visually grounded word meanings from a human tutor
(3.2 Mb)
PDF
Publisher Licence URL
http://creativecommons.org/licenses/by/4.0/
How Much do Robots Understand Rudeness? Challenges in Human-Robot Interaction
(2024)
Presentation / Conference Contribution
TaskMaster: A Novel Cross-platform Task-based Spoken Dialogue System for Human-Robot Interaction
(2023)
Presentation / Conference Contribution
MoDEsT: a Modular Dialogue Experiments and Demonstration Toolkit
(2023)
Presentation / Conference Contribution
A Visually-Aware Conversational Robot Receptionist
(2022)
Presentation / Conference Contribution
The CRECIL Corpus: a New Dataset for Extraction of Relations between Characters in Chinese Multi-party Dialogues
(2022)
Presentation / Conference Contribution
About Edinburgh Napier Research Repository
Administrator e-mail: repository@napier.ac.uk
This application uses the following open-source libraries:
Apache License Version 2.0 (http://www.apache.org/licenses/)
Apache License Version 2.0 (http://www.apache.org/licenses/)
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search