Projects per year
Abstract
This paper presents a multi-modal system for finding out where to direct the attention of a social robot in a dialog scenario, which is robust against environmental sounds (door slamming, phone ringing etc.) and short speech segments. The method is based on combining voice activity detection (VAD) and sound source localization (SSL) and furthermore apply post-processing to SSL to filter out short sounds. The system is tested against a baseline system in four different real-world experiments, where different sounds are used as interfering sounds. The results are promising and show a clear improvement.
Original language | English |
---|---|
Title of host publication | Multimodal Analyses enabling Artificial Agents in Human-Machine Interaction |
Number of pages | 10 |
Publisher | Springer Publishing Company |
Publication date | 2015 |
Pages | 25-34 |
ISBN (Print) | 978-3-319-15556-2 |
ISBN (Electronic) | 978-3-319-15557-9 |
DOIs | |
Publication status | Published - 2015 |
Event | 2014 2nd Multimodal Analyses Enabling Artificial Agents in Human-Machine Interaction - Singapore, Singapore Duration: 14 Sept 2014 → 14 Sept 2014 |
Conference
Conference | 2014 2nd Multimodal Analyses Enabling Artificial Agents in Human-Machine Interaction |
---|---|
Country/Territory | Singapore |
City | Singapore |
Period | 14/09/2014 → 14/09/2014 |
Series | Lecture Notes in Computer Science |
---|---|
ISSN | 0302-9743 |
Keywords
- Multi-modal tracking
- human-computer interaction
- sound-source localization
Fingerprint
Dive into the research topics of 'Improving Robustness against Environmental Sounds for Directing Attention of Social Robots'. Together they form a unique fingerprint.Projects
- 1 Finished
-
iSocioBot: Durable Interaction with Socially Intelligent Robots
Tan, Z., Jensen, S. H., Lindberg, B. & Thomsen, N. B.
01/08/2013 → 31/12/2017
Project: Research