iSocioBot; A Multimodal Interactive Social Robot

Zheng-Hua Tan, Nicolai Bæk Thomsen, Xiaodong Duan, Evgenios Vlachos, Sven Ewan Shepstone, Morten Højfeldt Rasmussen, Jesper Lisby Højvang

Research output: Contribution to journalJournal articleResearchpeer-review

5 Citations (Scopus)

Abstract

We present one way of constructing a social robot, such that it is able to interact with humans using multiple modalities. The robotic system is able to direct attention towards the dominant speaker using sound source localization and face detection, it is capable of identifying persons using face recognition and speaker identification and the system is able to communicate and engage in a dialog with humans by using speech recognition, speech synthesis and different facial expressions. The software is built upon the open-source robot operating system framework and our software is made publicly available. Furthermore, the electrical parts (sensors, laptop, base platform, etc.) are standard components, thus allowing for replicating the system. The design of the robot is unique and we justify why this design is suitable for our robot and the intended use. By making software, hardware and design accessible to everyone, we make research in social robotics available to a broader audience. To evaluate the properties and the appearance of the robot we invited users to interact with it in pairs (active interaction partner/observer) and collected their responses via an extended version of the Godspeed Questionnaire. Results suggest an overall positive impression of the robot and interaction experience, as well as significant differences in responses based on type of interaction and gender.

Original languageEnglish
JournalInternational Journal of Social Robotics
Volume10
Issue number1
Pages (from-to)5-19
Number of pages15
ISSN1875-4791
DOIs
Publication statusPublished - 1 Jan 2018

Fingerprint

Robots
Face recognition
Robotics
Speech synthesis
Speech recognition
Acoustic waves
Hardware
Sensors

Keywords

  • Human robot interaction
  • Image processing
  • Social robot
  • Speech processing

Cite this

Tan, Z-H., Thomsen, N. B., Duan, X., Vlachos, E., Shepstone, S. E., Rasmussen, M. H., & Højvang, J. L. (2018). iSocioBot; A Multimodal Interactive Social Robot. International Journal of Social Robotics, 10(1), 5-19. https://doi.org/10.1007/s12369-017-0426-7
Tan, Zheng-Hua ; Thomsen, Nicolai Bæk ; Duan, Xiaodong ; Vlachos, Evgenios ; Shepstone, Sven Ewan ; Rasmussen, Morten Højfeldt ; Højvang, Jesper Lisby. / iSocioBot; A Multimodal Interactive Social Robot. In: International Journal of Social Robotics. 2018 ; Vol. 10, No. 1. pp. 5-19.
@article{d06949a3f859421dacca0d542ed1f615,
title = "iSocioBot;: A Multimodal Interactive Social Robot",
abstract = "We present one way of constructing a social robot, such that it is able to interact with humans using multiple modalities. The robotic system is able to direct attention towards the dominant speaker using sound source localization and face detection, it is capable of identifying persons using face recognition and speaker identification and the system is able to communicate and engage in a dialog with humans by using speech recognition, speech synthesis and different facial expressions. The software is built upon the open-source robot operating system framework and our software is made publicly available. Furthermore, the electrical parts (sensors, laptop, base platform, etc.) are standard components, thus allowing for replicating the system. The design of the robot is unique and we justify why this design is suitable for our robot and the intended use. By making software, hardware and design accessible to everyone, we make research in social robotics available to a broader audience. To evaluate the properties and the appearance of the robot we invited users to interact with it in pairs (active interaction partner/observer) and collected their responses via an extended version of the Godspeed Questionnaire. Results suggest an overall positive impression of the robot and interaction experience, as well as significant differences in responses based on type of interaction and gender.",
keywords = "Human robot interaction, Image processing, Social robot, Speech processing",
author = "Zheng-Hua Tan and Thomsen, {Nicolai B{\ae}k} and Xiaodong Duan and Evgenios Vlachos and Shepstone, {Sven Ewan} and Rasmussen, {Morten H{\o}jfeldt} and H{\o}jvang, {Jesper Lisby}",
year = "2018",
month = "1",
day = "1",
doi = "10.1007/s12369-017-0426-7",
language = "English",
volume = "10",
pages = "5--19",
journal = "International Journal of Social Robotics",
issn = "1875-4791",
publisher = "Physica-Verlag",
number = "1",

}

Tan, Z-H, Thomsen, NB, Duan, X, Vlachos, E, Shepstone, SE, Rasmussen, MH & Højvang, JL 2018, 'iSocioBot; A Multimodal Interactive Social Robot', International Journal of Social Robotics, vol. 10, no. 1, pp. 5-19. https://doi.org/10.1007/s12369-017-0426-7

iSocioBot; A Multimodal Interactive Social Robot. / Tan, Zheng-Hua; Thomsen, Nicolai Bæk; Duan, Xiaodong; Vlachos, Evgenios; Shepstone, Sven Ewan; Rasmussen, Morten Højfeldt; Højvang, Jesper Lisby.

In: International Journal of Social Robotics, Vol. 10, No. 1, 01.01.2018, p. 5-19.

Research output: Contribution to journalJournal articleResearchpeer-review

TY - JOUR

T1 - iSocioBot;

T2 - A Multimodal Interactive Social Robot

AU - Tan, Zheng-Hua

AU - Thomsen, Nicolai Bæk

AU - Duan, Xiaodong

AU - Vlachos, Evgenios

AU - Shepstone, Sven Ewan

AU - Rasmussen, Morten Højfeldt

AU - Højvang, Jesper Lisby

PY - 2018/1/1

Y1 - 2018/1/1

N2 - We present one way of constructing a social robot, such that it is able to interact with humans using multiple modalities. The robotic system is able to direct attention towards the dominant speaker using sound source localization and face detection, it is capable of identifying persons using face recognition and speaker identification and the system is able to communicate and engage in a dialog with humans by using speech recognition, speech synthesis and different facial expressions. The software is built upon the open-source robot operating system framework and our software is made publicly available. Furthermore, the electrical parts (sensors, laptop, base platform, etc.) are standard components, thus allowing for replicating the system. The design of the robot is unique and we justify why this design is suitable for our robot and the intended use. By making software, hardware and design accessible to everyone, we make research in social robotics available to a broader audience. To evaluate the properties and the appearance of the robot we invited users to interact with it in pairs (active interaction partner/observer) and collected their responses via an extended version of the Godspeed Questionnaire. Results suggest an overall positive impression of the robot and interaction experience, as well as significant differences in responses based on type of interaction and gender.

AB - We present one way of constructing a social robot, such that it is able to interact with humans using multiple modalities. The robotic system is able to direct attention towards the dominant speaker using sound source localization and face detection, it is capable of identifying persons using face recognition and speaker identification and the system is able to communicate and engage in a dialog with humans by using speech recognition, speech synthesis and different facial expressions. The software is built upon the open-source robot operating system framework and our software is made publicly available. Furthermore, the electrical parts (sensors, laptop, base platform, etc.) are standard components, thus allowing for replicating the system. The design of the robot is unique and we justify why this design is suitable for our robot and the intended use. By making software, hardware and design accessible to everyone, we make research in social robotics available to a broader audience. To evaluate the properties and the appearance of the robot we invited users to interact with it in pairs (active interaction partner/observer) and collected their responses via an extended version of the Godspeed Questionnaire. Results suggest an overall positive impression of the robot and interaction experience, as well as significant differences in responses based on type of interaction and gender.

KW - Human robot interaction

KW - Image processing

KW - Social robot

KW - Speech processing

UR - http://www.scopus.com/inward/record.url?scp=85040838181&partnerID=8YFLogxK

U2 - 10.1007/s12369-017-0426-7

DO - 10.1007/s12369-017-0426-7

M3 - Journal article

VL - 10

SP - 5

EP - 19

JO - International Journal of Social Robotics

JF - International Journal of Social Robotics

SN - 1875-4791

IS - 1

ER -

Tan Z-H, Thomsen NB, Duan X, Vlachos E, Shepstone SE, Rasmussen MH et al. iSocioBot; A Multimodal Interactive Social Robot. International Journal of Social Robotics. 2018 Jan 1;10(1):5-19. https://doi.org/10.1007/s12369-017-0426-7