Multimodal Desktop Interaction: The Face –Object-Gesture–Voice Example

Nikolas Vidakis, Anastasios Vlasopoulos, Tsampikos Kounalakis, Petros Varchalamas, Michalis Dimitriou, Gregory Kalliatakis, Efthimios Syntychakis, John Christofakis, Georgios Triantafyllidis

Research output: Contribution to book/anthology/report/conference proceedingArticle in proceedingResearchpeer-review

3 Citations (Scopus)
605 Downloads (Pure)

Abstract

This paper presents a natural user interface system
based on multimodal human computer interaction, which
operates as an intermediate module between the user and the
operating system. The aim of this work is to demonstrate a
multimodal system which gives users the ability to interact with
desktop applications using face, objects, voice and gestures.
These human behaviors constitute the input qualifiers to the
system. Microsoft Kinect multi-sensor was utilized as input
device in order to succeed the natural user interaction, mainly
due to the multimodal capabilities offered by this device. We
demonstrate scenarios which contain all the functions and
capabilities of our system from the perspective of natural user
interaction.
Original languageEnglish
Title of host publication18th International Conference on Digital Signal Processing (DSP)
EditorsAthanasios Skodras
Number of pages8
PublisherWiley-IEEE press
Publication date2013
ISBN (Print)978-1-4673-5807-1
DOIs
Publication statusPublished - 2013
EventInternational conference on Digital Signal Processing - Fira, Greece
Duration: 1 Jul 20133 Jul 2013
Conference number: 18
http://dsp2013.dspconferences.org/

Conference

ConferenceInternational conference on Digital Signal Processing
Number18
CountryGreece
CityFira
Period01/07/201303/07/2013
Internet address
SeriesInternational Conference on Digital Signal Processing proceedings
ISSN1546-1874

Fingerprint

Human computer interaction
User interfaces
Sensors

Cite this

Vidakis, N., Vlasopoulos, A., Kounalakis, T., Varchalamas, P., Dimitriou, M., Kalliatakis, G., ... Triantafyllidis, G. (2013). Multimodal Desktop Interaction: The Face –Object-Gesture–Voice Example. In A. Skodras (Ed.), 18th International Conference on Digital Signal Processing (DSP) Wiley-IEEE press. International Conference on Digital Signal Processing proceedings https://doi.org/10.1109/ICDSP.2013.6622782
Vidakis, Nikolas ; Vlasopoulos, Anastasios ; Kounalakis, Tsampikos ; Varchalamas, Petros ; Dimitriou, Michalis ; Kalliatakis, Gregory ; Syntychakis, Efthimios ; Christofakis, John ; Triantafyllidis, Georgios. / Multimodal Desktop Interaction: The Face –Object-Gesture–Voice Example. 18th International Conference on Digital Signal Processing (DSP). editor / Athanasios Skodras. Wiley-IEEE press, 2013. (International Conference on Digital Signal Processing proceedings).
@inproceedings{16d443fa9eac4ae8899a0b96dbc3e0c3,
title = "Multimodal Desktop Interaction: The Face –Object-Gesture–Voice Example",
abstract = "This paper presents a natural user interface systembased on multimodal human computer interaction, whichoperates as an intermediate module between the user and theoperating system. The aim of this work is to demonstrate amultimodal system which gives users the ability to interact withdesktop applications using face, objects, voice and gestures.These human behaviors constitute the input qualifiers to thesystem. Microsoft Kinect multi-sensor was utilized as inputdevice in order to succeed the natural user interaction, mainlydue to the multimodal capabilities offered by this device. Wedemonstrate scenarios which contain all the functions andcapabilities of our system from the perspective of natural userinteraction.",
author = "Nikolas Vidakis and Anastasios Vlasopoulos and Tsampikos Kounalakis and Petros Varchalamas and Michalis Dimitriou and Gregory Kalliatakis and Efthimios Syntychakis and John Christofakis and Georgios Triantafyllidis",
year = "2013",
doi = "10.1109/ICDSP.2013.6622782",
language = "English",
isbn = "978-1-4673-5807-1",
series = "International Conference on Digital Signal Processing proceedings",
editor = "Athanasios Skodras",
booktitle = "18th International Conference on Digital Signal Processing (DSP)",
publisher = "Wiley-IEEE press",
address = "United States",

}

Vidakis, N, Vlasopoulos, A, Kounalakis, T, Varchalamas, P, Dimitriou, M, Kalliatakis, G, Syntychakis, E, Christofakis, J & Triantafyllidis, G 2013, Multimodal Desktop Interaction: The Face –Object-Gesture–Voice Example. in A Skodras (ed.), 18th International Conference on Digital Signal Processing (DSP). Wiley-IEEE press, International Conference on Digital Signal Processing proceedings, International conference on Digital Signal Processing, Fira, Greece, 01/07/2013. https://doi.org/10.1109/ICDSP.2013.6622782

Multimodal Desktop Interaction: The Face –Object-Gesture–Voice Example. / Vidakis, Nikolas; Vlasopoulos, Anastasios; Kounalakis, Tsampikos; Varchalamas, Petros; Dimitriou, Michalis; Kalliatakis, Gregory; Syntychakis, Efthimios; Christofakis, John; Triantafyllidis, Georgios.

18th International Conference on Digital Signal Processing (DSP). ed. / Athanasios Skodras. Wiley-IEEE press, 2013. (International Conference on Digital Signal Processing proceedings).

Research output: Contribution to book/anthology/report/conference proceedingArticle in proceedingResearchpeer-review

TY - GEN

T1 - Multimodal Desktop Interaction: The Face –Object-Gesture–Voice Example

AU - Vidakis, Nikolas

AU - Vlasopoulos, Anastasios

AU - Kounalakis, Tsampikos

AU - Varchalamas, Petros

AU - Dimitriou, Michalis

AU - Kalliatakis, Gregory

AU - Syntychakis, Efthimios

AU - Christofakis, John

AU - Triantafyllidis, Georgios

PY - 2013

Y1 - 2013

N2 - This paper presents a natural user interface systembased on multimodal human computer interaction, whichoperates as an intermediate module between the user and theoperating system. The aim of this work is to demonstrate amultimodal system which gives users the ability to interact withdesktop applications using face, objects, voice and gestures.These human behaviors constitute the input qualifiers to thesystem. Microsoft Kinect multi-sensor was utilized as inputdevice in order to succeed the natural user interaction, mainlydue to the multimodal capabilities offered by this device. Wedemonstrate scenarios which contain all the functions andcapabilities of our system from the perspective of natural userinteraction.

AB - This paper presents a natural user interface systembased on multimodal human computer interaction, whichoperates as an intermediate module between the user and theoperating system. The aim of this work is to demonstrate amultimodal system which gives users the ability to interact withdesktop applications using face, objects, voice and gestures.These human behaviors constitute the input qualifiers to thesystem. Microsoft Kinect multi-sensor was utilized as inputdevice in order to succeed the natural user interaction, mainlydue to the multimodal capabilities offered by this device. Wedemonstrate scenarios which contain all the functions andcapabilities of our system from the perspective of natural userinteraction.

U2 - 10.1109/ICDSP.2013.6622782

DO - 10.1109/ICDSP.2013.6622782

M3 - Article in proceeding

SN - 978-1-4673-5807-1

T3 - International Conference on Digital Signal Processing proceedings

BT - 18th International Conference on Digital Signal Processing (DSP)

A2 - Skodras, Athanasios

PB - Wiley-IEEE press

ER -

Vidakis N, Vlasopoulos A, Kounalakis T, Varchalamas P, Dimitriou M, Kalliatakis G et al. Multimodal Desktop Interaction: The Face –Object-Gesture–Voice Example. In Skodras A, editor, 18th International Conference on Digital Signal Processing (DSP). Wiley-IEEE press. 2013. (International Conference on Digital Signal Processing proceedings). https://doi.org/10.1109/ICDSP.2013.6622782