Semantic congruence in audio-haptic simulation of footsteps

Luca Turchet*, Stefania Serafin

*Corresponding author

Research output: Contribution to journalJournal articleResearchpeer-review

17 Citations (Scopus)


In this paper we present an experiment whose goal is to investigate subjects’ ability to match pairs of synthetic auditory and haptic stimuli which simulate the sensation of walking on different surfaces. In three non-interactive conditions the audio–haptic stimuli were passively presented through a desktop system, while in three interactive conditions participants produced the audio–haptic feedback interactively while walking. Results show that material typology (i.e., solid or aggregate) is processed very consistently in both the auditory and haptic modalities. Subjects expressed a higher level of semantic congruence for those audio–haptic pairs of materials which belonged to the same typology. Furthermore, better matching ability was found for the passive case compared to the interactive one, although this may be due to the limits of the technology used for the interactive haptic simulations.
Original languageEnglish
JournalApplied Acoustics
Issue number1
Pages (from-to)59-66
Number of pages8
Publication statusPublished - 1 Jan 2014


  • Audio
  • Haptic
  • Semantic congruence

Fingerprint Dive into the research topics of 'Semantic congruence in audio-haptic simulation of footsteps'. Together they form a unique fingerprint.

Cite this