Do we need individual head-related transfer functions for vertical localization? The case study of a spectral notch distance metric

Michele Geronazzo, Simone Spagnol, Federico Avanzini

Research output: Contribution to journalJournal articleResearchpeer-review

5 Citations (Scopus)

Abstract

This paper deals with the issue of individualizing the head-related transfer function (HRTF) rendering process for auditory elevation perception. Is it possible to find a nonindividual, personalized HRTF set that allows a listener to have an equally accurate localization performance than with his/her individual HRTFs? We propose a psychoacoustically motivated, anthropometry based mismatch function between HRTF pairs that exploits the close relation between the listener's pinna geometry and localization cues. This is evaluated using an auditory model that computes a mapping between HRTF spectra and perceived spatial locations. Results on a large number of subjects in the center for image processing and integrated computing (CIPIC) and acoustics research institute (ARI) HRTF databases suggest that there exists a nonindividual HRTF set, which allows a listener to have an equally accurate vertical localization than with individual HRTFs. Furthermore, we find the optimal parameterization of the proposed mismatch function, i.e., the one that best reflects the information given by the auditory model. Our findings show that the selection procedure yields statistically significant improvements with respect to dummy-head HRTFs or random HRTF selection, with potentially high impact from an applicative point of view.
Original languageEnglish
JournalIEEE/ACM Transactions on Audio, Speech, and Language Processing
Volume26
Issue number7
Pages (from-to)1247 - 1260
Number of pages14
ISSN2329-9290
DOIs
Publication statusPublished - 1 Jul 2018

Fingerprint

notches
transfer functions
Transfer functions
anthropometry
Anthropometry
dummies
cues
Parameterization
parameterization
image processing
Image processing
Acoustics
acoustics
Geometry
geometry

Keywords

  • HRTF selection
  • Spatial audio
  • auditory models
  • head-related transfer functions (HRTFs)
  • individualized HRTFs
  • spectral notch metric
  • vertical localization

Cite this

@article{8c6f3bb6dc884f85882491e58e3dd96a,
title = "Do we need individual head-related transfer functions for vertical localization? The case study of a spectral notch distance metric",
abstract = "This paper deals with the issue of individualizing the head-related transfer function (HRTF) rendering process for auditory elevation perception. Is it possible to find a nonindividual, personalized HRTF set that allows a listener to have an equally accurate localization performance than with his/her individual HRTFs? We propose a psychoacoustically motivated, anthropometry based mismatch function between HRTF pairs that exploits the close relation between the listener's pinna geometry and localization cues. This is evaluated using an auditory model that computes a mapping between HRTF spectra and perceived spatial locations. Results on a large number of subjects in the center for image processing and integrated computing (CIPIC) and acoustics research institute (ARI) HRTF databases suggest that there exists a nonindividual HRTF set, which allows a listener to have an equally accurate vertical localization than with individual HRTFs. Furthermore, we find the optimal parameterization of the proposed mismatch function, i.e., the one that best reflects the information given by the auditory model. Our findings show that the selection procedure yields statistically significant improvements with respect to dummy-head HRTFs or random HRTF selection, with potentially high impact from an applicative point of view.",
keywords = "HRTF selection, Spatial audio, auditory models, head-related transfer functions (HRTFs), individualized HRTFs, spectral notch metric, vertical localization",
author = "Michele Geronazzo and Simone Spagnol and Federico Avanzini",
year = "2018",
month = "7",
day = "1",
doi = "10.1109/TASLP.2018.2821846",
language = "English",
volume = "26",
pages = "1247 -- 1260",
journal = "IEEE/ACM Transactions on Audio, Speech, and Language Processing",
issn = "2329-9290",
publisher = "IEEE Signal Processing Society",
number = "7",

}

Do we need individual head-related transfer functions for vertical localization? The case study of a spectral notch distance metric. / Geronazzo, Michele; Spagnol, Simone; Avanzini, Federico.

In: IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 26, No. 7, 01.07.2018, p. 1247 - 1260.

Research output: Contribution to journalJournal articleResearchpeer-review

TY - JOUR

T1 - Do we need individual head-related transfer functions for vertical localization? The case study of a spectral notch distance metric

AU - Geronazzo, Michele

AU - Spagnol, Simone

AU - Avanzini, Federico

PY - 2018/7/1

Y1 - 2018/7/1

N2 - This paper deals with the issue of individualizing the head-related transfer function (HRTF) rendering process for auditory elevation perception. Is it possible to find a nonindividual, personalized HRTF set that allows a listener to have an equally accurate localization performance than with his/her individual HRTFs? We propose a psychoacoustically motivated, anthropometry based mismatch function between HRTF pairs that exploits the close relation between the listener's pinna geometry and localization cues. This is evaluated using an auditory model that computes a mapping between HRTF spectra and perceived spatial locations. Results on a large number of subjects in the center for image processing and integrated computing (CIPIC) and acoustics research institute (ARI) HRTF databases suggest that there exists a nonindividual HRTF set, which allows a listener to have an equally accurate vertical localization than with individual HRTFs. Furthermore, we find the optimal parameterization of the proposed mismatch function, i.e., the one that best reflects the information given by the auditory model. Our findings show that the selection procedure yields statistically significant improvements with respect to dummy-head HRTFs or random HRTF selection, with potentially high impact from an applicative point of view.

AB - This paper deals with the issue of individualizing the head-related transfer function (HRTF) rendering process for auditory elevation perception. Is it possible to find a nonindividual, personalized HRTF set that allows a listener to have an equally accurate localization performance than with his/her individual HRTFs? We propose a psychoacoustically motivated, anthropometry based mismatch function between HRTF pairs that exploits the close relation between the listener's pinna geometry and localization cues. This is evaluated using an auditory model that computes a mapping between HRTF spectra and perceived spatial locations. Results on a large number of subjects in the center for image processing and integrated computing (CIPIC) and acoustics research institute (ARI) HRTF databases suggest that there exists a nonindividual HRTF set, which allows a listener to have an equally accurate vertical localization than with individual HRTFs. Furthermore, we find the optimal parameterization of the proposed mismatch function, i.e., the one that best reflects the information given by the auditory model. Our findings show that the selection procedure yields statistically significant improvements with respect to dummy-head HRTFs or random HRTF selection, with potentially high impact from an applicative point of view.

KW - HRTF selection

KW - Spatial audio

KW - auditory models

KW - head-related transfer functions (HRTFs)

KW - individualized HRTFs

KW - spectral notch metric

KW - vertical localization

UR - http://www.scopus.com/inward/record.url?scp=85042443005&partnerID=8YFLogxK

U2 - 10.1109/TASLP.2018.2821846

DO - 10.1109/TASLP.2018.2821846

M3 - Journal article

VL - 26

SP - 1247

EP - 1260

JO - IEEE/ACM Transactions on Audio, Speech, and Language Processing

JF - IEEE/ACM Transactions on Audio, Speech, and Language Processing

SN - 2329-9290

IS - 7

ER -