Generating Accurate and Diverse Audio Captions Through Variational Autoencoder Framework

Yiming Zhang, Ruoyi Du, Zheng Hua Tan, Wenwu Wang, Zhanyu Ma*

*Kontaktforfatter

Publikation: Bidrag til tidsskriftTidsskriftartikelForskningpeer review

1 Citationer (Scopus)

Abstract

Generating both diverse and accurate descriptions is an essential goal in the audio captioning task. Traditional methods mainly focus on improving the accuracy of the generated captions but ignore their diversity. In contrast, recent methods have considered generating diverse captions for a given audio clip, but with the potential trade-off in caption accuracy. In this work, we propose a new diverse audio captioning method based on a variational autoencoder structure, dubbed AC-VAE, aiming to achieve a better trade-off between the diversity and accuracy of the generated captions. To improve diversity, AC-VAE learns the latent word distribution at each location based on contextual information. To uphold accuracy, AC-VAE incorporates an autoregressive prior module and a global constraint module, which enable precise modeling of word distribution and encourage semantic consistency of captions at the sentence level. We evaluate the proposed AC-VAE on the Clotho dataset. Experimental results show that AC-VAE achieves a better trade-off between diversity and accuracy compared to the state-of-the-art methods.

OriginalsprogEngelsk
TidsskriftIEEE Signal Processing Letters
Vol/bind31
Sider (fra-til)2520-2524
Antal sider5
ISSN1070-9908
DOI
StatusUdgivet - 2024

Bibliografisk note

Publisher Copyright:
© 1994-2012 IEEE.

Fingeraftryk

Dyk ned i forskningsemnerne om 'Generating Accurate and Diverse Audio Captions Through Variational Autoencoder Framework'. Sammen danner de et unikt fingeraftryk.

Citationsformater