Soft Dropout and Its Variational Bayes Approximation

Jiyang Xie, Zhanyu Ma, Guoqiang Zhang, Jing-Hao Xue, Zheng-Hua Tan, Jun Guo

Research output: Contribution to book/anthology/report/conference proceedingArticle in proceedingResearchpeer-review

7 Citations (Scopus)

Abstract

Soft dropout, a generalization of standard 'hard' dropout, is introduced to regularize the parameters in neural networks and prevent overfitting. We replace the 'hard' dropout mask following a Bernoulli distribution with the 'soft' mask following a beta distribution to drop the hidden nodes in different levels. The soft dropout method can introduce continuous mask coefficients in the interval of [0, 1], rather than only zero and one. Meanwhile, in order to implement the adaptive dropout rate via adaptive distribution parameters, we respectively utilize the half-Gaussian distributed and the half-Laplace distributed variables to approximate the beta distributed masks and apply a variation of variational Bayes optimization called stochastic gradient variational Bayes (SGVB) algorithm to optimize the distribution parameters. In the experiments, compared with the standard soft dropout with fixed dropout rate, the adaptive soft dropout method generally improves the performance. In addition, the proposed soft dropout and its adaptive versions achieve performance improvement compared with the referred methods on both image classification and regression tasks.

Original languageEnglish
Title of host publication2019 IEEE 29th International Workshop on Machine Learning for Signal Processing (MLSP)
Number of pages6
PublisherIEEE (Institute of Electrical and Electronics Engineers)
Publication date5 Dec 2019
Pages1-6
Article number8918818
ISBN (Print)978-1-7281-0825-4
ISBN (Electronic)978-1-7281-0824-7
DOIs
Publication statusPublished - 5 Dec 2019
Event2019 IEEE 29th International Workshop on Machine Learning for Signal Processing (MLSP) - Pittsburgh, United States
Duration: 13 Oct 201916 Oct 2019

Conference

Conference2019 IEEE 29th International Workshop on Machine Learning for Signal Processing (MLSP)
Country/TerritoryUnited States
CityPittsburgh
Period13/10/201916/10/2019
SeriesIEEE International Workshop on Machine Learning for Signal Processing (MLSP). Proceedings.

Keywords

  • Bayesian approximation
  • Neural networks
  • beta distribution
  • soft dropout

Fingerprint

Dive into the research topics of 'Soft Dropout and Its Variational Bayes Approximation'. Together they form a unique fingerprint.

Cite this