TY - JOUR
T1 - Rethinking Bayesian Learning for Data Analysis
T2 - The art of prior and inference in sparsity-aware modeling
AU - Cheng, Lei
AU - Yin, Feng
AU - Theodoridis, Sergios
AU - Chatzis, Soterios
AU - Chang, Tsung-Hui
PY - 2022/11
Y1 - 2022/11
N2 - Sparse modeling for signal processing and machine learning, in general, has been at the focus of scientific research for over two decades. Among others, supervised sparsity-aware learning (SAL) consists of two major paths paved by 1) discriminative methods that establish direct input-output mapping based on a regularized cost function optimization and 2) generative methods that learn the underlying distributions. The latter, more widely known as Bayesian methods, enable uncertainty evaluation with respect to the performed predictions. Furthermore, they can better exploit related prior information and also, in principle, can naturally introduce robustness into the model, due to their unique capacity to marginalize out uncertainties related to the parameter estimates. Moreover, hyperparameters (tuning parameters) associated with the adopted priors, which correspond to cost function regularizers, can be learned via the training data and not via costly cross-validation techniques, which is, in general, the case with the discriminative methods.
AB - Sparse modeling for signal processing and machine learning, in general, has been at the focus of scientific research for over two decades. Among others, supervised sparsity-aware learning (SAL) consists of two major paths paved by 1) discriminative methods that establish direct input-output mapping based on a regularized cost function optimization and 2) generative methods that learn the underlying distributions. The latter, more widely known as Bayesian methods, enable uncertainty evaluation with respect to the performed predictions. Furthermore, they can better exploit related prior information and also, in principle, can naturally introduce robustness into the model, due to their unique capacity to marginalize out uncertainties related to the parameter estimates. Moreover, hyperparameters (tuning parameters) associated with the adopted priors, which correspond to cost function regularizers, can be learned via the training data and not via costly cross-validation techniques, which is, in general, the case with the discriminative methods.
UR - http://www.scopus.com/inward/record.url?scp=85141517685&partnerID=8YFLogxK
U2 - 10.1109/MSP.2022.3198201
DO - 10.1109/MSP.2022.3198201
M3 - Journal article
SN - 1053-5888
VL - 39
SP - 18
EP - 52
JO - I E E E - Signal Processing Magazine
JF - I E E E - Signal Processing Magazine
IS - 6
ER -