TY - JOUR
T1 - Robust Bilinear Probabilistic PCA Using a Matrix Variate t Distribution
AU - Zhao, Jianhua
AU - Ma, Xuan
AU - Shi, Lei
AU - Wang, Zhen
N1 - Publisher Copyright:
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
PY - 2023/12/1
Y1 - 2023/12/1
N2 - —The bilinear probabilistic principal component analysis (BPPCA) was introduced recently as a model-based dimension reduction technique on matrix data. However, BPPCA is based on the Gaussian assumption and hence is vulnerable to potential outlying matrix-valued observations. In this article, we present a new robust extension of BPPCA, called BPPCA using a matrix variate t distribution (tBPPCA), that is built upon a matrix variate t distribution. Like the multivariate t, this distribution offers an additional robustness tuning parameter, which can downweight outliers. By introducing a Gamma distributed latent weight variable, this distribution can be represented hierarchically. With this representation, two efficient accelerated expectation–maximization (EM)-like algorithms for parameter estimation are developed. Experiments on a number of synthetic and real datasets are conducted to understand tBPPCA and compare with several closely related competitors, including its vector-based counterpart. The results reveal that tBPPCA is generally more robust and accurate in the presence of outliers. Moreover, the expected latent weights under tBPPCA can be effectively used for outliers’ detection, which is much more reliable than its vector-based counterpart due to its better robustness.
AB - —The bilinear probabilistic principal component analysis (BPPCA) was introduced recently as a model-based dimension reduction technique on matrix data. However, BPPCA is based on the Gaussian assumption and hence is vulnerable to potential outlying matrix-valued observations. In this article, we present a new robust extension of BPPCA, called BPPCA using a matrix variate t distribution (tBPPCA), that is built upon a matrix variate t distribution. Like the multivariate t, this distribution offers an additional robustness tuning parameter, which can downweight outliers. By introducing a Gamma distributed latent weight variable, this distribution can be represented hierarchically. With this representation, two efficient accelerated expectation–maximization (EM)-like algorithms for parameter estimation are developed. Experiments on a number of synthetic and real datasets are conducted to understand tBPPCA and compare with several closely related competitors, including its vector-based counterpart. The results reveal that tBPPCA is generally more robust and accurate in the presence of outliers. Moreover, the expected latent weights under tBPPCA can be effectively used for outliers’ detection, which is much more reliable than its vector-based counterpart due to its better robustness.
KW - Dimension reduction
KW - expectation–maximization (EM)
KW - matrix data
KW - matrix variate t distribution
KW - principal component analysis
UR - http://www.scopus.com/inward/record.url?scp=85132520174&partnerID=8YFLogxK
U2 - 10.1109/TNNLS.2022.3170797
DO - 10.1109/TNNLS.2022.3170797
M3 - 文章
C2 - 35533172
AN - SCOPUS:85132520174
SN - 2162-237X
VL - 34
SP - 10683
EP - 10697
JO - IEEE Transactions on Neural Networks and Learning Systems
JF - IEEE Transactions on Neural Networks and Learning Systems
IS - 12
ER -