TY - JOUR
T1 - A Revised Formation of Trace Ratio LDA for Small Sample Size Problem
AU - Li, Zhengxin
AU - Nie, Feiping
AU - Wang, Rong
AU - Li, Xuelong
N1 - Publisher Copyright:
IEEE
PY - 2024
Y1 - 2024
N2 - Linear discriminant analysis (LDA) is a classic tool for supervised dimensionality reduction. Because the projected samples can be classified effectively, LDA has been successfully applied in many applications. Among the variants of LDA, trace ratio LDA (TR-LDA) is a classic form due to its explicit meaning. Unfortunately, when the sample size is much smaller than the data dimension, the algorithm for solving TR-LDA does not converge. The so-called small sample size (SSS) problem severely limits the application of TR-LDA. To solve this problem, we propose a revised formation of TR-LDA, which can be applied to datasets with different sizes in a unified form. Then, we present an optimization algorithm to solve the proposed method, explain why it can avoid the SSS problem, and analyze the convergence and computational complexity of the optimization algorithm. Next, based on the introduced theorems, we quantitatively elaborate on when the SSS problem will occur in TR-LDA. Finally, the experimental results on real-world datasets demonstrate the effectiveness of the proposed method.
AB - Linear discriminant analysis (LDA) is a classic tool for supervised dimensionality reduction. Because the projected samples can be classified effectively, LDA has been successfully applied in many applications. Among the variants of LDA, trace ratio LDA (TR-LDA) is a classic form due to its explicit meaning. Unfortunately, when the sample size is much smaller than the data dimension, the algorithm for solving TR-LDA does not converge. The so-called small sample size (SSS) problem severely limits the application of TR-LDA. To solve this problem, we propose a revised formation of TR-LDA, which can be applied to datasets with different sizes in a unified form. Then, we present an optimization algorithm to solve the proposed method, explain why it can avoid the SSS problem, and analyze the convergence and computational complexity of the optimization algorithm. Next, based on the introduced theorems, we quantitatively elaborate on when the SSS problem will occur in TR-LDA. Finally, the experimental results on real-world datasets demonstrate the effectiveness of the proposed method.
KW - Classification
KW - Costs
KW - Dimensionality reduction
KW - Feature extraction
KW - Learning systems
KW - Linear programming
KW - Optimization
KW - Principal component analysis
KW - convergence
KW - dimensionality reduction
KW - small sample size (SSS) problem
KW - trace ratio LDA (TR-LDA)
UR - http://www.scopus.com/inward/record.url?scp=85186093365&partnerID=8YFLogxK
U2 - 10.1109/TNNLS.2024.3362512
DO - 10.1109/TNNLS.2024.3362512
M3 - 文章
AN - SCOPUS:85186093365
SN - 2162-237X
SP - 1
EP - 7
JO - IEEE Transactions on Neural Networks and Learning Systems
JF - IEEE Transactions on Neural Networks and Learning Systems
ER -