TY - JOUR
T1 - Bias correction for linear discriminant analysis
AU - Zollanvari, Amin
AU - Abibullaev, Berdakh
N1 - Funding Information:
The work was supported by the Nazarbayev University Faculty Development Competitive Research Grants Program under grant number 021220FD1151.
Publisher Copyright:
© 2021 Elsevier B.V.
PY - 2021/11
Y1 - 2021/11
N2 - Linear discriminant analysis (LDA) is perhaps one of the most fundamental statistical pattern recognition techniques. In this work, we explicitly present, for the first time, an asymptotically exact estimator of the LDA optimal intercept in terms of achieving the lowest overall risk in the classification of two multivariate Gaussian distributions with a common covariance matrix and arbitrary misclassification costs. The proposed estimator of the optimal bias term is developed based on the theory of random matrices of increasing dimension in which the observation dimension and the sample size tend to infinity while keeping their magnitudes comparable. The simple form of this estimator provides us with some analytical insights into the working mechanism of the bias correction in LDA. We then complement these analytical insights with numerical experiments. In particular, empirical results using real data show that insofar as the overall risk is concerned, the proposed bias-corrected form of LDA can outperform the conventional LDA classifier in a wide range of misclassification costs. At the same time, the superiority of the proposed form over LDA tends to be more evident as dimensionality or the ratio between class-specific costs increase.
AB - Linear discriminant analysis (LDA) is perhaps one of the most fundamental statistical pattern recognition techniques. In this work, we explicitly present, for the first time, an asymptotically exact estimator of the LDA optimal intercept in terms of achieving the lowest overall risk in the classification of two multivariate Gaussian distributions with a common covariance matrix and arbitrary misclassification costs. The proposed estimator of the optimal bias term is developed based on the theory of random matrices of increasing dimension in which the observation dimension and the sample size tend to infinity while keeping their magnitudes comparable. The simple form of this estimator provides us with some analytical insights into the working mechanism of the bias correction in LDA. We then complement these analytical insights with numerical experiments. In particular, empirical results using real data show that insofar as the overall risk is concerned, the proposed bias-corrected form of LDA can outperform the conventional LDA classifier in a wide range of misclassification costs. At the same time, the superiority of the proposed form over LDA tends to be more evident as dimensionality or the ratio between class-specific costs increase.
KW - Bias-correction
KW - Discriminant analysis
KW - Pattern recognition
KW - Machine Learning
UR - http://www.scopus.com/inward/record.url?scp=85112533678&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85112533678&partnerID=8YFLogxK
U2 - 10.1016/j.patrec.2021.07.026
DO - 10.1016/j.patrec.2021.07.026
M3 - Article
AN - SCOPUS:85112533678
SN - 0167-8655
VL - 151
SP - 41
EP - 47
JO - Pattern Recognition Letters
JF - Pattern Recognition Letters
ER -