On the relation between discriminant analysis and mutual information for supervised linear feature extraction

作者:

Highlights:

摘要

This paper provides a unifying view of three discriminant linear feature extraction methods: linear discriminant analysis, heteroscedastic discriminant analysis and maximization of mutual information. We propose a model-independent reformulation of the criteria related to these three methods that stresses their similarities and elucidates their differences. Based on assumptions for the probability distribution of the classification data, we obtain sufficient conditions under which two or more of the above criteria coincide. It is shown that these conditions also suffice for Bayes optimality of the criteria. Our approach results in an information-theoretic derivation of linear discriminant analysis and heteroscedastic discriminant analysis. Finally, regarding linear discriminant analysis, we discuss its relation to multidimensional independent component analysis and derive suboptimality bounds based on information theory.

论文关键词:Linear feature extraction,Linear discriminant analysis,Heteroscedastic discriminant analysis,Maximization of mutual information,Bayes error,Negentropy

论文评审过程:Received 27 November 2002, Revised 27 October 2003, Available online 31 January 2004.

论文官网地址:https://doi.org/10.1016/j.patcog.2003.12.002