Abstract
In our work, we propose a novel formulation for supervised dimensionality reduction based on a nonlinear dependency criterion called Statistical Distance Correlation, (Székely et al., 2007). We propose an objective which is free of distributional assumptions on regression variables and regression model assumptions. Our proposed formulation is based on learning a low-dimensional feature representation z, which maximizes the squared sum of Distance Correlations between low-dimensional features z and response y, and also between features z and covariates x. We propose a novel algorithm to optimize our proposed objective using the Generalized Minimization Maximization method of (Parizi et al., 2015). We show superior empirical results on multiple datasets proving the effectiveness of our proposed approach over several relevant state-of-the-art supervised dimensionality reduction methods.
Original language | English (US) |
---|---|
Pages (from-to) | 960-984 |
Number of pages | 25 |
Journal | Electronic Journal of Statistics |
Volume | 12 |
Issue number | 1 |
DOIs | |
State | Published - 2018 |
All Science Journal Classification (ASJC) codes
- Statistics and Probability
- Statistics, Probability and Uncertainty
Keywords
- Distance correlation
- Fixed point iteration
- Minorization maximization
- Multivariate statistical independence
- Optimization
- Representation learning
- Supervised dimensionality reduction