×

Expectile regression for analyzing heteroscedasticity in high dimension. (English) Zbl 1414.62324

Summary: High-dimensional data often display heteroscedasticity and this feature has attracted a lot of attention and discussion. In this paper, we propose regularized expectile regression with SCAD penalty for analyzing heteroscedasticity in high dimension when the error has finite moments. Since the corresponding optimization problem is nonconvex due to the SCAD penalty, we adopt the CCCP (coupling of the concave and convex procedure) algorithm to solve this problem. Under some regular conditions, we can prove that with probability tending to one, the proposed algorithm converges to the oracle estimator after several iterations. We should address that the higher order moment the error has, the higher dimension cardinality our procedure can handle with. If the error follows gaussian or sub-gaussian distribution, our method can be extended to deal with ultra high-dimensional data. Furthermore, by taking different expectile weight level \(\alpha\), we are able to detect heteroscedasticity and explore the entire conditional distribution of the response variable given all the covariates. We investigate the performances of our proposed method through Monte Carlo simulation study and real application and the numerical results show that the resulting estimator by our algorithm enjoys good performance and demonstrate the usefulness of our proposed method to analyze heteroscedasticity.

MSC:

62J07 Ridge regression; shrinkage estimators (Lasso)
65C05 Monte Carlo methods
PDFBibTeX XMLCite
Full Text: DOI

References:

[1] Aigner, D. J.; Amemiya, T.; Poirier, D. J., On the estimation of production frontiers: maximum likelihood estimation of the parameters of a discontinuous density function, Internat. Econom. Rev., 377-396, (1976) · Zbl 0339.62083
[2] Breiman, L., Heuristics of instability and stabilization in model selection, Ann. Statist., 24, 6, 2350-2383, (1996) · Zbl 0867.62055
[3] Fan, J.; Fan, Y.; Barut, E., Adaptive robust variable selection, Ann. Statist., 42, 1, 324-351, (2014) · Zbl 1296.62144
[4] Fan, J.; Li, R., Variable selection via nonconcave penalized likelihood and its oracle properties, J. Amer. Statist. Assoc., 96, 456, 1348-1360, (2001) · Zbl 1073.62547
[5] Fan, J.; Li, Q.; Wang, Y., Estimation of high dimensional mean regression in the absence of symmetry and light tail assumptions, J. Roy. Statist. Soc., 79, 1, 247-265, (2017) · Zbl 1414.62178
[6] Fan, J.; Lv, J., Nonconcave penalized likelihood with np-dimensionality, IEEE Trans. Inform. Theory, 57, 8, 5467-5484, (2011) · Zbl 1365.62277
[7] Fan, J.; Peng, H., Nonconcave penalized likelihood with a diverging number of parameters, Ann. Statist., 32, 3, 928-961, (2004) · Zbl 1092.62031
[8] Gu, Y.; Zou, H., High-dimensional generalizations of asymmetric least squares regression and their applications, Ann. Statist., 44, 6, 2661-2694, (2016) · Zbl 1364.62185
[9] Huang, C. C.; Liu, K.; Pope, R. M.; Du, P.; Lin, S.; Rajamannan, N. M., Activated tlr signaling in atherosclerosis among women with lower framingham risk score: the multi-ethnic study of atherosclerosis, PLoS One, 6, 6, e21067, (2011)
[10] Kim, Y.; Choi, H.; Oh, H. S., Smoothly clipped absolute deviation on high dimensions, J. Amer. Statist. Assoc., 103, 484, 1665-1673, (2008) · Zbl 1286.62062
[11] Le Thi Hoai, A.; Tao, P. D., Solving a class of linearly constrained indefinite quadratic problems by DC algorithms, J. Global Optim., 11, 3, 253-285, (1997) · Zbl 0905.90131
[12] Lilliefors, H. W., On the Kolmogorov-Smirnov test for normality with mean and variance unknown, J. Amer. Statist. Assoc., 62, 318, 399-402, (1967)
[13] Negahban, S.; Yu, B.; Wainwright, M. J.; Ravikumar, P. K., A unified framework for high-dimensional analysis of \(M\)-estimators with decomposable regularizers, (Advances in Neural Information Processing Systems, (2009)), 1348-1356
[14] Newey, W. K.; Powell, J. L., Asymmetric least squares estimation and testing, Econometrica, 819-847, (1987) · Zbl 0625.62047
[15] Schnabel, S. K.; Eilers, P. H., Optimal expectile smoothing, Comput. Statist. Data Anal., 53, 12, 4168-4177, (2009) · Zbl 1453.62192
[16] Sobotka, F.; Kneib, T., Geoadditive expectile regression, Comput. Statist. Data Anal., 56, 4, 755-767, (2012) · Zbl 1241.62058
[17] Tibshirani, R., Regression shrinkage and selection via the lasso, J. R. Stat. Soc. Ser. B Stat. Methodol., 267-288, (1996) · Zbl 0850.62538
[18] Waltrup, L. S.; Sobotka, F.; Kneib, T.; Kauermann, G., Expectile and quantile regressiondavid and goliath?, Stat. Model., 15, 5, 433-456, (2015) · Zbl 07258997
[19] Wang, L.; Wu, Y.; Li, R., Quantile regression for analyzing heterogeneity in ultra-high dimension, J. Amer. Statist. Assoc., 107, 497, 214-222, (2012) · Zbl 1328.62468
[20] Wu, Y.; Liu, Y., Variable selection in quantile regression, Statist. Sinica, 801-817, (2009) · Zbl 1166.62012
[21] Ye, F.; Zhang, C. H., Rate minimaxity of the lasso and Dantzig selector for the lq loss in lr balls, J. Mach. Learn. Res., 11, Dec, 3519-3540, (2010) · Zbl 1242.62074
[22] Yuille, A. L.; Rangarajan, A., The concave-convex procedure, Neural Comput., 15, 4, 915-936, (2003) · Zbl 1022.68112
This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. In some cases that data have been complemented/enhanced by data from zbMATH Open. This attempts to reflect the references listed in the original paper as accurately as possible without claiming completeness or a perfect matching.