Skip to main content
Top

2015 | OriginalPaper | Chapter

77. Model Selection for High-Dimensional Problems

Authors : Jing-Zhi Huang, Zhan Shi, Wei Zhong

Published in: Handbook of Financial Econometrics and Statistics

Publisher: Springer New York

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

High-dimensional data analysis is becoming more and more important to both academics and practitioners in finance and economics but is also very challenging because the number of variables or parameters in connection with such data can be larger than the sample size. Recently, several variable selection approaches have been developed and used to help us select significant variables and construct a parsimonious model simultaneously. In this chapter, we first provide an overview of model selection approaches in the context of penalized least squares. We then review independence screening, a recently developed method for analyzing ultrahigh-dimensional data where the number of variables or parameters can be exponentially larger than the sample size. Finally, we discuss and advocate multistage procedures that combine independence screening and variable selection and that may be especially suitable for analyzing high-frequency financial data.
Penalized least squares seek to keep important predictors in a model while penalizing coefficients associated with irrelevant predictors. As such, under certain conditions, penalized least squares can lead to a sparse solution for linear models and achieve asymptotic consistency in separating relevant variables from irrelevant ones. Independence screening selects relevant variables based on certain measures of marginal correlations between candidate variables and the response.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Appendix
Available only for authorised users
Literature
go back to reference Akaike, H. (1973). Information theory and an extension of the maximum likelihood principle. In B. N. Petrov & F. Csaki (Eds.), Second international symposium on information theory (Vol. 1, pp. 267–281). Budapest: Akademiai Kiado. Akaike, H. (1973). Information theory and an extension of the maximum likelihood principle. In B. N. Petrov & F. Csaki (Eds.), Second international symposium on information theory (Vol. 1, pp. 267–281). Budapest: Akademiai Kiado.
go back to reference Allen, D. (1974). The relationship between variable selection and data augmentation and a method for prediction. Technometrics, 16, 125–127.CrossRef Allen, D. (1974). The relationship between variable selection and data augmentation and a method for prediction. Technometrics, 16, 125–127.CrossRef
go back to reference Anderson, H. M., & Vahid, F. (2007). Forecasting the volatility of Australian stock returns. Journal of Business and Economic Statistics, 25, 76–90.CrossRef Anderson, H. M., & Vahid, F. (2007). Forecasting the volatility of Australian stock returns. Journal of Business and Economic Statistics, 25, 76–90.CrossRef
go back to reference Antoniadis, A., & Fan, J. (2001). Regularization of wavelets approximations (with discussion). Journal of the American Statistical Association, 96, 939–967.CrossRef Antoniadis, A., & Fan, J. (2001). Regularization of wavelets approximations (with discussion). Journal of the American Statistical Association, 96, 939–967.CrossRef
go back to reference Bai, J., & Ng, S. (2008). Forecasting economic time series using targeted predictors. Journal of Econometrics, 146, 304–317.CrossRef Bai, J., & Ng, S. (2008). Forecasting economic time series using targeted predictors. Journal of Econometrics, 146, 304–317.CrossRef
go back to reference Bellman, R. (1961). Adaptive control processes: A guided tour. Pattern Recognition Letters, 29, 1351–1357. Bellman, R. (1961). Adaptive control processes: A guided tour. Pattern Recognition Letters, 29, 1351–1357.
go back to reference Breiman, L. (1996). Heuristics of instability and stabilization in model selection. Annals of Statistics, 24, 2350–2383.CrossRef Breiman, L. (1996). Heuristics of instability and stabilization in model selection. Annals of Statistics, 24, 2350–2383.CrossRef
go back to reference Campbell, J. Y., Lo, A. W., & MacKinlay, A. C. (1997). The econometrics of financial markets. Princeton: Princeton University Press. Campbell, J. Y., Lo, A. W., & MacKinlay, A. C. (1997). The econometrics of financial markets. Princeton: Princeton University Press.
go back to reference Candes, E., & Tao, T. (2007). The Dantzig selector: Statistical estimation when p is much larger than n (with discussion). Annals of Statistics, 35, 2313–2404.CrossRef Candes, E., & Tao, T. (2007). The Dantzig selector: Statistical estimation when p is much larger than n (with discussion). Annals of Statistics, 35, 2313–2404.CrossRef
go back to reference Chen, X., Zou, C., & Cook, R. D. (2010). Coordinate-independent sparse sufficient dimension reduction and variable selection. Annals of Statistics, 38, 3696–3723.CrossRef Chen, X., Zou, C., & Cook, R. D. (2010). Coordinate-independent sparse sufficient dimension reduction and variable selection. Annals of Statistics, 38, 3696–3723.CrossRef
go back to reference Comon, P. (1994). Independent component analysis, a new concept? Signal Processing, 36, 287–314.CrossRef Comon, P. (1994). Independent component analysis, a new concept? Signal Processing, 36, 287–314.CrossRef
go back to reference Cook, R. D. (1998). Regression graphics: Ideas for studying regressions through graphics. New York: John Wiley & Sons. Cook, R. D. (1998). Regression graphics: Ideas for studying regressions through graphics. New York: John Wiley & Sons.
go back to reference Cook, R. D., & Weisberg, S. (1991). Sliced inverse regression for dimension reduction: Comment. Journal of the American Statistical Association, 86, 328–332. Cook, R. D., & Weisberg, S. (1991). Sliced inverse regression for dimension reduction: Comment. Journal of the American Statistical Association, 86, 328–332.
go back to reference Craven, P., & Wahba, G. (1979). Smoothing noisy data with spline functions: Estimating the correct degree of smoothing by the method of generalized cross-validation. Numerische Mathematik, 31, 377–403.CrossRef Craven, P., & Wahba, G. (1979). Smoothing noisy data with spline functions: Estimating the correct degree of smoothing by the method of generalized cross-validation. Numerische Mathematik, 31, 377–403.CrossRef
go back to reference Demartines, P., & Hérault, J. (1997). Curvilinear component analysis: A self-organizing neural network for nonlinear mapping of data sets. Neural Networks, 8, 148–154.CrossRef Demartines, P., & Hérault, J. (1997). Curvilinear component analysis: A self-organizing neural network for nonlinear mapping of data sets. Neural Networks, 8, 148–154.CrossRef
go back to reference Efron, B., Hastie, T., Johnstone, I., & Tibshirani, R. (2004). Least angle regression (with discussion). Annals of Statistics, 32, 409–499. Efron, B., Hastie, T., Johnstone, I., & Tibshirani, R. (2004). Least angle regression (with discussion). Annals of Statistics, 32, 409–499.
go back to reference Fan, J., & Li, R. (2001). Variable selection via nonconcave penalized likelihood and it oracle properties. Journal of the American Statistical Association, 96, 1348–1360.CrossRef Fan, J., & Li, R. (2001). Variable selection via nonconcave penalized likelihood and it oracle properties. Journal of the American Statistical Association, 96, 1348–1360.CrossRef
go back to reference Fan, J., & Li, R. (2006). Statistical challenges with high dimensionality: Feature selection in knowledge discovery. Proceedings of the International Congress of Mathematicians, European Mathematical Society, Zurich, III, 595–622. Fan, J., & Li, R. (2006). Statistical challenges with high dimensionality: Feature selection in knowledge discovery. Proceedings of the International Congress of Mathematicians, European Mathematical Society, Zurich, III, 595–622.
go back to reference Fan, J., & Lv, J. (2008). Sure independence screening for ultrahigh dimensional feature space (with discussion). Journal of the Royal Statistical Society, Series B, 70, 849–911.CrossRef Fan, J., & Lv, J. (2008). Sure independence screening for ultrahigh dimensional feature space (with discussion). Journal of the Royal Statistical Society, Series B, 70, 849–911.CrossRef
go back to reference Fan, J., & Song, R. (2010). Sure independence screening in generalized linear models with NP-dimensionality. The Annals of Statistics, 38, 3567–3604.CrossRef Fan, J., & Song, R. (2010). Sure independence screening in generalized linear models with NP-dimensionality. The Annals of Statistics, 38, 3567–3604.CrossRef
go back to reference Fan, J., Samworth, R., & Wu, Y. (2009). Ultrahigh dimensional feature selection: Beyond the linear model. Journal of Machine Learning Research, 10, 1829–1853. Fan, J., Samworth, R., & Wu, Y. (2009). Ultrahigh dimensional feature selection: Beyond the linear model. Journal of Machine Learning Research, 10, 1829–1853.
go back to reference Fan, J., Feng, Y., & Song, R. (2011a). Nonparametric independence screening in sparse ultra-high dimensional additive models. Journal of the American Statistical Association, 106, 544–557.CrossRef Fan, J., Feng, Y., & Song, R. (2011a). Nonparametric independence screening in sparse ultra-high dimensional additive models. Journal of the American Statistical Association, 106, 544–557.CrossRef
go back to reference Fan, J., Lv, J., & Qi, L. (2011b). Sparse high-dimensional models in economics. Annual Review of Economics, 3, 291–317.CrossRef Fan, J., Lv, J., & Qi, L. (2011b). Sparse high-dimensional models in economics. Annual Review of Economics, 3, 291–317.CrossRef
go back to reference Frank, I. E., & Friedman, J. H. (1993). A statistical view of some chemometrics regression tools. Technometrics, 35, 109–148.CrossRef Frank, I. E., & Friedman, J. H. (1993). A statistical view of some chemometrics regression tools. Technometrics, 35, 109–148.CrossRef
go back to reference Friedman, J. (2008). Fast sparse regression and classification. International Journal of Forecasting, 28, 722–738.CrossRef Friedman, J. (2008). Fast sparse regression and classification. International Journal of Forecasting, 28, 722–738.CrossRef
go back to reference Goto, S., & Xu, Y. (2010). On mean variance portfolio optimization: Improving performance through better use of hedging relations (Working paper), University of Rhode Island. Goto, S., & Xu, Y. (2010). On mean variance portfolio optimization: Improving performance through better use of hedging relations (Working paper), University of Rhode Island.
go back to reference Hall, P., & Miller, H. (2009). Using generalized correlation to effect variable selection in very high dimensional problems. Journal of Computational and Graphical Statistics, 18, 533–550.CrossRef Hall, P., & Miller, H. (2009). Using generalized correlation to effect variable selection in very high dimensional problems. Journal of Computational and Graphical Statistics, 18, 533–550.CrossRef
go back to reference Hastie, T. (1994). Principal curves and surfaces. DTIC Document. Hastie, T. (1994). Principal curves and surfaces. DTIC Document.
go back to reference Hastie, T., Tibshirani, R., & Friedman, J. (2009). The elements of statistical learning: Data mining, inference, and prediction (2nd ed.). New York: Springer.CrossRef Hastie, T., Tibshirani, R., & Friedman, J. (2009). The elements of statistical learning: Data mining, inference, and prediction (2nd ed.). New York: Springer.CrossRef
go back to reference Hoerl, A., & Kennard, R. (1970). Ridge regression: Biased estimation for nonorthogonal problems. Technometrics, 12, 55–67.CrossRef Hoerl, A., & Kennard, R. (1970). Ridge regression: Biased estimation for nonorthogonal problems. Technometrics, 12, 55–67.CrossRef
go back to reference Ji, P., & Jin, J. (2012). UPS delivers optimal phase diagram in high dimensional variable selection. Annals of Statistics, 40, 73–103.CrossRef Ji, P., & Jin, J. (2012). UPS delivers optimal phase diagram in high dimensional variable selection. Annals of Statistics, 40, 73–103.CrossRef
go back to reference Li, K. C. (1991). Sliced inverse regression for dimension reduction. Journal of the American Statistical Association, 86, 285–299.CrossRef Li, K. C. (1991). Sliced inverse regression for dimension reduction. Journal of the American Statistical Association, 86, 285–299.CrossRef
go back to reference Li, L. (2007). Sparse sufficient dimension reduction. Biometrika, 94, 603–613.CrossRef Li, L. (2007). Sparse sufficient dimension reduction. Biometrika, 94, 603–613.CrossRef
go back to reference Li, R., Zhong, W., Zhu, L. (2012). Feature screening via distance correlation learning. Journal of the American Statistical Association, 107, 1129–1139. Li, R., Zhong, W., Zhu, L. (2012). Feature screening via distance correlation learning. Journal of the American Statistical Association, 107, 1129–1139.
go back to reference Mallows, C. L. (1973). Some comments on Cp. Technometrics, 15, 661–675. Mallows, C. L. (1973). Some comments on Cp. Technometrics, 15, 661–675.
go back to reference Mardia, K. V., Kent, J. T., & Bibby, J. M. (1980). Multivariate analysis. London: Academic. Mardia, K. V., Kent, J. T., & Bibby, J. M. (1980). Multivariate analysis. London: Academic.
go back to reference Miller, A. J. (2002). Subset selection in regression (2nd ed.). New York: Chapman & HALL/CRC.CrossRef Miller, A. J. (2002). Subset selection in regression (2nd ed.). New York: Chapman & HALL/CRC.CrossRef
go back to reference Schwartz, G. (1978). Estimating the dimension of a model. Annals of Statistics, 6, 461–464.CrossRef Schwartz, G. (1978). Estimating the dimension of a model. Annals of Statistics, 6, 461–464.CrossRef
go back to reference Shao, J. (1997). An asymptotic theory for linear model selection. Statistica Sinica, 7, 221–264. Shao, J. (1997). An asymptotic theory for linear model selection. Statistica Sinica, 7, 221–264.
go back to reference Stock, J. H., & Watson, M. W. (2002). Forecasting using principal components from a large number of predictors. Journal of the American Statistical Association, 97, 1167–1179.CrossRef Stock, J. H., & Watson, M. W. (2002). Forecasting using principal components from a large number of predictors. Journal of the American Statistical Association, 97, 1167–1179.CrossRef
go back to reference Tibshirani, R. (1996). Regression shrinkage and selection via LASSO. Journal of the Royal Statistical Society, Series B, 58, 267–288. Tibshirani, R. (1996). Regression shrinkage and selection via LASSO. Journal of the Royal Statistical Society, Series B, 58, 267–288.
go back to reference Wang, H., Li, R., & Tsai, C. L. (2007). Tuning parameter selectors for the smoothly clipped absolute deviation method. Biometrika, 94, 553–568.CrossRef Wang, H., Li, R., & Tsai, C. L. (2007). Tuning parameter selectors for the smoothly clipped absolute deviation method. Biometrika, 94, 553–568.CrossRef
go back to reference Zhang, C. (2010). Nearly unbiased variable selection under minimax concave penalty. Annals of Statistics, 38, 894–942.CrossRef Zhang, C. (2010). Nearly unbiased variable selection under minimax concave penalty. Annals of Statistics, 38, 894–942.CrossRef
go back to reference Zhou, J., & He, X. (2008). Dimension reduction based on constrained canonical correlation and variable filtering. Annals of Statistics, 36, 1649–1668.CrossRef Zhou, J., & He, X. (2008). Dimension reduction based on constrained canonical correlation and variable filtering. Annals of Statistics, 36, 1649–1668.CrossRef
go back to reference Zhu, L., Li, L., Li, R., & Zhu, L. (2011). Model-free feature screening for ultrahigh dimensional data. Journal of the American Statistical Association, 106, 1464–1475.CrossRef Zhu, L., Li, L., Li, R., & Zhu, L. (2011). Model-free feature screening for ultrahigh dimensional data. Journal of the American Statistical Association, 106, 1464–1475.CrossRef
go back to reference Zou, H. (2006). The adaptive lasso and its oracle properties. Journal of the American Statistical Association, 101, 1418–1429.CrossRef Zou, H. (2006). The adaptive lasso and its oracle properties. Journal of the American Statistical Association, 101, 1418–1429.CrossRef
go back to reference Zou, H., & Hastie, T. (2005). Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society, Series B, 67, 301–320.CrossRef Zou, H., & Hastie, T. (2005). Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society, Series B, 67, 301–320.CrossRef
go back to reference Zou, H., & Li, R. (2008). One-step sparse estimates in nonconcave penalized likelihood models. Annals of Statistics, 36, 1509–1533.CrossRef Zou, H., & Li, R. (2008). One-step sparse estimates in nonconcave penalized likelihood models. Annals of Statistics, 36, 1509–1533.CrossRef
go back to reference Zou, H., & Zhang, H. H. (2009). On the adaptive elastic-net with a diverging number of parameters. Annals of Statistics, 37, 1733–1751.CrossRef Zou, H., & Zhang, H. H. (2009). On the adaptive elastic-net with a diverging number of parameters. Annals of Statistics, 37, 1733–1751.CrossRef
Metadata
Title
Model Selection for High-Dimensional Problems
Authors
Jing-Zhi Huang
Zhan Shi
Wei Zhong
Copyright Year
2015
Publisher
Springer New York
DOI
https://doi.org/10.1007/978-1-4614-7750-1_77