Skip to main content
Erschienen in: Neural Processing Letters 1/2020

15.10.2019

Tuning Parameter Selection Based on Blocked \(3\times 2\) Cross-Validation for High-Dimensional Linear Regression Model

verfasst von: Xingli Yang, Yu Wang, Ruibo Wang, Mengmeng Chen, Jihong Li

Erschienen in: Neural Processing Letters | Ausgabe 1/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

In high-dimensional linear regression, selecting an appropriate tuning parameter is essential for the penalized linear models. From the perspective of the expected prediction error of the model, cross-validation methods are commonly used to select the tuning parameter in machine learning. In this paper, blocked \(3\times 2\) cross-validation (\(3\times 2\) BCV) is proposed as the tuning parameter selection method because of its small variance for the prediction error estimation. Under some weaker conditions than leave-\(n_v\)-out cross-validation, the tuning parameter selection method based on \(3\times 2\) BCV is proved to be consistent for the high-dimensional linear regression model. Furthermore, simulated and real data experiments support the theoretical results and demonstrate that the proposed method works well in several criteria about selecting the true model.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Anhänge
Nur mit Berechtigung zugänglich
Literatur
1.
Zurück zum Zitat Chen J, Chen Z (2008) Extended Bayesian information criteria for model selection with large model spaces. Biometrika 95:759–771 Chen J, Chen Z (2008) Extended Bayesian information criteria for model selection with large model spaces. Biometrika 95:759–771
2.
Zurück zum Zitat Fan J, Li R (2001) Variable selection via nonconcave penalized likelihood and its oracle properties. J Am Stat Assoc 96:1348–1360 Fan J, Li R (2001) Variable selection via nonconcave penalized likelihood and its oracle properties. J Am Stat Assoc 96:1348–1360
3.
Zurück zum Zitat Hastie T, Tibshirani R, Friedman J (2009) The elements of statistical learning: data mining, inference, and prediction. Springer, New York Hastie T, Tibshirani R, Friedman J (2009) The elements of statistical learning: data mining, inference, and prediction. Springer, New York
4.
Zurück zum Zitat Meinshausen N, Buhlmann P (2006) High-dimensional graphs and variable selection with the LASSO. Ann Stat 34(3):1436–1462 Meinshausen N, Buhlmann P (2006) High-dimensional graphs and variable selection with the LASSO. Ann Stat 34(3):1436–1462
5.
Zurück zum Zitat Ng S (2013) Variable selection in predictive regressions. Handb Econ Forecast 2B:753–789 Ng S (2013) Variable selection in predictive regressions. Handb Econ Forecast 2B:753–789
6.
Zurück zum Zitat Tibshirani R (1996) Regression shrinkage and selection via the LASSO. J R Stat Soc Ser B 58:267–288 Tibshirani R (1996) Regression shrinkage and selection via the LASSO. J R Stat Soc Ser B 58:267–288
7.
Zurück zum Zitat Zhang C (2010) Nearly unbiased variable selection under minimax concave penalty. Ann Stat 38:894–942 Zhang C (2010) Nearly unbiased variable selection under minimax concave penalty. Ann Stat 38:894–942
8.
Zurück zum Zitat Zou H, Hastie T (2005) Regularization and variable selection via the elastic net. J R Stat Soc Ser B 67:301–320 Zou H, Hastie T (2005) Regularization and variable selection via the elastic net. J R Stat Soc Ser B 67:301–320
9.
Zurück zum Zitat Zou H, Zhang HH (2009) On the adaptive elastic-net with a diverging number of parameters. Ann Stat 37(4):1733–1751 Zou H, Zhang HH (2009) On the adaptive elastic-net with a diverging number of parameters. Ann Stat 37(4):1733–1751
10.
Zurück zum Zitat Akaike H (1974) A new look at the statistical model identification. IEEE Trans Autom Control 19(6):716–723 Akaike H (1974) A new look at the statistical model identification. IEEE Trans Autom Control 19(6):716–723
11.
Zurück zum Zitat Schawarz G (1978) Estimating the dimension of a model. Ann Stat 6:461–464 Schawarz G (1978) Estimating the dimension of a model. Ann Stat 6:461–464
12.
Zurück zum Zitat Wang H, Li B, Leng C (2009) Shrinkage tuning parameter selection with a diverging number of parameters. J R Stat Soc Ser B 71:671–683 Wang H, Li B, Leng C (2009) Shrinkage tuning parameter selection with a diverging number of parameters. J R Stat Soc Ser B 71:671–683
13.
Zurück zum Zitat Alpaydin E (1999) Combined 5 \(\times \) 2 cv F test for comparing supervised classification learning algorithms. Neural Comput 11(8):1885–1892 Alpaydin E (1999) Combined 5 \(\times \) 2 cv F test for comparing supervised classification learning algorithms. Neural Comput 11(8):1885–1892
14.
Zurück zum Zitat Yang Y (2007) Consistency of cross validation for comparing regression procedures. Ann Stat 35:2450–2473 Yang Y (2007) Consistency of cross validation for comparing regression procedures. Ann Stat 35:2450–2473
15.
Zurück zum Zitat Wang Y, Wang R, Jia H, Li J (2014) Blocked \(3\times 2\) cross-validated t-test for comparing supervised classification learning algorithms. Neural Comput 26(1):208–235 Wang Y, Wang R, Jia H, Li J (2014) Blocked \(3\times 2\) cross-validated t-test for comparing supervised classification learning algorithms. Neural Comput 26(1):208–235
16.
Zurück zum Zitat Zhang Y, Yang Y (2015) Cross-validation for selecting a model selection procedure. J Econom 187(1):95–112 Zhang Y, Yang Y (2015) Cross-validation for selecting a model selection procedure. J Econom 187(1):95–112
17.
Zurück zum Zitat Dietterich T (1998) Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput 10(7):1895–1924 Dietterich T (1998) Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput 10(7):1895–1924
18.
Zurück zum Zitat Feng Y, Yu Y (2013) Consistent cross-validation for tuning parameter selection in high-dimensional variable selection. In: World statistics congress Feng Y, Yu Y (2013) Consistent cross-validation for tuning parameter selection in high-dimensional variable selection. In: World statistics congress
19.
Zurück zum Zitat Rao C, Wu Y (1989) A strongly consistent procedure for model selection in a regression problem. Biometrika 76:369–374 Rao C, Wu Y (1989) A strongly consistent procedure for model selection in a regression problem. Biometrika 76:369–374
20.
Zurück zum Zitat Wang T, Zhu L (2011) Consistent tuning parameter selection in high dimensional sparse linear regression. J Multivar Anal 102:1141–1151 Wang T, Zhu L (2011) Consistent tuning parameter selection in high dimensional sparse linear regression. J Multivar Anal 102:1141–1151
21.
Zurück zum Zitat Fan J, Guo S, Hao N (2012) Variance estimation using refitted cross-validation in ultrahigh dimensional regression. J R Stat Soc Ser B. 74(1):37–65 Fan J, Guo S, Hao N (2012) Variance estimation using refitted cross-validation in ultrahigh dimensional regression. J R Stat Soc Ser B. 74(1):37–65
22.
Zurück zum Zitat Shao J (1993) Linear model selection by cross-validation. Stat Assoc 88:486–494 Shao J (1993) Linear model selection by cross-validation. Stat Assoc 88:486–494
23.
Zurück zum Zitat Wang Y, Li J, Li Y (2017) Choosing between two classification learning algorithms based on calibrated balanced 5\(\times \) 2 cross-validated F-test. Neural Process Lett 46(1):1–13 Wang Y, Li J, Li Y (2017) Choosing between two classification learning algorithms based on calibrated balanced 5\(\times \) 2 cross-validated F-test. Neural Process Lett 46(1):1–13
24.
Zurück zum Zitat Wang R, Wang Y, Li J, Yang X, Yang J (2017) Block-regularized \(m \times 2\) cross-validated estimator of the generalization error. Neural Comput 29(2):519–544 Wang R, Wang Y, Li J, Yang X, Yang J (2017) Block-regularized \(m \times 2\) cross-validated estimator of the generalization error. Neural Comput 29(2):519–544
25.
Zurück zum Zitat Yang Y (2006) Comparing learning methods for classification. Stat Sin 16:635–657 Yang Y (2006) Comparing learning methods for classification. Stat Sin 16:635–657
26.
Zurück zum Zitat Zhang C, Huang J (2008) The sparsity and bias of the LASSO selection in high dimensional linear regression. Ann Stat 36(4):1567–1594 Zhang C, Huang J (2008) The sparsity and bias of the LASSO selection in high dimensional linear regression. Ann Stat 36(4):1567–1594
27.
Zurück zum Zitat Buza K (2014) Feedback prediction for blogs. In: Spiliopoulou M, Schmidt-Thieme L, Janning R (eds) Data analysis, machine learning and knowledge discovery. Springer International Publishing, New York, pp 145–152 Buza K (2014) Feedback prediction for blogs. In: Spiliopoulou M, Schmidt-Thieme L, Janning R (eds) Data analysis, machine learning and knowledge discovery. Springer International Publishing, New York, pp 145–152
Metadaten
Titel
Tuning Parameter Selection Based on Blocked Cross-Validation for High-Dimensional Linear Regression Model
verfasst von
Xingli Yang
Yu Wang
Ruibo Wang
Mengmeng Chen
Jihong Li
Publikationsdatum
15.10.2019
Verlag
Springer US
Erschienen in
Neural Processing Letters / Ausgabe 1/2020
Print ISSN: 1370-4621
Elektronische ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-019-10105-w

Weitere Artikel der Ausgabe 1/2020

Neural Processing Letters 1/2020 Zur Ausgabe

Neuer Inhalt