Skip to main content
Erschienen in: Optimization and Engineering 4/2016

29.09.2016

Parallel block coordinate minimization with application to group regularized regression

verfasst von: Giuseppe C. Calafiore

Erschienen in: Optimization and Engineering | Ausgabe 4/2016

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

This paper proposes a method for parallel block coordinate-wise minimization of convex functions. Each iteration involves a first phase where n independent minimizations are performed over the n variable blocks, followed by a phase where the results of the first phase are coordinated to obtain the whole variable update. Convergence of the method to the global optimum is proved for functions composed of a smooth part plus a possibly non-smooth but separable term. The method is also proved to have a linear rate of convergence, for functions that are smooth and strongly convex. The proposed algorithm can give computational advantage over the more standard serial block coordinate-wise minimization methods, when run over a parallel, multi-worker, computing architecture. The method is suitable for regularized regression problems, such as the group Lasso, group Ridge regression, and group Elastic Net. Numerical tests are run on such types of regression problems to exemplify the performance of the proposed method.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Audet C, Dennis JE Jr (2006) Mesh adaptive direct search algorithms for constrained optimization. SIAM J Optim 17(1):188–217MathSciNetCrossRefMATH Audet C, Dennis JE Jr (2006) Mesh adaptive direct search algorithms for constrained optimization. SIAM J Optim 17(1):188–217MathSciNetCrossRefMATH
Zurück zum Zitat Bradley JK, Kyrola A, Bickson D, Guestrin C (2011) Parallel coordinate descent for \(\ell _1\)-regularized loss minimization. In: 8th International conference on machine learning Bradley JK, Kyrola A, Bickson D, Guestrin C (2011) Parallel coordinate descent for \(\ell _1\)-regularized loss minimization. In: 8th International conference on machine learning
Zurück zum Zitat Calafiore G, El Ghaoui L (2014) Optimization models. Cambridge University Press, CambridgeMATH Calafiore G, El Ghaoui L (2014) Optimization models. Cambridge University Press, CambridgeMATH
Zurück zum Zitat Kelley CT (2003) Solving nonlinear equations with Newton’s method, vol 1 of Fundamentals of Algorithms. SIAM Kelley CT (2003) Solving nonlinear equations with Newton’s method, vol 1 of Fundamentals of Algorithms. SIAM
Zurück zum Zitat Li X, Zhao T, Arora R, Liu H, Hong M (2016) An improved convergence analysis of cyclic block coordinate gradient descent methods for strongly convex minimization. In: AISTATS Li X, Zhao T, Arora R, Liu H, Hong M (2016) An improved convergence analysis of cyclic block coordinate gradient descent methods for strongly convex minimization. In: AISTATS
Zurück zum Zitat Liu J, Wright SJ (2015) Asynchronous stochastic coordinate descent: parallelism and convergence properties. SIAM J Optim 25(1):352–376MathSciNetCrossRef Liu J, Wright SJ (2015) Asynchronous stochastic coordinate descent: parallelism and convergence properties. SIAM J Optim 25(1):352–376MathSciNetCrossRef
Zurück zum Zitat Luo ZQ, Tseng P (1992) On the convergence of the coordinate descent method for convex differentiable minimization. J Optim Theory Appl 72(1):7–35MathSciNetCrossRefMATH Luo ZQ, Tseng P (1992) On the convergence of the coordinate descent method for convex differentiable minimization. J Optim Theory Appl 72(1):7–35MathSciNetCrossRefMATH
Zurück zum Zitat Magnus JR, Neudecker H (1988) Matrix differential calculus with applications in statistics and econometrics. Wiley, New YorkMATH Magnus JR, Neudecker H (1988) Matrix differential calculus with applications in statistics and econometrics. Wiley, New YorkMATH
Zurück zum Zitat Meshi O, Jaakkola T, Globerson A (2012) Convergence rate analysis of map coordinate minimization algorithms. In: NIPS Meshi O, Jaakkola T, Globerson A (2012) Convergence rate analysis of map coordinate minimization algorithms. In: NIPS
Zurück zum Zitat Necoara I, Clipici D (2013) Efficient parallel coordinate descent algorithm for convex optimization problems with separable constraints: application to distributed mpc. J Process Control 23:243–253CrossRef Necoara I, Clipici D (2013) Efficient parallel coordinate descent algorithm for convex optimization problems with separable constraints: application to distributed mpc. J Process Control 23:243–253CrossRef
Zurück zum Zitat Necoara I, Clipici D (2016) Parallel random coordinate descent method for composite minimization: convergence analysis and error bounds. SIAM J Optim 26(1):197–226MathSciNetCrossRefMATH Necoara I, Clipici D (2016) Parallel random coordinate descent method for composite minimization: convergence analysis and error bounds. SIAM J Optim 26(1):197–226MathSciNetCrossRefMATH
Zurück zum Zitat Ng AY (2004) Feature selection, \(\ell _1\) vs. \(\ell _2\) regularization, and rotational invariance. In: Proceedings of the 21st international conference on machine learning Ng AY (2004) Feature selection, \(\ell _1\) vs. \(\ell _2\) regularization, and rotational invariance. In: Proceedings of the 21st international conference on machine learning
Zurück zum Zitat Nutini J, Schmidt M, Laradji IH, Friedlander M, Koepke H (2015) Coordinate descent converges faster with the Gauss-Southwell rule than random selection. In: Proceedings of the 32nd international conference on machine learning, Lille, France Nutini J, Schmidt M, Laradji IH, Friedlander M, Koepke H (2015) Coordinate descent converges faster with the Gauss-Southwell rule than random selection. In: Proceedings of the 32nd international conference on machine learning, Lille, France
Zurück zum Zitat Qin Z, Scheinberg K, Goldfarb D (2013) Efficient block-coordinate descent algorithms for the group lasso. Math Programm Comput 5(2):143–169MathSciNetCrossRefMATH Qin Z, Scheinberg K, Goldfarb D (2013) Efficient block-coordinate descent algorithms for the group lasso. Math Programm Comput 5(2):143–169MathSciNetCrossRefMATH
Zurück zum Zitat Richtárik P, Takáč M (2012) Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function. In: Mathematical programming, pp 1–38 Richtárik P, Takáč M (2012) Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function. In: Mathematical programming, pp 1–38
Zurück zum Zitat Richtárik P, Takáč M (2013) Parallel coordinate descent methods for big data optimization. ArXiv.org, 1212.0873v2 Richtárik P, Takáč M (2013) Parallel coordinate descent methods for big data optimization. ArXiv.org, 1212.0873v2
Zurück zum Zitat Stewart BT, Venkat AN, Rawlings JB, Wright SJ, Pannocchia G (2010) Cooperative distributed model predictive control. Syst Control Lett 59(8):460–469MathSciNetCrossRefMATH Stewart BT, Venkat AN, Rawlings JB, Wright SJ, Pannocchia G (2010) Cooperative distributed model predictive control. Syst Control Lett 59(8):460–469MathSciNetCrossRefMATH
Zurück zum Zitat Tibshirani R (1996) Regression shrinkage and selection via the lasso. J R Stat Soc 58(1):267–288MathSciNetMATH Tibshirani R (1996) Regression shrinkage and selection via the lasso. J R Stat Soc 58(1):267–288MathSciNetMATH
Zurück zum Zitat Tibshirani R (2011) Regression shrinkage and selection via the lasso: a retrospective. J R Stat Soc Stat Methodol 73(3):273–282MathSciNetCrossRef Tibshirani R (2011) Regression shrinkage and selection via the lasso: a retrospective. J R Stat Soc Stat Methodol 73(3):273–282MathSciNetCrossRef
Zurück zum Zitat Tseng P (2001) Convergence of a block coordinate descent method for nondifferentiable minimization. J Optim Theory Appl 109(3):475–494MathSciNetCrossRefMATH Tseng P (2001) Convergence of a block coordinate descent method for nondifferentiable minimization. J Optim Theory Appl 109(3):475–494MathSciNetCrossRefMATH
Zurück zum Zitat Tsitsiklis JN, Bertsekas D (1989) Parallel and distributed computation: numerical methods. Prentice-Hall, Upper Saddle RiverMATH Tsitsiklis JN, Bertsekas D (1989) Parallel and distributed computation: numerical methods. Prentice-Hall, Upper Saddle RiverMATH
Metadaten
Titel
Parallel block coordinate minimization with application to group regularized regression
verfasst von
Giuseppe C. Calafiore
Publikationsdatum
29.09.2016
Verlag
Springer US
Erschienen in
Optimization and Engineering / Ausgabe 4/2016
Print ISSN: 1389-4420
Elektronische ISSN: 1573-2924
DOI
https://doi.org/10.1007/s11081-016-9336-z

Weitere Artikel der Ausgabe 4/2016

Optimization and Engineering 4/2016 Zur Ausgabe

    Marktübersichten

    Die im Laufe eines Jahres in der „adhäsion“ veröffentlichten Marktübersichten helfen Anwendern verschiedenster Branchen, sich einen gezielten Überblick über Lieferantenangebote zu verschaffen.