Skip to main content
Top

2017 | OriginalPaper | Chapter

Online Regression with Controlled Label Noise Rate

Authors : Edward Moroshko, Koby Crammer

Published in: Machine Learning and Knowledge Discovery in Databases

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Many online regression (and adaptive filtering) algorithms are linear, use additive update and designed for the noise-free setting. We consider the practical setting where the algorithm’s feedback is noisy, rather than a clean label. We propose a new family of algorithms which modifies the learning rate based on the noise-variance of the feedback (labels), by shrinking both inputs and feedbacks, based on the amount of noise per input instance. We consider both settings, where the noise is either given or estimated. Empirical study with both synthetic and real-world speech data shows that our algorithms improve the overall performance of the regressor, even when there is no additional explicit information (i.e. amount of noise). We also consider a more general setting where an algorithm can sample more than single (noisy) label, yet there is a total (or average) budget for the feedback. We propose a few strategies how to effectively spend the given budget, which are based on noise-variance estimation and our shrinkage rule. We show empirically that our approach outperforms other naive approaches.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Benesty, J., Rey, H., Vega, L.R., Tressens, S.: A nonparametric VSS NLMS algorithm. IEEE Sig. Process. Lett. 13(10), 581–584 (2006)CrossRef Benesty, J., Rey, H., Vega, L.R., Tressens, S.: A nonparametric VSS NLMS algorithm. IEEE Sig. Process. Lett. 13(10), 581–584 (2006)CrossRef
2.
go back to reference Bershad, N.J.: Analysis of the normalized LMS algorithm with Gaussian inputs. IEEE Trans. Acoust. Speech Sig. Process. 34(4), 793–806 (1986)CrossRef Bershad, N.J.: Analysis of the normalized LMS algorithm with Gaussian inputs. IEEE Trans. Acoust. Speech Sig. Process. 34(4), 793–806 (1986)CrossRef
3.
go back to reference Cesa-Bianchi, N., Gentile, C., Zaniboni, L.: Worst-case analysis of selective sampling for linear classification. J. Mach. Learn. Res. 7, 1205–1230 (2006)MathSciNetMATH Cesa-Bianchi, N., Gentile, C., Zaniboni, L.: Worst-case analysis of selective sampling for linear classification. J. Mach. Learn. Res. 7, 1205–1230 (2006)MathSciNetMATH
4.
go back to reference Cesa-Bianchi, N., Long, P.M., Warmuth, M.K.: Worst case quadratic loss bounds for on-line prediction of linear functions by gradient descent. Technical Report IR-418, University of California, Santa Cruz, CA, USA (1993) Cesa-Bianchi, N., Long, P.M., Warmuth, M.K.: Worst case quadratic loss bounds for on-line prediction of linear functions by gradient descent. Technical Report IR-418, University of California, Santa Cruz, CA, USA (1993)
5.
go back to reference Cesa-Bianchi, N., Shalev-Shwartz, S., Shamir, O.: Online learning of noisy data. IEEE Trans. Inf. Theory 57(12), 7907–7931 (2011)MathSciNetCrossRefMATH Cesa-Bianchi, N., Shalev-Shwartz, S., Shamir, O.: Online learning of noisy data. IEEE Trans. Inf. Theory 57(12), 7907–7931 (2011)MathSciNetCrossRefMATH
6.
go back to reference Chawla, S., Dwork, C., McSherry, F., Smith, A.D., Wee, H.: Toward privacy in public databases. In: Proceedings of Theory of Cryptography, Second Theory of Cryptography Conference, TCC 2005, Cambridge, MA, USA, 10–12 February 2005, pp. 363–385 (2005) Chawla, S., Dwork, C., McSherry, F., Smith, A.D., Wee, H.: Toward privacy in public databases. In: Proceedings of Theory of Cryptography, Second Theory of Cryptography Conference, TCC 2005, Cambridge, MA, USA, 10–12 February 2005, pp. 363–385 (2005)
7.
go back to reference Ciochină, S., Paleologu, C., Benesty, J.: An optimized NLMS algorithm for system identification. Sig. Process. 118(C), 115–121 (2016)CrossRef Ciochină, S., Paleologu, C., Benesty, J.: An optimized NLMS algorithm for system identification. Sig. Process. 118(C), 115–121 (2016)CrossRef
8.
go back to reference Crammer, K., Dekel, O., Keshet, J., Shalev-Shwartz, S., Singer, Y.: Online passive-aggressive algorithms. JMLR 7, 551–585 (2006)MathSciNetMATH Crammer, K., Dekel, O., Keshet, J., Shalev-Shwartz, S., Singer, Y.: Online passive-aggressive algorithms. JMLR 7, 551–585 (2006)MathSciNetMATH
9.
go back to reference Crammer, K., Kulesza, A., Dredze, M.: Adaptive regularization of weighted vectors. In: Advances in Neural Information Processing Systems, vol. 23 (2009) Crammer, K., Kulesza, A., Dredze, M.: Adaptive regularization of weighted vectors. In: Advances in Neural Information Processing Systems, vol. 23 (2009)
10.
go back to reference Dredze, M., Crammer, K., Pereira, F.: Confidence-weighted linear classification. In: International Conference on Machine Learning (2008) Dredze, M., Crammer, K., Pereira, F.: Confidence-weighted linear classification. In: International Conference on Machine Learning (2008)
12.
go back to reference Gollamudi, S., Nagaraj, S., Kapoor, S., Huang, Y.F.: Set-membership filtering and a set-membership normalized LMS algorithm with an adaptive step size. IEEE Sig. Process. Lett. 5(5), 111–114 (1998)CrossRef Gollamudi, S., Nagaraj, S., Kapoor, S., Huang, Y.F.: Set-membership filtering and a set-membership normalized LMS algorithm with an adaptive step size. IEEE Sig. Process. Lett. 5(5), 111–114 (1998)CrossRef
13.
go back to reference Kivinen, J., Warmuth, M.K.: Exponential gradient versus gradient descent for linear predictors. Inf. Comput. 132, 132–163 (1997)CrossRefMATH Kivinen, J., Warmuth, M.K.: Exponential gradient versus gradient descent for linear predictors. Inf. Comput. 132, 132–163 (1997)CrossRefMATH
14.
go back to reference Liu, T., Tao, D.: Classification with noisy labels by importance reweighting. IEEE Trans. Pattern Anal. Mach. Intell. 38(3), 447–461 (2016)MathSciNetCrossRef Liu, T., Tao, D.: Classification with noisy labels by importance reweighting. IEEE Trans. Pattern Anal. Mach. Intell. 38(3), 447–461 (2016)MathSciNetCrossRef
15.
go back to reference Moroshko, E., Crammer, K.: Weighted last-step min-max algorithm with improved sub-logarithmic regret. Theor. Comput. Sci. 558, 107–124 (2014)MathSciNetCrossRefMATH Moroshko, E., Crammer, K.: Weighted last-step min-max algorithm with improved sub-logarithmic regret. Theor. Comput. Sci. 558, 107–124 (2014)MathSciNetCrossRefMATH
16.
go back to reference Natarajan, N., Dhillon, I.S., Ravikumar, P.K., Tewari, A.: Learning with noisy labels. Adv. Neural Inf. Process. Syst. 26, 1196–1204 (2013) Natarajan, N., Dhillon, I.S., Ravikumar, P.K., Tewari, A.: Learning with noisy labels. Adv. Neural Inf. Process. Syst. 26, 1196–1204 (2013)
17.
go back to reference Paleologu, C., Ciochină, S., Benesty, J., Grant, S.L.: An overview on optimized NLMS algorithms for acoustic echo cancellation. EURASIP J. Appl. Sig. Process. 2015, 97 (2015)CrossRef Paleologu, C., Ciochină, S., Benesty, J., Grant, S.L.: An overview on optimized NLMS algorithms for acoustic echo cancellation. EURASIP J. Appl. Sig. Process. 2015, 97 (2015)CrossRef
18.
go back to reference Paleologu, C., Ciochina, S., Benesty, J.: Variable step-size NLMS algorithm for under-modeling acoustic echo cancellation. IEEE Sig. Process. Lett. 15, 5–8 (2008)CrossRef Paleologu, C., Ciochina, S., Benesty, J.: Variable step-size NLMS algorithm for under-modeling acoustic echo cancellation. IEEE Sig. Process. Lett. 15, 5–8 (2008)CrossRef
20.
go back to reference Vaits, N., Crammer, K.: Re-adapting the regularization of weights for non-stationary regression. In: The 22nd International Conference on Algorithmic Learning Theory, ALT 2011 (2011) Vaits, N., Crammer, K.: Re-adapting the regularization of weights for non-stationary regression. In: The 22nd International Conference on Algorithmic Learning Theory, ALT 2011 (2011)
21.
go back to reference Valin, J., Collings, I.B.: Interference-normalized least mean square algorithm. IEEE Sig. Process. Lett. 14(12), 988–991 (2007)CrossRef Valin, J., Collings, I.B.: Interference-normalized least mean square algorithm. IEEE Sig. Process. Lett. 14(12), 988–991 (2007)CrossRef
22.
go back to reference Vega, L., Rey, H., Benesty, J., Tressens, S.: A new robust variable step-size NLMS algorithm. Trans. Sig. Proc. 56(5), 1878–1893 (2008)MathSciNetCrossRef Vega, L., Rey, H., Benesty, J., Tressens, S.: A new robust variable step-size NLMS algorithm. Trans. Sig. Proc. 56(5), 1878–1893 (2008)MathSciNetCrossRef
24.
go back to reference Widrow, B., Hoff Jr., M.E.: Adaptive switching circuits (1960) Widrow, B., Hoff Jr., M.E.: Adaptive switching circuits (1960)
25.
go back to reference Yu, Y., Zhao, H.: A novel variable step size NLMS algorithm based on the power estimate of the system noise. CoRR abs/1504.05323 (2015) Yu, Y., Zhao, H.: A novel variable step size NLMS algorithm based on the power estimate of the system noise. CoRR abs/1504.05323 (2015)
Metadata
Title
Online Regression with Controlled Label Noise Rate
Authors
Edward Moroshko
Koby Crammer
Copyright Year
2017
DOI
https://doi.org/10.1007/978-3-319-71246-8_22

Premium Partner