Skip to main content
Erschienen in: International Journal of Machine Learning and Cybernetics 5/2023

27.11.2022 | Original Article

SCMP-IL: an incremental learning method with super constraints on model parameters

verfasst von: Jidong Han, Zhaoying Liu, Yujian Li, Ting Zhang

Erschienen in: International Journal of Machine Learning and Cybernetics | Ausgabe 5/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Deep learning technology has played an important role in our life. Since deep learning technology relies on the neural network model, it is still plagued by the catastrophic forgetting problem, which refers to the neural network model will forget what it has learned after learning new knowledge. The neural network model learns knowledge through labeled samples, and its knowledge is stored in its parameters. Therefore, many methods try to solve this problem from the perspective of constraint parameters and stored samples. There are few ways to solve this problem from the perspective of constraining features output of neural network models. This paper proposes an incremental learning method with super constraints on model parameters. This method not only calculates the parameter similarity loss of the old and new models, but also calculates the layer output feature similarity loss of the old and new models, and finally suppresses the change of model parameters from two directions. In addition, we also propose a new strategy for selecting representative samples from dataset and tackling the imbalance between stored samples and new task samples. Finally, we utilize the neural kernel mapping support vector machine theory to increase the interpretability of the model. In order to better meet the actual situation, five sample sets with different categories and amounts were employed in experiments. Experiments show the effectiveness of our method. For example, after learning the last task, our method is at least 1.930% and 0.562% higher than other methods on the training set and test set, respectively.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Weitere Produktempfehlungen anzeigen
Anhänge
Nur mit Berechtigung zugänglich
Literatur
2.
Zurück zum Zitat Wang Q, Wu T, Zheng H, Guo G (2020) Hierarchical pyramid diverse attention networks for face recognition. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. pp 8323–8332 Wang Q, Wu T, Zheng H, Guo G (2020) Hierarchical pyramid diverse attention networks for face recognition. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. pp 8323–8332
23.
Zurück zum Zitat Vaswani A, Shazeer N, Parmar N, et al (2017) Attention is all you need. In: Advances in neural information processing systems. pp 5998–6008 Vaswani A, Shazeer N, Parmar N, et al (2017) Attention is all you need. In: Advances in neural information processing systems. pp 5998–6008
24.
Zurück zum Zitat Dosovitskiy A, Beyer L, Kolesnikov A, et al (2020) An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. Int Conf Learn Represent Dosovitskiy A, Beyer L, Kolesnikov A, et al (2020) An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. Int Conf Learn Represent
25.
Zurück zum Zitat Carion N, Massa F, Synnaeve G et al (2020) End-to-end object detection with transformers. In: Vedaldi A, Bischof H, Brox T, Frahm J-M (eds) Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics). Springer International Publishing, Cham, pp 213–229 Carion N, Massa F, Synnaeve G et al (2020) End-to-end object detection with transformers. In: Vedaldi A, Bischof H, Brox T, Frahm J-M (eds) Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics). Springer International Publishing, Cham, pp 213–229
26.
Zurück zum Zitat Touvron H, Cord M, Douze M et al (2021) Training data-efficient image transformers & distillation through attention. In: International conference on machine learning. PMLR, pp 10347–10357 Touvron H, Cord M, Douze M et al (2021) Training data-efficient image transformers & distillation through attention. In: International conference on machine learning. PMLR, pp 10347–10357
27.
Zurück zum Zitat Tolstikhin I, Houlsby N, Kolesnikov A, et al (2021) MLP-Mixer: An all-MLP Architecture for Vision. arXiv Prepr arXiv210501601 Tolstikhin I, Houlsby N, Kolesnikov A, et al (2021) MLP-Mixer: An all-MLP Architecture for Vision. arXiv Prepr arXiv210501601
30.
Zurück zum Zitat Zenke F, Poole B, Ganguli S (2017) Continual learning through synaptic intelligence. In: Precup D, Teh YW (eds) Proceedings of the 34th International Conference on Machine Learning. PMLR, pp 3987–3995 Zenke F, Poole B, Ganguli S (2017) Continual learning through synaptic intelligence. In: Precup D, Teh YW (eds) Proceedings of the 34th International Conference on Machine Learning. PMLR, pp 3987–3995
31.
Zurück zum Zitat Nguyen C V, Li Y, Bui TD, Turner RE (2018) Variational continual learning. In: 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings Nguyen C V, Li Y, Bui TD, Turner RE (2018) Variational continual learning. In: 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings
33.
Zurück zum Zitat Belouadah E, Popescu A (2019) IL2M: Class incremental learning with dual memory. In: Proceedings of the IEEE International Conference on Computer Vision. pp 583–592 Belouadah E, Popescu A (2019) IL2M: Class incremental learning with dual memory. In: Proceedings of the IEEE International Conference on Computer Vision. pp 583–592
34.
Zurück zum Zitat Shin H, Lee JK, Kim J, Kim J (2017) Continual learning with deep generative replay. In: Advances in Neural Information Processing Systems. pp 2991–3000 Shin H, Lee JK, Kim J, Kim J (2017) Continual learning with deep generative replay. In: Advances in Neural Information Processing Systems. pp 2991–3000
35.
Zurück zum Zitat He C, Wang R, Shan S, Chen X (2019) Exemplar-supported generative reproduction for class incremental learning. In: British Machine Vision Conference 2018, BMVC 2018. p 98 He C, Wang R, Shan S, Chen X (2019) Exemplar-supported generative reproduction for class incremental learning. In: British Machine Vision Conference 2018, BMVC 2018. p 98
36.
Zurück zum Zitat Hayes TL, Kafle K, Shrestha R, et al (2020) Remind your neural network to prevent catastrophic forgetting. In: European Conference on Computer Vision. Springer, pp 466–483 Hayes TL, Kafle K, Shrestha R, et al (2020) Remind your neural network to prevent catastrophic forgetting. In: European Conference on Computer Vision. Springer, pp 466–483
37.
Zurück zum Zitat Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. arXiv Prepr arXiv150302531 Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. arXiv Prepr arXiv150302531
39.
Zurück zum Zitat Yao Z, Wang Y, Long M, Wang J (2020) Unsupervised transfer learning for spatiotemporal predictive networks. In: 37th International Conference on Machine Learning, ICML 2020. PMLR, pp 10709–10719 Yao Z, Wang Y, Long M, Wang J (2020) Unsupervised transfer learning for spatiotemporal predictive networks. In: 37th International Conference on Machine Learning, ICML 2020. PMLR, pp 10709–10719
40.
Zurück zum Zitat Li Z, Hoiem D (2017) Learning without forgetting. IEEE Trans Pattern Anal Mach Intell 40:2935–2947CrossRef Li Z, Hoiem D (2017) Learning without forgetting. IEEE Trans Pattern Anal Mach Intell 40:2935–2947CrossRef
41.
Zurück zum Zitat Castro FM, Marín-Jiménez MJ, Guil N, et al (2018) End-to-end incremental learning. In: Proceedings of the European conference on computer vision (ECCV). pp 233–248 Castro FM, Marín-Jiménez MJ, Guil N, et al (2018) End-to-end incremental learning. In: Proceedings of the European conference on computer vision (ECCV). pp 233–248
42.
Zurück zum Zitat Cheraghian A, Rahman S, Fang P, et al (2021) Semantic-aware Knowledge Distillation for Few-Shot Class-Incremental Learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 2534–2543 Cheraghian A, Rahman S, Fang P, et al (2021) Semantic-aware Knowledge Distillation for Few-Shot Class-Incremental Learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 2534–2543
43.
Zurück zum Zitat Zhang J, Zhang J, Ghosh S, et al (2020) Class-incremental Learning via Deep Model Consolidation. In: 2020 IEEE Winter Conference on Applications of Computer Vision (WACV). pp 1120–1129 Zhang J, Zhang J, Ghosh S, et al (2020) Class-incremental Learning via Deep Model Consolidation. In: 2020 IEEE Winter Conference on Applications of Computer Vision (WACV). pp 1120–1129
44.
Zurück zum Zitat Wu Y, Chen Y, Wang L, et al (2019) Large scale incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 374–382 Wu Y, Chen Y, Wang L, et al (2019) Large scale incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 374–382
45.
Zurück zum Zitat Zhao B, Xiao X, Gan G, et al (2020) Maintaining discrimination and fairness in class incremental learning. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. pp 13205–13214 Zhao B, Xiao X, Gan G, et al (2020) Maintaining discrimination and fairness in class incremental learning. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. pp 13205–13214
46.
Zurück zum Zitat Belouadah E, Popescu A (2020) ScaIL: Classifier weights scaling for class incremental learning. In: Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020. pp 1255–1264 Belouadah E, Popescu A (2020) ScaIL: Classifier weights scaling for class incremental learning. In: Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020. pp 1255–1264
47.
Zurück zum Zitat Zhu F, Zhang X-Y, Liu C-L (2021) Calibration for Non-Exemplar Based Class-Incremental Learning. In: 2021 IEEE International Conference on Multimedia and Expo (ICME). IEEE, pp 1–6 Zhu F, Zhang X-Y, Liu C-L (2021) Calibration for Non-Exemplar Based Class-Incremental Learning. In: 2021 IEEE International Conference on Multimedia and Expo (ICME). IEEE, pp 1–6
49.
Zurück zum Zitat You Y, Li J, Reddi S, et al (2019) Large batch optimization for deep learning: training BERT in 76 minutes. Int Conf Learn Represent You Y, Li J, Reddi S, et al (2019) Large batch optimization for deep learning: training BERT in 76 minutes. Int Conf Learn Represent
Metadaten
Titel
SCMP-IL: an incremental learning method with super constraints on model parameters
verfasst von
Jidong Han
Zhaoying Liu
Yujian Li
Ting Zhang
Publikationsdatum
27.11.2022
Verlag
Springer Berlin Heidelberg
Erschienen in
International Journal of Machine Learning and Cybernetics / Ausgabe 5/2023
Print ISSN: 1868-8071
Elektronische ISSN: 1868-808X
DOI
https://doi.org/10.1007/s13042-022-01725-1

Weitere Artikel der Ausgabe 5/2023

International Journal of Machine Learning and Cybernetics 5/2023 Zur Ausgabe