Skip to main content
Erschienen in: Neural Processing Letters 1/2020

04.07.2019

Non-iterative Knowledge Fusion in Deep Convolutional Neural Networks

verfasst von: Mikhail Iu. Leontev, Viktoriia Islenteva, Sergey V. Sukhov

Erschienen in: Neural Processing Letters | Ausgabe 1/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Incorporation of new knowledge into neural networks with simultaneous preservation of the previous knowledge is known to be a nontrivial problem. This problem becomes even more complex when the new knowledge is contained not in new training examples, but inside the parameters (e.g., connection weights) of another neural network. In this correspondence, we propose and test two methods of combining knowledge contained in separate networks. The first method is based on a summation of weights. The second incorporates new knowledge by modification of weights nonessential for the preservation of previously stored information. We show that with these methods, the knowledge can be transferred non-iteratively from one network to another without requiring additional training sessions. The fused network operates efficiently, performing classification at a level similar to that of an ensemble of networks. The efficiency of the methods is quantified on several publicly available data sets in classification tasks both for shallow and deep feedforward neural networks.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Anhänge
Nur mit Berechtigung zugänglich
Literatur
2.
Zurück zum Zitat Yu J, Yang X, Gao F, Tao D (2016) Deep multimodal distance metric learning using click constraints for image ranking. IEEE Trans Cybern 47:4014–4024 Yu J, Yang X, Gao F, Tao D (2016) Deep multimodal distance metric learning using click constraints for image ranking. IEEE Trans Cybern 47:4014–4024
6.
Zurück zum Zitat Li Z, Hoiem D (2016) Learning without forgetting. In: Proceedings of the European conference on computer vision (ECCV), pp 614–629 Li Z, Hoiem D (2016) Learning without forgetting. In: Proceedings of the European conference on computer vision (ECCV), pp 614–629
7.
Zurück zum Zitat Donahue J, Jia Y, Vinyals O, et al (2014) DeCAF: a deep convolutional activation feature for generic visual recognition. In: Xing EP, Jebara T (eds) Proceedings of the 31st international conference on machine learning, PMLR, Bejing, China, pp 647–655 Donahue J, Jia Y, Vinyals O, et al (2014) DeCAF: a deep convolutional activation feature for generic visual recognition. In: Xing EP, Jebara T (eds) Proceedings of the 31st international conference on machine learning, PMLR, Bejing, China, pp 647–655
8.
Zurück zum Zitat Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, pp 580–587 Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, pp 580–587
9.
Zurück zum Zitat French RM, Ans B, Rousset S (2001) Pseudopatterns and dual‐network memory models: advantages and shortcomings. In: French R, Sougné J (eds) Connectionist models of learning, development and evolution. Springer, London, pp 13–22 French RM, Ans B, Rousset S (2001) Pseudopatterns and dual‐network memory models: advantages and shortcomings. In: French R, Sougné J (eds) Connectionist models of learning, development and evolution. Springer, London, pp 13–22
11.
Zurück zum Zitat Dietterich TG (2000) Ensemble methods in machine learning. In: Multiple classifier systems, pp 1–15 Dietterich TG (2000) Ensemble methods in machine learning. In: Multiple classifier systems, pp 1–15
13.
Zurück zum Zitat Domingos P (2000) Bayesian averaging of classifiers and the overfitting problem. In: 17th International conference on machine learning, San Francisco, pp 223–230 Domingos P (2000) Bayesian averaging of classifiers and the overfitting problem. In: 17th International conference on machine learning, San Francisco, pp 223–230
16.
Zurück zum Zitat Fujii T, Asama H, Fujita T, et al (1996) Knowledge sharing among multiple autonomous mobile robots through indirect communication using intelligent data carriers. In: Proceedings of IEEE/RSJ international conference on intelligent robots and systems, IROS’96. IEEE, pp 1466–1471 Fujii T, Asama H, Fujita T, et al (1996) Knowledge sharing among multiple autonomous mobile robots through indirect communication using intelligent data carriers. In: Proceedings of IEEE/RSJ international conference on intelligent robots and systems, IROS’96. IEEE, pp 1466–1471
17.
Zurück zum Zitat Yu J, Kuang Z, Zhang B et al (2018) Leveraging content sensitiveness and user trustworthiness to recommend fine-grained privacy settings for social image sharing. IEEE Trans Inf Forensics Secur 13:1317–1332 Yu J, Kuang Z, Zhang B et al (2018) Leveraging content sensitiveness and user trustworthiness to recommend fine-grained privacy settings for social image sharing. IEEE Trans Inf Forensics Secur 13:1317–1332
18.
Zurück zum Zitat Paul R, Hawkins SH, Balagurunathan Y et al (2016) Deep feature transfer learning in combination with traditional features predicts survival among patients with lung adenocarcinoma. Tomography 2:388 Paul R, Hawkins SH, Balagurunathan Y et al (2016) Deep feature transfer learning in combination with traditional features predicts survival among patients with lung adenocarcinoma. Tomography 2:388
19.
Zurück zum Zitat Zeng X, Martinez TR (2000) Using a neural network to approximate an ensemble of classifiers. Neural Process Lett 12:225–237 Zeng X, Martinez TR (2000) Using a neural network to approximate an ensemble of classifiers. Neural Process Lett 12:225–237
20.
Zurück zum Zitat Buciluǎ C, Caruana R, Niculescu-Mizil A (2006) Model compression. In: Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining - KDD’06. ACM Press, New York, New York, USA, p 535 Buciluǎ C, Caruana R, Niculescu-Mizil A (2006) Model compression. In: Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining - KDD’06. ACM Press, New York, New York, USA, p 535
23.
Zurück zum Zitat Alexandra C, Cristea P, Okamoto T (1997) Neural network knowledge extraction. Rev Roum des Sci Tech Ser EE (Electrotechn Energ) 42:477–491 Alexandra C, Cristea P, Okamoto T (1997) Neural network knowledge extraction. Rev Roum des Sci Tech Ser EE (Electrotechn Energ) 42:477–491
24.
Zurück zum Zitat Towell GG, Shavlik JW (1993) Extracting refined rules from knowledge-based neural networks. Mach Learn 13:71–101 Towell GG, Shavlik JW (1993) Extracting refined rules from knowledge-based neural networks. Mach Learn 13:71–101
25.
Zurück zum Zitat Kolman E, Margaliot M (2005) Are artificial neural networks white boxes? IEEE Trans Neural Netw 16:844–852 Kolman E, Margaliot M (2005) Are artificial neural networks white boxes? IEEE Trans Neural Netw 16:844–852
26.
Zurück zum Zitat Mantas CJ (2008) A generic fuzzy aggregation operator: rules extraction from and insertion into artificial neural networks. Soft Comput 12:493–514 Mantas CJ (2008) A generic fuzzy aggregation operator: rules extraction from and insertion into artificial neural networks. Soft Comput 12:493–514
27.
Zurück zum Zitat Hruschka ER, Ebecken NFF (2006) Extracting rules from multilayer perceptrons in classification problems: a clustering-based approach. Neurocomputing 70:384–397 Hruschka ER, Ebecken NFF (2006) Extracting rules from multilayer perceptrons in classification problems: a clustering-based approach. Neurocomputing 70:384–397
28.
Zurück zum Zitat McGarry KJ, MacIntyre J (1999) Knowledge extraction and insertion from radial basis function networks. In: IEE Colloquium on Applied Statistical Pattern Recognition (Ref. No. 1999/063), pp 15/1–15/6 McGarry KJ, MacIntyre J (1999) Knowledge extraction and insertion from radial basis function networks. In: IEE Colloquium on Applied Statistical Pattern Recognition (Ref. No. 1999/063), pp 15/1–15/6
29.
Zurück zum Zitat Kasabov N, Woodford B (1999) Rule insertion and rule extraction from evolving fuzzy neural networks: algorithms and applications for building adaptive, intelligent expert systems. In: Proceedings of the FUZZ-IEEE, pp 1406–1411 Kasabov N, Woodford B (1999) Rule insertion and rule extraction from evolving fuzzy neural networks: algorithms and applications for building adaptive, intelligent expert systems. In: Proceedings of the FUZZ-IEEE, pp 1406–1411
30.
Zurück zum Zitat Tran SN, Garcez ADA (2013) Knowledge extraction from deep belief networks for images. In: IJCAI-2013 workshop on neural-symbolic learning and reasoning Tran SN, Garcez ADA (2013) Knowledge extraction from deep belief networks for images. In: IJCAI-2013 workshop on neural-symbolic learning and reasoning
31.
Zurück zum Zitat Tran SN, Garcez ADA (2018) Deep logic networks: inserting and extracting knowledge from deep belief networks. IEEE Trans Neural Netw Learn Syst 29:246–258 Tran SN, Garcez ADA (2018) Deep logic networks: inserting and extracting knowledge from deep belief networks. IEEE Trans Neural Netw Learn Syst 29:246–258
32.
Zurück zum Zitat Utans J (1996) Weight averaging for neural networks and local resampling schemes. In: AAAI-96 Workshop on integrating multiple learned models, pp 133–138 Utans J (1996) Weight averaging for neural networks and local resampling schemes. In: AAAI-96 Workshop on integrating multiple learned models, pp 133–138
33.
Zurück zum Zitat Smith J, Gashler M (2017) An investigation of how neural networks learn from the experiences of peers through periodic weight averaging. In: 2017 16th IEEE international conference on machine learning and applications (ICMLA). IEEE, pp 731–736 Smith J, Gashler M (2017) An investigation of how neural networks learn from the experiences of peers through periodic weight averaging. In: 2017 16th IEEE international conference on machine learning and applications (ICMLA). IEEE, pp 731–736
36.
Zurück zum Zitat Thrun S, Pratt L (1998) Learning to learn. Springer, Berlin Thrun S, Pratt L (1998) Learning to learn. Springer, Berlin
37.
Zurück zum Zitat Hu J, Lu J, Tan Y-P (2015) Deep transfer metric learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 325–333 Hu J, Lu J, Tan Y-P (2015) Deep transfer metric learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 325–333
38.
Zurück zum Zitat Wu Y, Ji Q (2016) Constrained deep transfer feature learning and its applications. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5101–5109 Wu Y, Ji Q (2016) Constrained deep transfer feature learning and its applications. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5101–5109
39.
Zurück zum Zitat Huang G-B, Wang DH, Lan Y (2011) Extreme learning machines: a survey. Int J Mach Learn Cybern 2:107–122 Huang G-B, Wang DH, Lan Y (2011) Extreme learning machines: a survey. Int J Mach Learn Cybern 2:107–122
40.
Zurück zum Zitat Krizhevsky A, Sutskever I, Hinton GE (2012) ImageNet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097–1105 Krizhevsky A, Sutskever I, Hinton GE (2012) ImageNet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097–1105
41.
Zurück zum Zitat Blundell C, Cornebise J, Kavukcuoglu K, Wierstra D (2015) Weight uncertainty in neural networks. In: Proceedings of the 32nd international conference on international conference on machine learning-volume 37, JMLR.org, pp 1613–1622 Blundell C, Cornebise J, Kavukcuoglu K, Wierstra D (2015) Weight uncertainty in neural networks. In: Proceedings of the 32nd international conference on international conference on machine learning-volume 37, JMLR.org, pp 1613–1622
42.
Zurück zum Zitat Schmidt WF, Kraaijveld MA, Duin RPW (1992) Feedforward neural networks with random weights. In: Proceedings 11th IAPR international conference on pattern recognition, Vol. II. Conference B: pattern recognition methodology and systems. IEEE Computer Society Press, pp 1–4 Schmidt WF, Kraaijveld MA, Duin RPW (1992) Feedforward neural networks with random weights. In: Proceedings 11th IAPR international conference on pattern recognition, Vol. II. Conference B: pattern recognition methodology and systems. IEEE Computer Society Press, pp 1–4
43.
Zurück zum Zitat Bellido I, Fiesler E (1993) Do backpropagation trained neural networks have normal weight distributions? In: ICANN’93. Springer, London, pp 772–775 Bellido I, Fiesler E (1993) Do backpropagation trained neural networks have normal weight distributions? In: ICANN’93. Springer, London, pp 772–775
44.
Zurück zum Zitat French RM, Chater N (2002) Using noise to compute error surfaces in connectionist networks: a novel means of reducing catastrophic forgetting. Neural Comput 14:1755–1769 French RM, Chater N (2002) Using noise to compute error surfaces in connectionist networks: a novel means of reducing catastrophic forgetting. Neural Comput 14:1755–1769
45.
Zurück zum Zitat Kirkpatrick J, Pascanu R, Rabinowitz N et al (2017) Overcoming catastrophic forgetting in neural networks. Proc Natl Acad Sci 114:3521–3526 Kirkpatrick J, Pascanu R, Rabinowitz N et al (2017) Overcoming catastrophic forgetting in neural networks. Proc Natl Acad Sci 114:3521–3526
48.
Zurück zum Zitat Ashmore S, Gashler M (2015) A method for finding similarity between multi-layer perceptrons by Forward Bipartite Alignment. In: 2015 International joint conference on neural networks (IJCNN), IEEE, pp 1–7 Ashmore S, Gashler M (2015) A method for finding similarity between multi-layer perceptrons by Forward Bipartite Alignment. In: 2015 International joint conference on neural networks (IJCNN), IEEE, pp 1–7
49.
Zurück zum Zitat Zhang J, Yu J, Tao D (2018) Local deep-feature alignment for unsupervised dimension reduction. IEEE Trans Image Process 27:2420–2432 Zhang J, Yu J, Tao D (2018) Local deep-feature alignment for unsupervised dimension reduction. IEEE Trans Image Process 27:2420–2432
50.
Zurück zum Zitat Kuhn HW (1955) The Hungarian algorithm for the assignment problem. Nav Res Logist Q 2:83–97 Kuhn HW (1955) The Hungarian algorithm for the assignment problem. Nav Res Logist Q 2:83–97
54.
Zurück zum Zitat Breiman L (1996) Bagging predictors. Mach Learn 24:123–140 Breiman L (1996) Bagging predictors. Mach Learn 24:123–140
55.
Zurück zum Zitat Freund Y (1995) Boosting a weak learning algorithm by majority. Inf Comput 121:256–285 Freund Y (1995) Boosting a weak learning algorithm by majority. Inf Comput 121:256–285
56.
Zurück zum Zitat Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images. Technical Report, University of Toronto 1:7 Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images. Technical Report, University of Toronto 1:7
58.
Zurück zum Zitat Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv Preprint arXiv:1409.1556 Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv Preprint arXiv:​1409.​1556
Metadaten
Titel
Non-iterative Knowledge Fusion in Deep Convolutional Neural Networks
verfasst von
Mikhail Iu. Leontev
Viktoriia Islenteva
Sergey V. Sukhov
Publikationsdatum
04.07.2019
Verlag
Springer US
Erschienen in
Neural Processing Letters / Ausgabe 1/2020
Print ISSN: 1370-4621
Elektronische ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-019-10074-0

Weitere Artikel der Ausgabe 1/2020

Neural Processing Letters 1/2020 Zur Ausgabe

Neuer Inhalt