Skip to main content
Top
Published in: Wireless Personal Communications 1/2021

08-08-2020

Eagle Eye CBVR Based on Unique Key Frame Extraction and Deep Belief Neural Network

Authors: T. Prathiba, R. Shantha Selva Kumari

Published in: Wireless Personal Communications | Issue 1/2021

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Great efforts have been paid to develop Content Based Video Retrieval (CBVR) due to the exponential growth of video datasets with several types of data such as visual, audio and metadata. This paper proposes Eagle Eye a new Content Based Video Retrieval Framework through applying new methods for the major three important parts of CBVR. A Subspace Clustering Algorithm is implemented with binary codes that are extracted from different videos. We introduce a Local Histogram based shot boundary detection algorithm to detect shot boundaries. Then Unique Key Frame Summarization algorithm with detected shots is applied. An integrated Local Binary Pattern, Coherence based filter and HOG algorithm is used for the Feature extraction and best features are selected and stored along with the video indices. The extracted features were then utilized to train a classifier. A Deep Belief Neural Network classifier trains each extracted feature of query frame and matches the best features of the query frames with other video features. Then we enhance the video retrieval by introducing video hashing method, combining low level and high-level semantic features for the elimination of repetitive video in the retrieval. We used precision, accuracy, sensitivity and specificity metrics to assess the applicability of the projected technique. Xiph.org and Youtube datasets are used for the valuation analysis. The experimental results show that the Eagle Eye provides better performance and less processing time compared to the other methods.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Hu, W., Xie, N., Li, L., Zeng, X., & Maybank, S. (2011). A survey on visual content-based video indexing and retrieval. IEEE Transactions on Systems Man and Cybernetics Part C (Applications and Reviews), 41(6), 797–819.CrossRef Hu, W., Xie, N., Li, L., Zeng, X., & Maybank, S. (2011). A survey on visual content-based video indexing and retrieval. IEEE Transactions on Systems Man and Cybernetics Part C (Applications and Reviews), 41(6), 797–819.CrossRef
2.
go back to reference Bakar, Z. A., Kassim, M., Sahroni, M. N., & Anuar, N. (2017). A survey: Framework to develop retrieval algorithms of indexing techniques on learning material. Journal of Telecommunication, Electronic and Computer Engineering, 9(2–5), 43–46. Bakar, Z. A., Kassim, M., Sahroni, M. N., & Anuar, N. (2017). A survey: Framework to develop retrieval algorithms of indexing techniques on learning material. Journal of Telecommunication, Electronic and Computer Engineering, 9(2–5), 43–46.
3.
go back to reference Pasumarthi, N., & Malleswari, L. (2016). An empirical study and comparative analysis of Content Based Image Retrieval (CBIR) techniques with various similarity measures Pasumarthi, N., & Malleswari, L. (2016). An empirical study and comparative analysis of Content Based Image Retrieval (CBIR) techniques with various similarity measures
4.
go back to reference Pentland, A. P., Picard, R. W., & Scarloff, S. (1994). Photobook: Tools for content-based manipulation of image databases. In Storage and retrieval for image and video databases II (vol. 2185, pp. 34–48). International Society for Optics and Photonics. Pentland, A. P., Picard, R. W., & Scarloff, S. (1994). Photobook: Tools for content-based manipulation of image databases. In Storage and retrieval for image and video databases II (vol. 2185, pp. 34–48). International Society for Optics and Photonics.
5.
go back to reference Deng, Y., & Manjunath, B. (1998). NeTra-V: Toward an object-based video representation. IEEE Transactions on Circuits and Systems for Video Technology, 8(5), 616–627.CrossRef Deng, Y., & Manjunath, B. (1998). NeTra-V: Toward an object-based video representation. IEEE Transactions on Circuits and Systems for Video Technology, 8(5), 616–627.CrossRef
6.
go back to reference Sclaroff, S., Taycher, L., & La Cascia, M. (1997). Imagerover: A content-based image browser for the world wide web. In 1997 Proceedings IEEE workshop on content-based access of image and video libraries (pp. 2–9) IEEE. Sclaroff, S., Taycher, L., & La Cascia, M. (1997). Imagerover: A content-based image browser for the world wide web. In 1997 Proceedings IEEE workshop on content-based access of image and video libraries (pp. 2–9) IEEE.
7.
go back to reference Hampapur, A., et al. (1997). Virage video engine. In: Storage and retrieval for image and video databases V (vol 3022, pp. 188–199). International Society for Optics and Photonics. Hampapur, A., et al. (1997). Virage video engine. In: Storage and retrieval for image and video databases V (vol 3022, pp. 188–199). International Society for Optics and Photonics.
8.
go back to reference Beebe, C. (2000). Image indexing for multiple needs. Journal of the Art Libraries Society of North America, 19(2), 16–21. Beebe, C. (2000). Image indexing for multiple needs. Journal of the Art Libraries Society of North America, 19(2), 16–21.
9.
go back to reference Rui, Y., Huang, T. S., & Mehrotra, S. (1997). Content-based image retrieval with relevance feedback in MARS. In Proceedings of international conference on image processing (vol. 2, pp. 815–818). IEEE. Rui, Y., Huang, T. S., & Mehrotra, S. (1997). Content-based image retrieval with relevance feedback in MARS. In Proceedings of international conference on image processing (vol. 2, pp. 815–818). IEEE.
10.
go back to reference Carson, C., Belongie, S., Greenspan, H., & Malik, J. (2002). Blobworld: Image segmentation using expectation-maximization and its application to image querying. IEEE Transactions on Pattern Analysis and Machine Intelligence, 8, 1026–1038.CrossRef Carson, C., Belongie, S., Greenspan, H., & Malik, J. (2002). Blobworld: Image segmentation using expectation-maximization and its application to image querying. IEEE Transactions on Pattern Analysis and Machine Intelligence, 8, 1026–1038.CrossRef
11.
go back to reference Cheikh, F. A. (2004). MUVIS-a system for content-based image retrieval. Tampere University of Technology. Cheikh, F. A. (2004). MUVIS-a system for content-based image retrieval. Tampere University of Technology.
12.
go back to reference Laaksonen, J., Koskela, M., Laakso, S., & Oja, E. (2000). PicSOM–content-based image retrieval with self-organizing maps. Pattern Recognition Letters, 21(13–14), 1199–1207.MATHCrossRef Laaksonen, J., Koskela, M., Laakso, S., & Oja, E. (2000). PicSOM–content-based image retrieval with self-organizing maps. Pattern Recognition Letters, 21(13–14), 1199–1207.MATHCrossRef
13.
go back to reference Patil, D., Potey, M. (2015). Survey of content based lecture video retrieval. International Journal of Computer Trends and Technology 19(1) Patil, D., Potey, M. (2015). Survey of content based lecture video retrieval. International Journal of Computer Trends and Technology 19(1)
14.
go back to reference Kaur, M., & Kumar, H. (2015). Keyword-based search and ranking in NPTEL lecture videos. In Proceedings of the international conference on transformations in engineering education (pp. 79–88). Springer Kaur, M., & Kumar, H. (2015). Keyword-based search and ranking in NPTEL lecture videos. In Proceedings of the international conference on transformations in engineering education (pp. 79–88). Springer
15.
go back to reference Song, J., Gao, L., Liu, L., Zhu, X., & Sebe, N. (2018). Quantization-based hashing: a general framework for scalable image and video retrieval. Pattern Recognition, 75, 175–187.CrossRef Song, J., Gao, L., Liu, L., Zhu, X., & Sebe, N. (2018). Quantization-based hashing: a general framework for scalable image and video retrieval. Pattern Recognition, 75, 175–187.CrossRef
16.
go back to reference Sun, J., Liu, X., Wan, W., Li, J., Zhao, D., & Zhang, H. J. N. (2016). Video hashing based on appearance and attention features fusion via DBN. Neurocomputing, 213, 84–94.CrossRef Sun, J., Liu, X., Wan, W., Li, J., Zhao, D., & Zhang, H. J. N. (2016). Video hashing based on appearance and attention features fusion via DBN. Neurocomputing, 213, 84–94.CrossRef
17.
go back to reference Ansari, A., & Mohammed, M. (2015). Content based video retrieval systems-methods, techniques, trends and challenges. International Journal of Computer Applications, 112(7), 13–22. Ansari, A., & Mohammed, M. (2015). Content based video retrieval systems-methods, techniques, trends and challenges. International Journal of Computer Applications, 112(7), 13–22.
18.
go back to reference Khusro, S., Ali, Z., & Ullah, I. (2016). Recommender systems: issues, challenges, and research opportunities. In: Information science and applications (ICISA) 2016 (pp. 1179–1189). Springer Khusro, S., Ali, Z., & Ullah, I. (2016). Recommender systems: issues, challenges, and research opportunities. In: Information science and applications (ICISA) 2016 (pp. 1179–1189). Springer
19.
go back to reference Fernandez-Beltran, R., & Pla, F. (2016). Latent topics-based relevance feedback for video retrieval. Pattern Recognition, 51, 72–84.CrossRef Fernandez-Beltran, R., & Pla, F. (2016). Latent topics-based relevance feedback for video retrieval. Pattern Recognition, 51, 72–84.CrossRef
20.
go back to reference Guo, J.-M., & Prasetyo, H. (2015). Content-based image retrieval using features extracted from halftoning-based block truncation coding. IEEE Transactions on Image Processing, 24(3), 1010–1024.MathSciNetMATHCrossRef Guo, J.-M., & Prasetyo, H. (2015). Content-based image retrieval using features extracted from halftoning-based block truncation coding. IEEE Transactions on Image Processing, 24(3), 1010–1024.MathSciNetMATHCrossRef
21.
go back to reference Basu, S., Yu, Y., & Zimmermann, R. (2016). Fuzzy clustering of lecture videos based on topic modeling. In: 2016 14th international workshop on content-based multimedia indexing (CBMI) (pp. 1–6). IEEE. Basu, S., Yu, Y., & Zimmermann, R. (2016). Fuzzy clustering of lecture videos based on topic modeling. In: 2016 14th international workshop on content-based multimedia indexing (CBMI) (pp. 1–6). IEEE.
22.
go back to reference Song, J., Zhang, H., Li, X., Gao, L., Wang, M., & Hong, R. (2018). Self-supervised video hashing with hierarchical binary auto-encoder. IEEE Transactions on Image Processing, 27(7), 3210–3221.MathSciNetMATHCrossRef Song, J., Zhang, H., Li, X., Gao, L., Wang, M., & Hong, R. (2018). Self-supervised video hashing with hierarchical binary auto-encoder. IEEE Transactions on Image Processing, 27(7), 3210–3221.MathSciNetMATHCrossRef
23.
go back to reference Ye, Y., Zhao, Z., Li, Y., Chen, L., Xiao, J., & Zhuang, Y. (2017). Video question answering via attribute-augmented attention network learning. In: Proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval, pp. 829–832. ACM. Ye, Y., Zhao, Z., Li, Y., Chen, L., Xiao, J., & Zhuang, Y. (2017). Video question answering via attribute-augmented attention network learning. In: Proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval, pp. 829–832. ACM.
24.
go back to reference Hauptmann, A. (2015). Fast and accurate content-based semantic search in 100M Internet Videos. In MM 2015-proceedings of the 2015 ACM multimedia conference, 2015. Hauptmann, A. (2015). Fast and accurate content-based semantic search in 100M Internet Videos. In MM 2015-proceedings of the 2015 ACM multimedia conference, 2015.
25.
go back to reference Deldjoo, Y., Elahi, M., Quadrana, M., & Cremonesi, P. (2018). Using visual features based on MPEG-7 and deep learning for movie recommendation. International Journal of Multimedia Information Retrieval, 7(4), 207–219.CrossRef Deldjoo, Y., Elahi, M., Quadrana, M., & Cremonesi, P. (2018). Using visual features based on MPEG-7 and deep learning for movie recommendation. International Journal of Multimedia Information Retrieval, 7(4), 207–219.CrossRef
26.
go back to reference Liong, V. E., Lu, J., Tan, Y.-P., & Zhou, J. (2016). Deep video hashing. IEEE Transactions on Multimedia, 19(6), 1209–1219.CrossRef Liong, V. E., Lu, J., Tan, Y.-P., & Zhou, J. (2016). Deep video hashing. IEEE Transactions on Multimedia, 19(6), 1209–1219.CrossRef
27.
go back to reference Thakre, K., Rajurkar, A., & Manthalkar, R. (2016). Video partitioning and secured keyframe extraction of MPEG video. Procedia Computer Science, 78, 790–798.CrossRef Thakre, K., Rajurkar, A., & Manthalkar, R. (2016). Video partitioning and secured keyframe extraction of MPEG video. Procedia Computer Science, 78, 790–798.CrossRef
28.
go back to reference Guo, H., Wang, J., & Lu, H. (2016). Multiple deep features learning for object retrieval in surveillance videos. IET Computer Vision, 10(4), 268–272.CrossRef Guo, H., Wang, J., & Lu, H. (2016). Multiple deep features learning for object retrieval in surveillance videos. IET Computer Vision, 10(4), 268–272.CrossRef
29.
go back to reference Younus, Z. S., et al. (2015). Content-based image retrieval using PSO and k-means clustering algorithm. Arabian Journal of Geosciences, 8(8), 6211–6224.CrossRef Younus, Z. S., et al. (2015). Content-based image retrieval using PSO and k-means clustering algorithm. Arabian Journal of Geosciences, 8(8), 6211–6224.CrossRef
30.
go back to reference Lu, Y., et al. (2016). GeoUGV: user-generated mobile video dataset with fine granularity spatial metadata. In: Proceedings of the 7th international conference on multimedia systems, 2016. ACM. Lu, Y., et al. (2016). GeoUGV: user-generated mobile video dataset with fine granularity spatial metadata. In: Proceedings of the 7th international conference on multimedia systems, 2016. ACM.
31.
go back to reference Bui, T., & Collomosse, J. (2015). Scalable sketch-based image retrieval using color gradient features. In: Proceedings of the IEEE international conference on computer vision workshops, pp. 1–8. Bui, T., & Collomosse, J. (2015). Scalable sketch-based image retrieval using color gradient features. In: Proceedings of the IEEE international conference on computer vision workshops, pp. 1–8.
32.
go back to reference de Oliveira Barra, G., Lux, M., & Giro-i-Nieto, X. (2016). Large scale content-based video retrieval with LIvRE. In: 2016 14th international workshop on content-based multimedia indexing (CBMI) (pp. 1–4). IEEE. de Oliveira Barra, G., Lux, M., & Giro-i-Nieto, X. (2016). Large scale content-based video retrieval with LIvRE. In: 2016 14th international workshop on content-based multimedia indexing (CBMI) (pp. 1–4). IEEE.
33.
go back to reference Chakraborty, S., Tickoo, O., & Iyer, R. (2015). Adaptive keyframe selection for video summarization. In: 2015 IEEE winter conference on applications of computer vision, 2015, pp. 702–709: IEEE. Chakraborty, S., Tickoo, O., & Iyer, R. (2015). Adaptive keyframe selection for video summarization. In: 2015 IEEE winter conference on applications of computer vision, 2015, pp. 702–709: IEEE.
34.
go back to reference Schoeffmann, K., Del Fabro, M., Szkaliczki, T., Böszörmenyi, L., & Keckstein, J. (2015). Keyframe extraction in endoscopic video. Multimedia Tools and Applications, 74(24), 11187–11206.CrossRef Schoeffmann, K., Del Fabro, M., Szkaliczki, T., Böszörmenyi, L., & Keckstein, J. (2015). Keyframe extraction in endoscopic video. Multimedia Tools and Applications, 74(24), 11187–11206.CrossRef
35.
go back to reference Panagiotakis, C., Papoutsakis, K., & Argyros, A. (2018). A graph-based approach for detecting common actions in motion capture data and videos. Pattern Recognition, 79, 1–11.CrossRef Panagiotakis, C., Papoutsakis, K., & Argyros, A. (2018). A graph-based approach for detecting common actions in motion capture data and videos. Pattern Recognition, 79, 1–11.CrossRef
36.
go back to reference Ejaz, N., Baik, S. W., Majeed, H., Chang, H., & Mehmood, I. (2018). Multi-scale contrast and relative motion-based key frame extraction. Journal on Image and Video Processing, 1, 40.CrossRef Ejaz, N., Baik, S. W., Majeed, H., Chang, H., & Mehmood, I. (2018). Multi-scale contrast and relative motion-based key frame extraction. Journal on Image and Video Processing, 1, 40.CrossRef
37.
go back to reference Liu, W., Mei, T., Zhang, Y., Che, C., & Luo, J. (2015). Multi-task deep visual-semantic embedding for video thumbnail selection. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3707–3715. Liu, W., Mei, T., Zhang, Y., Che, C., & Luo, J. (2015). Multi-task deep visual-semantic embedding for video thumbnail selection. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3707–3715.
38.
go back to reference Wu, J., Zhong, S.-H., Jiang, J., & Yang, Y. (2017). A novel clustering method for static video summarization. Multimedia Tools and Applications, 76(7), 9625–9641.CrossRef Wu, J., Zhong, S.-H., Jiang, J., & Yang, Y. (2017). A novel clustering method for static video summarization. Multimedia Tools and Applications, 76(7), 9625–9641.CrossRef
39.
go back to reference Meng, J., Wang, H., Yuan, J., & Tan, Y.-P. (2016). From keyframes to key objects: Video summarization by representative object proposal selection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 1039–1048. Meng, J., Wang, H., Yuan, J., & Tan, Y.-P. (2016). From keyframes to key objects: Video summarization by representative object proposal selection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 1039–1048.
40.
go back to reference Ioannidis, A., Chasanis, V., & Likas, A. (2016). Weighted multi-view key-frame extraction. Pattern Recognition Letters., 72, 52–61.CrossRef Ioannidis, A., Chasanis, V., & Likas, A. (2016). Weighted multi-view key-frame extraction. Pattern Recognition Letters., 72, 52–61.CrossRef
41.
go back to reference Chen, A. Y., & Corso, J. J. (2010).Propagating multi-class pixel labels throughout video frames. In 2010 Western New York image processing workshop, 2010. IEEE, pp. 14–17 Chen, A. Y., & Corso, J. J. (2010).Propagating multi-class pixel labels throughout video frames. In 2010 Western New York image processing workshop, 2010. IEEE, pp. 14–17
42.
go back to reference Liu, A.-A., Su, Y.-T., Nie, W.-Z., & Kankanhalli, M. (2017). Hierarchical clustering multi-task learning for joint human action grouping and recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39(1), 102–114.CrossRef Liu, A.-A., Su, Y.-T., Nie, W.-Z., & Kankanhalli, M. (2017). Hierarchical clustering multi-task learning for joint human action grouping and recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39(1), 102–114.CrossRef
43.
go back to reference Wang, J., Liu, Z., Wu, Y., & Yuan, J. (2012). Mining actionlet ensemble for action recognition with depth cameras. In: 2012 IEEE conference on computer vision and pattern recognition, 2012. IEEE, pp. 1290–1297 Wang, J., Liu, Z., Wu, Y., & Yuan, J. (2012). Mining actionlet ensemble for action recognition with depth cameras. In: 2012 IEEE conference on computer vision and pattern recognition, 2012. IEEE, pp. 1290–1297
44.
go back to reference Ullah, A., Ahmad, J., Muhammad, K., Sajjad, M., & Baik, S. W. J. I. A. (2018). Action recognition in video sequences using deep bi-directional LSTM with CNN features. IEEE Access, 6, 1155–1166.CrossRef Ullah, A., Ahmad, J., Muhammad, K., Sajjad, M., & Baik, S. W. J. I. A. (2018). Action recognition in video sequences using deep bi-directional LSTM with CNN features. IEEE Access, 6, 1155–1166.CrossRef
45.
go back to reference Sandeep, R., Sharma, S., Thakur, M., & Bora, P. K. (2016). Perceptual video hashing based on Tucker decomposition with application to indexing and retrieval of near-identical videos. Multimedia Tools and Applications, 75(13), 7779–7797.CrossRef Sandeep, R., Sharma, S., Thakur, M., & Bora, P. K. (2016). Perceptual video hashing based on Tucker decomposition with application to indexing and retrieval of near-identical videos. Multimedia Tools and Applications, 75(13), 7779–7797.CrossRef
Metadata
Title
Eagle Eye CBVR Based on Unique Key Frame Extraction and Deep Belief Neural Network
Authors
T. Prathiba
R. Shantha Selva Kumari
Publication date
08-08-2020
Publisher
Springer US
Published in
Wireless Personal Communications / Issue 1/2021
Print ISSN: 0929-6212
Electronic ISSN: 1572-834X
DOI
https://doi.org/10.1007/s11277-020-07721-4

Other articles of this Issue 1/2021

Wireless Personal Communications 1/2021 Go to the issue