Skip to main content

2016 | OriginalPaper | Buchkapitel

Ego2Top: Matching Viewers in Egocentric and Top-View Videos

verfasst von : Shervin Ardeshir, Ali Borji

Erschienen in: Computer Vision – ECCV 2016

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Egocentric cameras are becoming increasingly popular and provide us with large amounts of videos, captured from the first person perspective. At the same time, surveillance cameras and drones offer an abundance of visual information, often captured from top-view. Although these two sources of information have been separately studied in the past, they have not been collectively studied and related. Having a set of egocentric cameras and a top-view camera capturing the same area, we propose a framework to identify the egocentric viewers in the top-view video. We utilize two types of features for our assignment procedure. Unary features encode what a viewer (seen from top-view or recording an egocentric video) visually experiences over time. Pairwise features encode the relationship between the visual content of a pair of viewers. Modeling each view (egocentric or top) by a graph, the assignment process is formulated as spectral graph matching. Evaluating our method over a dataset of 50 top-view and 188 egocentric videos taken in different scenarios demonstrates the efficiency of the proposed approach in assigning egocentric viewers to identities present in top-view camera. We also study the effect of different parameters such as the number of egocentric viewers and visual features.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Fathi, A., Farhadi, A., Rehg, J.: Understanding egocentric activities. In: 2011 IEEE International Conference on Computer Vision (ICCV). IEEE (2011) Fathi, A., Farhadi, A., Rehg, J.: Understanding egocentric activities. In: 2011 IEEE International Conference on Computer Vision (ICCV). IEEE (2011)
2.
Zurück zum Zitat Fathi, A., Li, Y., Rehg, J.M.: Learning to recognize daily actions using gaze. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7572, pp. 314–327. Springer, Heidelberg (2012). doi:10.1007/978-3-642-33718-5_23 Fathi, A., Li, Y., Rehg, J.M.: Learning to recognize daily actions using gaze. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7572, pp. 314–327. Springer, Heidelberg (2012). doi:10.​1007/​978-3-642-33718-5_​23
3.
Zurück zum Zitat Bettadapura, V., Essa, I., Pantofaru, C.: Egocentric field-of-view localization using first-person point-of-view devices. In: IEEE Winter Conference on Applications of Computer Vision (WACV) (2015) Bettadapura, V., Essa, I., Pantofaru, C.: Egocentric field-of-view localization using first-person point-of-view devices. In: IEEE Winter Conference on Applications of Computer Vision (WACV) (2015)
4.
Zurück zum Zitat Egozi, A., Keller, Y., Guterman, H.: A probabilistic approach to spectral graph matching. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 18–27 (2013)CrossRef Egozi, A., Keller, Y., Guterman, H.: A probabilistic approach to spectral graph matching. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 18–27 (2013)CrossRef
5.
Zurück zum Zitat Dicle, C., Campsm, O., Sznaier., M.: The way they move: tracking multiple targets with similar appearance. In: Proceedings of the IEEE International Conference on Computer Vision (2013) Dicle, C., Campsm, O., Sznaier., M.: The way they move: tracking multiple targets with similar appearance. In: Proceedings of the IEEE International Conference on Computer Vision (2013)
6.
Zurück zum Zitat Kanade, T., Hebert, M.: First-person vision. Proc. IEEE 100(8), 2442–2453 (2012)CrossRef Kanade, T., Hebert, M.: First-person vision. Proc. IEEE 100(8), 2442–2453 (2012)CrossRef
7.
Zurück zum Zitat Betancourt, A., Morerio, P., Regazzoni, C.S., Rauterberg, M.: The evolution of first person vision methods: a survey. IEEE Trans. Circ. Syst. Video Technol. 25(5), 744–760 (2015)CrossRef Betancourt, A., Morerio, P., Regazzoni, C.S., Rauterberg, M.: The evolution of first person vision methods: a survey. IEEE Trans. Circ. Syst. Video Technol. 25(5), 744–760 (2015)CrossRef
8.
Zurück zum Zitat Fathi, A., Ren, X., Rehg, J.M.: Learning to recognize objects in egocentric activities. In: 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2011) Fathi, A., Ren, X., Rehg, J.M.: Learning to recognize objects in egocentric activities. In: 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2011)
9.
Zurück zum Zitat Lu, Z., Grauman, K.: Story-driven summarization for egocentric video. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2013) Lu, Z., Grauman, K.: Story-driven summarization for egocentric video. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2013)
10.
Zurück zum Zitat Li, Y., Fathi, A., Rehg, J.: Learning to predict gaze in egocentric video. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3216–3223 (2013) Li, Y., Fathi, A., Rehg, J.: Learning to predict gaze in egocentric video. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3216–3223 (2013)
11.
Zurück zum Zitat Polatsek, P., Benesova, W., Paletta, L., Perko, R.: Novelty-based spatiotemporal saliency detection for prediction of gaze in egocentric video. IEEE Sig. Process. Lett. 23(3), 394–398 (2016)CrossRef Polatsek, P., Benesova, W., Paletta, L., Perko, R.: Novelty-based spatiotemporal saliency detection for prediction of gaze in egocentric video. IEEE Sig. Process. Lett. 23(3), 394–398 (2016)CrossRef
12.
Zurück zum Zitat Borji, A., Sihite, D.N., Itti, L.: What/where to look next? modeling top-down visual attention in complex interactive environments. IEEE Trans. Syst., Man Cybern.: Syst. 44(5), 523–538 (2014)CrossRef Borji, A., Sihite, D.N., Itti, L.: What/where to look next? modeling top-down visual attention in complex interactive environments. IEEE Trans. Syst., Man Cybern.: Syst. 44(5), 523–538 (2014)CrossRef
13.
Zurück zum Zitat Alahi, A., Bierlaire, M., Kunt, M.: Object detection and matching with mobile cameras collaborating with fixed cameras. In: Workshop on Multi-Camera and Multi-Modal Sensor Fusion Algorithms and Applications-M2SFA2 (2008) Alahi, A., Bierlaire, M., Kunt, M.: Object detection and matching with mobile cameras collaborating with fixed cameras. In: Workshop on Multi-Camera and Multi-Modal Sensor Fusion Algorithms and Applications-M2SFA2 (2008)
14.
Zurück zum Zitat Alahi, A., Marimon, D., Bierlaire, M., Kunt, M.: A master-slave approach for object detection and matching with fixed and mobile cameras. In: 15th IEEE International Conference on Image Processing, ICIP 2008 (2008) Alahi, A., Marimon, D., Bierlaire, M., Kunt, M.: A master-slave approach for object detection and matching with fixed and mobile cameras. In: 15th IEEE International Conference on Image Processing, ICIP 2008 (2008)
15.
Zurück zum Zitat Ferland, F., Pomerleau, F., Le Dinh, C., Michaud, F.: Egocentric and exocentric teleoperation interface using real-time, 3d video projection. In: 2009 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI) (2009) Ferland, F., Pomerleau, F., Le Dinh, C., Michaud, F.: Egocentric and exocentric teleoperation interface using real-time, 3d video projection. In: 2009 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI) (2009)
16.
Zurück zum Zitat Park, H., Jain, E., Sheikh, Y.: Predicting primary gaze behavior using social saliency fields. In: Proceedings of the IEEE International Conference on Computer Vision (2013) Park, H., Jain, E., Sheikh, Y.: Predicting primary gaze behavior using social saliency fields. In: Proceedings of the IEEE International Conference on Computer Vision (2013)
17.
Zurück zum Zitat Hoshen, Y., Ben-Artzi, G., Peleg, S.: Wisdom of the crowd in egocentric video curation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (2014) Hoshen, Y., Ben-Artzi, G., Peleg, S.: Wisdom of the crowd in egocentric video curation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (2014)
18.
Zurück zum Zitat Fathi, A., Hodgins, J.K., Rehg, J.M.: Social interactions: a first-person perspective. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2012) Fathi, A., Hodgins, J.K., Rehg, J.M.: Social interactions: a first-person perspective. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2012)
19.
Zurück zum Zitat Yan, Y., et al.: Egocentric daily activity recognition via multitask clustering. IEEE Trans. Image Process. 24(10), 2984–2995 (2015)MathSciNetCrossRef Yan, Y., et al.: Egocentric daily activity recognition via multitask clustering. IEEE Trans. Image Process. 24(10), 2984–2995 (2015)MathSciNetCrossRef
20.
Zurück zum Zitat Damen, D., Leelasawassuk, T., Haines, O., Calway, A., Mayol-Cuevas, W.: You-do, i-learn: discovering task relevant objects and their modes of interaction from multi-user egocentric video. In: BMVC (2014) Damen, D., Leelasawassuk, T., Haines, O., Calway, A., Mayol-Cuevas, W.: You-do, i-learn: discovering task relevant objects and their modes of interaction from multi-user egocentric video. In: BMVC (2014)
21.
Zurück zum Zitat Cheng, D.S., Cristani, M., Stoppa, M., Bazzani, L., Murino, V.: Custom pictorial structures for re-identification. In: BMVC (2011) Cheng, D.S., Cristani, M., Stoppa, M., Bazzani, L., Murino, V.: Custom pictorial structures for re-identification. In: BMVC (2011)
22.
Zurück zum Zitat Bak, S., Corvee, E., Brémond, F., Thonnat, M.: Multiple-shot human re-identification by mean riemannian covariance grid. In: 8th IEEE International Conference on Advanced Video and Signal-Based Surveillance (AVSS) (2011) Bak, S., Corvee, E., Brémond, F., Thonnat, M.: Multiple-shot human re-identification by mean riemannian covariance grid. In: 8th IEEE International Conference on Advanced Video and Signal-Based Surveillance (AVSS) (2011)
23.
Zurück zum Zitat Bazzani, L., Cristani, M., Murino, V.: Symmetry-driven accumulation of local features for human characterization and re-identification. In: Computer Vision and Image Understanding (2013) Bazzani, L., Cristani, M., Murino, V.: Symmetry-driven accumulation of local features for human characterization and re-identification. In: Computer Vision and Image Understanding (2013)
24.
Zurück zum Zitat Poleg, Y., Arora, C., Peleg, S.: Head motion signatures from egocentric videos. In: Cremers, D., Reid, I., Saito, H., Yang, M.-H. (eds.) ACCV 2014. LNCS, vol. 9005, pp. 315–329. Springer, Heidelberg (2015) Poleg, Y., Arora, C., Peleg, S.: Head motion signatures from egocentric videos. In: Cremers, D., Reid, I., Saito, H., Yang, M.-H. (eds.) ACCV 2014. LNCS, vol. 9005, pp. 315–329. Springer, Heidelberg (2015)
25.
Zurück zum Zitat Yonetani, R., Kitani, K.M., Sato, Y.: Ego-surfing first person videos. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2015) Yonetani, R., Kitani, K.M., Sato, Y.: Ego-surfing first person videos. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2015)
26.
Zurück zum Zitat Zamir, A.R., Ardeshir, S., Shah, M.: Robust refinement of GPS-tags using random walks with an adaptive damping factor. In: IEEE International Conference on Computer Vision and Pattern Recognition (CVPR) (2014) Zamir, A.R., Ardeshir, S., Shah, M.: Robust refinement of GPS-tags using random walks with an adaptive damping factor. In: IEEE International Conference on Computer Vision and Pattern Recognition (CVPR) (2014)
27.
Zurück zum Zitat Kiefer, P., Giannopoulos, I., Raubal, M.: Where am i? investigating map matching during selflocalization with mobile eye tracking in an urban environment. Trans. GIS 18(5), 660–686 (2014)CrossRef Kiefer, P., Giannopoulos, I., Raubal, M.: Where am i? investigating map matching during selflocalization with mobile eye tracking in an urban environment. Trans. GIS 18(5), 660–686 (2014)CrossRef
28.
Zurück zum Zitat Ardeshir, S., Zamir, A.R., Torroella, A., Shah, M.: GIS-assisted object detection and geospatial localization. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014, Part VI. LNCS, vol. 8694, pp. 602–617. Springer, Heidelberg (2014) Ardeshir, S., Zamir, A.R., Torroella, A., Shah, M.: GIS-assisted object detection and geospatial localization. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014, Part VI. LNCS, vol. 8694, pp. 602–617. Springer, Heidelberg (2014)
29.
Zurück zum Zitat Ardeshir, S., Collins-Sibley, K.M., Shah, M.: Geo-semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR (2015) Ardeshir, S., Collins-Sibley, K.M., Shah, M.: Geo-semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR (2015)
30.
31.
Zurück zum Zitat Felzenszwalb, P.F., Girshick, R.B., McAllester, D., Ramanan, D.: Object detection with discriminatively trained part based models. IEEE Trans. Pattern Anal. Mach. Intell. 32(9), 1627–1645 (2010)CrossRef Felzenszwalb, P.F., Girshick, R.B., McAllester, D., Ramanan, D.: Object detection with discriminatively trained part based models. IEEE Trans. Pattern Anal. Mach. Intell. 32(9), 1627–1645 (2010)CrossRef
Metadaten
Titel
Ego2Top: Matching Viewers in Egocentric and Top-View Videos
verfasst von
Shervin Ardeshir
Ali Borji
Copyright-Jahr
2016
DOI
https://doi.org/10.1007/978-3-319-46454-1_16