2008 | OriginalPaper | Buchkapitel
Motion Context: A New Representation for Human Action Recognition
verfasst von : Ziming Zhang, Yiqun Hu, Syin Chan, Liang-Tien Chia
Erschienen in: Computer Vision – ECCV 2008
Verlag: Springer Berlin Heidelberg
Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.
Wählen Sie Textabschnitte aus um mit Künstlicher Intelligenz passenden Patente zu finden. powered by
Markieren Sie Textabschnitte, um KI-gestützt weitere passende Inhalte zu finden. powered by
One of the key challenges in human action recognition from video sequences is how to model an action sufficiently. Therefore, in this paper we propose a novel motion-based representation called
Motion Context
(MC), which is insensitive to the scale and direction of an action, by employing image representation techniques. A MC captures the distribution of the
motion words
(MWs) over relative locations in a local region of the
motion image
(MI) around a reference point and thus summarizes the local motion information in a rich 3D MC descriptor. In this way, any human action can be represented as a 3D descriptor by summing up all the MC descriptors of this action. For action recognition, we propose 4 different recognition configurations: MW+pLSA, MW+SVM, MC+
w
3
-pLSA (a new direct graphical model by extending pLSA), and MC+SVM. We test our approach on two human action video datasets from KTH and Weizmann Institute of Science (WIS) and our performances are quite promising. For the KTH dataset, the proposed MC representation achieves the highest performance using the proposed
w
3
-pLSA. For the WIS dataset, the best performance of the proposed MC is comparable to the state of the art.