English
 
User Manual Privacy Policy Disclaimer Contact us
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
  Combining appearance and motion for human action classification in videos

Dhillon, P., Nowozin, S., & Lampert, C. (2009). Combining appearance and motion for human action classification in videos. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (pp. 22-29). Piscataway, NJ, USA: IEEE Service Center.

Item is

Basic

show hide
Item Permalink: http://hdl.handle.net/11858/00-001M-0000-0013-C473-3 Version Permalink: http://hdl.handle.net/21.11116/0000-0002-EAEF-A
Genre: Conference Paper

Files

show Files

Locators

show
hide
Description:
-

Creators

show
hide
 Creators:
Dhillon, PS, Author              
Nowozin, S1, 2, Author              
Lampert, C1, 2, Author              
Affiliations:
1Department Empirical Inference, Max Planck Institute for Biological Cybernetics, Max Planck Society, ou_1497795              
2Max Planck Institute for Biological Cybernetics, Max Planck Society, Spemannstrasse 38, 72076 Tübingen, DE, ou_1497794              

Content

show
hide
Free keywords: -
 Abstract: An important cue to high level scene understanding is to analyze the objects in the scene and their behavior and interactions. In this paper, we study the problem of classification of activities in videos, as this is an integral component of any scene understanding system, and present a novel approach for recognizing human action categories in videos by combining information from appearance and motion of human body parts. Our approach is based on tracking human body parts by using mixture particle filters and then clustering the particles using local non - parametric clustering, hence associating a local set of particles to each cluster mode. The trajectory of these cluster modes provides the “motion” information and the “appearance” information is provided by the statistical information about the relative motion of these local set of particles over a number of frames. Later we use a “Bag of Words” model to build one histogram per video sequence from the set of these robust appearance and motion descriptors. These histograms provide us characteristic information which helps us to discriminate among various human actions which ultimately helps us in better understanding of the complete scene. We tested our approach on the standard KTH and Weizmann human action datasets and the results were comparable to the state of the art methods. Additionally our approach is able to distinguish between activities that involve the motion of complete body from those in which only certain body parts move. In other words, our method discriminates well between activities with “global body motion” like running, jogging etc. and “local motion” like waving, boxing etc.

Details

show
hide
Language(s):
 Dates: 2009-06
 Publication Status: Published in print
 Pages: -
 Publishing info: -
 Table of Contents: -
 Rev. Method: -
 Identifiers: DOI: 10.1109/CVPR.2009.5204237
BibTex Citekey: 5900
 Degree: -

Event

show
hide
Title: 1st International Workshop on Visual Scene Understanding
Place of Event: Miami, FL, USA
Start-/End Date: 2009-06-20 - 2009-06-25

Legal Case

show

Project information

show

Source 1

show
hide
Title: 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
Source Genre: Proceedings
 Creator(s):
Affiliations:
Publ. Info: Piscataway, NJ, USA : IEEE Service Center
Pages: - Volume / Issue: - Sequence Number: - Start / End Page: 22 - 29 Identifier: ISBN: 978-1-4244-3993-5