English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Conference Paper

Automatic synthesis of sequences of human movements by linear combination of learned example patterns

MPS-Authors
/persons/resource/persons84018

Knappmeyer,  B
Max Planck Institute for Biological Cybernetics, Max Planck Society;
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

/persons/resource/persons83839

Bülthoff,  HH
Max Planck Institute for Biological Cybernetics, Max Planck Society;
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)
There are no public fulltexts stored in PuRe
Supplementary Material (public)
There is no public supplementary material available
Citation

Giese, M., Knappmeyer, B., & Bülthoff, H. (2002). Automatic synthesis of sequences of human movements by linear combination of learned example patterns. In H. Bülthoff, C. Wallraven, S.-W. Lee, & T. Poggio (Eds.), Biologically Motivated Computer Vision: Second International Workshop, BMCV 2002 Tübingen, Germany, November 22–24, 2002 (pp. 538-547). Berlin, Germany: Springer.


Cite as: https://hdl.handle.net/11858/00-001M-0000-0013-DE50-2
Abstract
We present a method for the synthesis of sequences of realistically looking human movements from learned example patterns. We apply this technique for the synthesis of dynamic facial expressions. Sequences of facial movements are decomposed into individual movement elements which are modeled by linear combinations of learned examples. The weights of the linear combinations define an abstract pattern space that permits a simple modification and parameterization of the style of the individual movement elements. The elements are defined in a way that is suitable for a simple automatic resynthesis of longer sequences from movement elements with different styles. We demonstrate the efficiency of this technique for the animation of a 3D head model and discuss how it can be used to generate spatio-temporally exaggerated sequences of facial expressions for psychophysical experiments on caricature effects.