English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Conference Paper

Performance Capture from Sparse Multi-view Video

MPS-Authors
/persons/resource/persons43977

de Aguiar,  Edilson
Computer Graphics, MPI for Informatics, Max Planck Society;

/persons/resource/persons45610

Theobalt,  Christian       
Computer Graphics, MPI for Informatics, Max Planck Society;

/persons/resource/persons45557

Stoll,  Carsten
Computer Graphics, MPI for Informatics, Max Planck Society;

/persons/resource/persons43978

Ahmed,  Naveed
Computer Graphics, MPI for Informatics, Max Planck Society;

/persons/resource/persons45449

Seidel,  Hans-Peter       
Computer Graphics, MPI for Informatics, Max Planck Society;

External Resource
No external resources are shared
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)
There are no public fulltexts stored in PuRe
Supplementary Material (public)
There is no public supplementary material available
Citation

de Aguiar, E., Theobalt, C., Stoll, C., Ahmed, N., Seidel, H.-P., & Thrun, S. (2008). Performance Capture from Sparse Multi-view Video. In G. Turk (Ed.), SIGGRAPH '08: ACM SIGGRAPH 2008 papers (pp. 98:1-98:10). New York, NY: ACM.


Cite as: https://hdl.handle.net/11858/00-001M-0000-000F-1CA4-0
Abstract
This paper proposes a new marker-less approach to capturing human performances
from multi-view video. Our algorithm can jointly reconstruct spatio-temporally
coherent geometry, motion and textural surface appearance of actors that
perform complex and rapid moves. Furthermore, since our algorithm is purely
meshbased and makes as few as possible prior assumptions about the type of
subject being tracked, it can even capture performances of people wearing wide
apparel, such as a dancer wearing a skirt. To serve this purpose our method
efficiently and effectively combines the power of surface- and volume-based
shape deformation techniques with a new mesh-based analysis-through-synthesis
framework. This framework extracts motion constraints from video and makes the
laser-scan of the tracked subject mimic the recorded performance. Also
small-scale time-varying shape detail is recovered by applying model-guided
multi-view stereo to refine the model surface. Our method delivers captured
performance data at higher level of detail, is highly versatile, and is
applicable to many complex types of scenes that could not be handled by
alternative marker-based or marker-free recording techniques.