English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Paper

VideoForensicsHQ: Detecting High-quality Manipulated Face Videos

MPS-Authors
/persons/resource/persons255864

Fox,  Gereon
Computer Graphics, MPI for Informatics, Max Planck Society;

/persons/resource/persons239547

Liu,  Wentao
Computer Graphics, MPI for Informatics, Max Planck Society;

/persons/resource/persons127713

Kim,  Hyeongwoo
Computer Graphics, MPI for Informatics, Max Planck Society;

/persons/resource/persons45449

Seidel,  Hans-Peter       
Computer Graphics, MPI for Informatics, Max Planck Society;

/persons/resource/persons229949

Elgharib,  Mohamed
Computer Graphics, MPI for Informatics, Max Planck Society;

/persons/resource/persons45610

Theobalt,  Christian       
Computer Graphics, MPI for Informatics, Max Planck Society;

External Resource
No external resources are shared
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)

arXiv:2005.10360.pdf
(Preprint), 11MB

Supplementary Material (public)
There is no public supplementary material available
Citation

Fox, G., Liu, W., Kim, H., Seidel, H.-P., Elgharib, M., & Theobalt, C. (2020). VideoForensicsHQ: Detecting High-quality Manipulated Face Videos. Retrieved from https://arxiv.org/abs/2005.10360.


Cite as: https://hdl.handle.net/21.11116/0000-0007-B109-7
Abstract
New approaches to synthesize and manipulate face videos at very high quality
have paved the way for new applications in computer animation, virtual and
augmented reality, or face video analysis. However, there are concerns that
they may be used in a malicious way, e.g. to manipulate videos of public
figures, politicians or reporters, to spread false information. The research
community therefore developed techniques for automated detection of modified
imagery, and assembled benchmark datasets showing manipulatons by
state-of-the-art techniques. In this paper, we contribute to this initiative in
two ways: First, we present a new audio-visual benchmark dataset. It shows some
of the highest quality visual manipulations available today. Human observers
find them significantly harder to identify as forged than videos from other
benchmarks. Furthermore we propose new family of deep-learning-based fake
detectors, demonstrating that existing detectors are not well-suited for
detecting fakes of a quality as high as presented in our dataset. Our detectors
examine spatial and temporal features. This allows them to outperform existing
approaches both in terms of high detection accuracy and generalization to
unseen fake generation methods and unseen identities.