English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
 
 
DownloadE-Mail
  Where and When: Space-Time Attention for Audio-Visual Explanations

Chen, Y., Hummel, T., Koepke, A. S., & Akata, Z. (2021). Where and When: Space-Time Attention for Audio-Visual Explanations. Retrieved from https://arxiv.org/abs/2105.01517.

Item is

Basic

show hide
Genre: Paper
Latex : Where and When: {S}pace-Time Attention for Audio-Visual Explanations

Files

show Files
hide Files
:
arXiv:2105.01517.pdf (Preprint), 7MB
Name:
arXiv:2105.01517.pdf
Description:
File downloaded from arXiv at 2021-11-29 08:46
OA-Status:
Visibility:
Public
MIME-Type / Checksum:
application/pdf / [MD5]
Technical Metadata:
Copyright Date:
-
Copyright Info:
-

Locators

show

Creators

show
hide
 Creators:
Chen, Yanbei1, Author
Hummel, Thomas1, Author
Koepke, A. Sophia1, Author
Akata, Zeynep2, Author           
Affiliations:
1External Organizations, ou_persistent22              
2Computer Vision and Machine Learning, MPI for Informatics, Max Planck Society, ou_1116547              

Content

show
hide
Free keywords: Computer Science, Computer Vision and Pattern Recognition, cs.CV,Computer Science, Artificial Intelligence, cs.AI,Computer Science, Learning, cs.LG
 Abstract: Explaining the decision of a multi-modal decision-maker requires to determine
the evidence from both modalities. Recent advances in XAI provide explanations
for models trained on still images. However, when it comes to modeling multiple
sensory modalities in a dynamic world, it remains underexplored how to
demystify the mysterious dynamics of a complex multi-modal model. In this work,
we take a crucial step forward and explore learnable explanations for
audio-visual recognition. Specifically, we propose a novel space-time attention
network that uncovers the synergistic dynamics of audio and visual data over
both space and time. Our model is capable of predicting the audio-visual video
events, while justifying its decision by localizing where the relevant visual
cues appear, and when the predicted sounds occur in videos. We benchmark our
model on three audio-visual video event datasets, comparing extensively to
multiple recent multi-modal representation learners and intrinsic explanation
models. Experimental results demonstrate the clear superior performance of our
model over the existing methods on audio-visual video event recognition.
Moreover, we conduct an in-depth study to analyze the explainability of our
model based on robustness analysis via perturbation tests and pointing games
using human annotations.

Details

show
hide
Language(s): eng - English
 Dates: 2021-05-042021
 Publication Status: Published online
 Pages: 13 p.
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: arXiv: 2105.01517
BibTex Citekey: Chen2105.01517
URI: https://arxiv.org/abs/2105.01517
 Degree: -

Event

show

Legal Case

show

Project information

show hide
Project name : DEXIM
Grant ID : 853489
Funding program : Horizon 2020 (H2020)
Funding organization : European Commission (EC)

Source

show