English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
 
 
DownloadE-Mail
  Understanding Low- and High-Level Contributions to Fixation Prediction

Kümmerer, M., Wallis, T., Gatys, L., & Bethge, M. (2017). Understanding Low- and High-Level Contributions to Fixation Prediction. In 19th IEEE International Conference on Computer Vision (ICCV 2017) (pp. 4799-4808). Piscataway, NJ, USA: IEEE.

Item is

Basic

show hide
Genre: Conference Paper

Files

show Files

Locators

show
hide
Locator:
Link (Any fulltext)
Description:
-
OA-Status:

Creators

show
hide
 Creators:
Kümmerer, M1, Author           
Wallis, TSA1, Author
Gatys, LA1, Author
Bethge, M1, Author           
Affiliations:
1University of Tübingen, Centre for Integrative Neuroscience, ou_persistent22              

Content

show
hide
Free keywords: -
 Abstract: Understanding where people look in images is an important problem in computer vision. Despite significant research, it remains unclear to what extent human fixations can be predicted by low-level (contrast) compared to highlevel (presence of objects) image features. Here we address this problem by introducing two novel models that use different feature spaces but the same readout architecture. The first model predicts human fixations based on deep neural network features trained on object recognition. This model sets a new state-of-the art in fixation prediction by achieving top performance in area under the curve metrics on the MIT300 hold-out benchmark (AUC = 88, sAUC = 77, NSS = 2.34). The second model uses purely low-level (isotropic contrast) features. This model achieves better performance than all models not using features pretrained on object recognition, making it a strong baseline to assess the utility of high-level features. We then evaluate and visualize which fixations are better explained by lowlevel compared to high-level image features. Surprisingly we find that a substantial proportion of fixations are better explained by the simple low-level model than the stateof- the-art model. Comparing different features within the same powerful readout architecture allows us to better understand the relevance of low- versus high-level features in predicting fixation locations, while simultaneously achieving state-of-the-art saliency prediction.

Details

show
hide
Language(s):
 Dates: 2017-10
 Publication Status: Issued
 Pages: -
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: DOI: 10.1109/ICCV.2017.513
BibTex Citekey: KummererWGB2017
 Degree: -

Event

show
hide
Title: 19th IEEE International Conference on Computer Vision (ICCV 2017)
Place of Event: Venezia, Italy
Start-/End Date: -

Legal Case

show

Project information

show

Source 1

show
hide
Title: 19th IEEE International Conference on Computer Vision (ICCV 2017)
Source Genre: Proceedings
 Creator(s):
Affiliations:
Publ. Info: Piscataway, NJ, USA : IEEE
Pages: - Volume / Issue: - Sequence Number: - Start / End Page: 4799 - 4808 Identifier: ISBN: 978-1-5386-1032-9