English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
  Comparing Search Strategies of Humans and Machines in Clutter

Michaelis, C., Weller, M., Funke, C., Ecker, A., Wallis, T., & Bethge, M. (2019). Comparing Search Strategies of Humans and Machines in Clutter. Poster presented at Nineteenth Annual Meeting of the Vision Sciences Society (VSS 2019), St. Pete Beach, FL, USA. doi:10.1167/19.10.309c.

Item is

Files

show Files

Locators

show
hide
Description:
-

Creators

show
hide
 Creators:
Michaelis, C, Author
Weller, M, Author
Funke, C, Author
Ecker, AS, Author              
Wallis , TSA, Author
Bethge, M1, 2, Author              
Affiliations:
1Max Planck Institute for Biological Cybernetics, Max Planck Society, ou_1497794              
2Research Group Computational Vision and Neuroscience, Max Planck Institute for Biological Cybernetics, Max Planck Society, ou_1497805              

Content

show
hide
Free keywords: -
 Abstract: While many perceptual tasks become more difficult in the presence of clutter, in general the human visual system has evolved tolerance to cluttered environments. In contrast, current machine learning approaches struggle in the presence of clutter. We compare human observers and CNNs on two target localization tasks with cluttered images created from characters or rendered objects. Each task sample consists of such a cluttered image as well as a separate image of one object which has to be localized. Human observers are asked to identify wether the object lies in the left or right half of the image and accuracy, reaction time and eye movements are recorded. CNNs are trained to segment the object and the position of the center of mass of the segmentation mask is then used to predict the position. Clutter levels are defined by the set-size ranging from 2 to 256 objects per image. We find that for humans processing times increase with the amount of clutter while for machine learning models accuracy drops. This points to a critical difference in human and machine processing: humans search serially whereas current machine learning models typically process a whole image in one pass. Following this line of thought we show that machine learning models with two iterations of processing perform significantly better than the purely feed-forward CNNs dominating in current object recognition applications. This finding suggests that confronted with challenging scenes iterative processing might be just as important for machines as it is for humans.

Details

show
hide
Language(s):
 Dates: 2019-052019-09
 Publication Status: Published in print
 Pages: -
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: DOI: 10.1167/19.10.309c
 Degree: -

Event

show
hide
Title: Nineteenth Annual Meeting of the Vision Sciences Society (VSS 2019)
Place of Event: St. Pete Beach, FL, USA
Start-/End Date: 2019-05-17 - 2019-05-22

Legal Case

show

Project information

show

Source 1

show
hide
Title: Journal of Vision
Source Genre: Journal
 Creator(s):
Affiliations:
Publ. Info: Charlottesville, VA : Scholar One, Inc.
Pages: - Volume / Issue: 19 (10) Sequence Number: 63.412 Start / End Page: 309 - 310 Identifier: ISSN: 1534-7362
CoNE: https://pure.mpg.de/cone/journals/resource/111061245811050