English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Report

Eye-Hand Coordination for 3-D Oriented Objects

MPS-Authors
/persons/resource/persons84073

Mamassian,  P
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;
Max Planck Institute for Biological Cybernetics, Max Planck Society;

/persons/resource/persons83839

Bülthoff,  HH
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;
Max Planck Institute for Biological Cybernetics, Max Planck Society;

/persons/resource/persons84011

Kersten,  D
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;
Max Planck Institute for Biological Cybernetics, Max Planck Society;

External Resource
No external resources are shared
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)

MPIK-TR-12.pdf
(Publisher version), 3MB

Supplementary Material (public)
There is no public supplementary material available
Citation

Mamassian, P., Bülthoff, H., & Kersten, D.(1995). Eye-Hand Coordination for 3-D Oriented Objects (12). Tübingen, Germany: Max Planck Institute for Biological Cybernetics.


Cite as: https://hdl.handle.net/11858/00-001M-0000-0013-ECBE-0
Abstract
When reaching for an object in space, the distance of the object, its orientation and shape should all be correctly estimated well before the hand arrives in contact with the object. We investigate here how the visual information about the object's orientation is incorporated into the motor program which guides the hand to the vicinity of the object. This motor program governs in particular the transportation of the hand, the opening of the hand, and the
rotations of the forearm and wrist. These variables are analyzed in the light of the ``visuo-motor channels'' framework proposed by Jeannerod (1981), in which the information about the object's distance and shape is processed independently, from perception to action.
Our results seem to favor an alternative framework, in which an interpretation of the three-dimensional world should be built before a motor command can be generated. Due to time constraints, this mechanism appears to be mainly feedforward.