English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
  Learning Manipulation under Physics Constraints with Visual Perception

Li, W., Leonardis, A., Bohg, J., & Fritz, M. (2019). Learning Manipulation under Physics Constraints with Visual Perception. Retrieved from http://arxiv.org/abs/1904.09860.

Item is

Files

show Files
hide Files
:
arXiv:1904.09860.pdf (Preprint), 7MB
Name:
arXiv:1904.09860.pdf
Description:
File downloaded from arXiv at 2019-07-03 11:19
OA-Status:
Visibility:
Public
MIME-Type / Checksum:
application/pdf / [MD5]
Technical Metadata:
Copyright Date:
-
Copyright Info:
-

Locators

show

Creators

show
hide
 Creators:
Li, Wenbin1, Author           
Leonardis, Aleš2, Author
Bohg, Jeannette2, Author
Fritz, Mario1, Author           
Affiliations:
1Computer Vision and Machine Learning, MPI for Informatics, Max Planck Society, ou_1116547              
2External Organizations, ou_persistent22              

Content

show
hide
Free keywords: Computer Science, Robotics, cs.RO
 Abstract: Understanding physical phenomena is a key competence that enables humans and
animals to act and interact under uncertain perception in previously unseen
environments containing novel objects and their configurations. In this work,
we consider the problem of autonomous block stacking and explore solutions to
learning manipulation under physics constraints with visual perception inherent
to the task. Inspired by the intuitive physics in humans, we first present an
end-to-end learning-based approach to predict stability directly from
appearance, contrasting a more traditional model-based approach with explicit
3D representations and physical simulation. We study the model's behavior
together with an accompanied human subject test. It is then integrated into a
real-world robotic system to guide the placement of a single wood block into
the scene without collapsing existing tower structure. To further automate the
process of consecutive blocks stacking, we present an alternative approach
where the model learns the physics constraint through the interaction with the
environment, bypassing the dedicated physics learning as in the former part of
this work. In particular, we are interested in the type of tasks that require
the agent to reach a given goal state that may be different for every new
trial. Thereby we propose a deep reinforcement learning framework that learns
policies for stacking tasks which are parametrized by a target structure.

Details

show
hide
Language(s): eng - English
 Dates: 2019-04-192019
 Publication Status: Published online
 Pages: 23 p.
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: arXiv: 1904.09860
URI: http://arxiv.org/abs/1904.09860
BibTex Citekey: Li_arXiv1904.09860
 Degree: -

Event

show

Legal Case

show

Project information

show

Source

show