English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Conference Paper

Using Reward-weighted Regression for Reinforcement Learning of Task Space Control

MPS-Authors
There are no MPG-Authors in the publication available
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)
There are no public fulltexts stored in PuRe
Supplementary Material (public)
There is no public supplementary material available
Citation

Peters, J., & Schaal, S. (2007). Using Reward-weighted Regression for Reinforcement Learning of Task Space Control. In 2007 IEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning (pp. 262-267). Piscataway, NJ, USA: IEEE Computer Society.


Cite as: https://hdl.handle.net/11858/00-001M-0000-0013-CE29-8
Abstract
Many robot control problems of practical importance, including task or operational space control, can be
reformulated as immediate reward reinforcement learning problems.
However, few of the known optimization or reinforcement
learning algorithms can be used in online learning control
for robots, as they are either prohibitively slow, do not scale
to interesting domains of complex robots, or require trying
out policies generated by random search, which are infeasible
for a physical system. Using a generalization of the EM-base
reinforcement learning framework suggested by Dayan amp; Hinton,
we reduce the problem of learning with immediate rewards to a
reward-weighted regression problem with an adaptive, integrated
reward transformation for faster convergence. The resulting
algorithm is efficient, learns smoothly without dangerous jumps
in solution space, and works well in applications of complex high
degree-of-freedom robots.