Deutsch
 
Hilfe Datenschutzhinweis Impressum
  DetailsucheBrowse

Datensatz

 
 
DownloadE-Mail
  Reinforcement Learning for Operational Space Control

Peters, J., & Schaal, S. (2007). Reinforcement Learning for Operational Space Control. In 2007 IEEE International Conference on Robotics and Automation (pp. 2111-2116). Los Alamitos, CA, USA: IEEE Computer Society.

Item is

Externe Referenzen

einblenden:
ausblenden:
Beschreibung:
-
OA-Status:

Urheber

einblenden:
ausblenden:
 Urheber:
Peters, J1, Autor           
Schaal, S, Autor           
Affiliations:
1External Organizations, ou_persistent22              

Inhalt

einblenden:
ausblenden:
Schlagwörter: -
 Zusammenfassung: While operational space control is of essential importance
for robotics and well-understood from an analytical
point of view, it can be prohibitively hard to achieve accurate
control in face of modeling errors, which are inevitable in
complex robots, e.g., humanoid robots. In such cases, learning
control methods can offer an interesting alternative to analytical
control algorithms. However, the resulting supervised learning
problem is ill-defined as it requires to learn an inverse mapping
of a usually redundant system, which is well known to suffer
from the property of non-convexity of the solution space, i.e.,
the learning system could generate motor commands that try
to steer the robot into physically impossible configurations. The
important insight that many operational space control algorithms
can be reformulated as optimal control problems, however, allows
addressing this inverse learning problem in the framework of
reinforcement learning. However, few of the known optimization
or reinforcement learning algorithms can be used in online
learning control for robots, as they are either prohibitively
slow, do not scale to interesting domains of complex robots,
or require trying out policies generated by random search,
which are infeasible for a physical system. Using a generalization
of the EM-based reinforcement learning framework suggested
by Dayan amp; Hinton, we reduce the problem of learning with
immediate rewards to a reward-weighted regression problem
with an adaptive, integrated reward transformation for faster
convergence. The resulting algorithm is efficient, learns smoothly
without dangerous jumps in solution space, and works well in
applications of complex high degree-of-freedom robots.

Details

einblenden:
ausblenden:
Sprache(n):
 Datum: 2007-04
 Publikationsstatus: Erschienen
 Seiten: -
 Ort, Verlag, Ausgabe: -
 Inhaltsverzeichnis: -
 Art der Begutachtung: -
 Identifikatoren: DOI: 10.1109/ROBOT.2007.363633
BibTex Citekey: 4723
 Art des Abschluß: -

Veranstaltung

einblenden:
ausblenden:
Titel: 2007 IEEE International Conference on Robotics and Automation (ICRA 2007)
Veranstaltungsort: Roma, Italy
Start-/Enddatum: 2007-04-10 - 2007-04-14

Entscheidung

einblenden:

Projektinformation

einblenden:

Quelle 1

einblenden:
ausblenden:
Titel: 2007 IEEE International Conference on Robotics and Automation
Genre der Quelle: Konferenzband
 Urheber:
Affiliations:
Ort, Verlag, Ausgabe: Los Alamitos, CA, USA : IEEE Computer Society
Seiten: - Band / Heft: - Artikelnummer: - Start- / Endseite: 2111 - 2116 Identifikator: ISBN: 1-4244-0602-1