To Örebro University

oru.seÖrebro University Publications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Learning Impedance Actions for Safe Reinforcement Learning in Contact-Rich Tasks
Örebro University, School of Science and Technology. (Autonomous Mobile Manipulation Lab (AMM))ORCID iD: 0000-0001-5655-0990
Department of Computer Science, Lund University, Sweden.
Department of Computer Science, Lund University, Sweden.
Örebro University, School of Science and Technology. (Autonomous Mobile Manipulation Lab (AMM))ORCID iD: 0000-0003-3958-6179
Show others and affiliations
2021 (English)In: NeurIPS 2021 Workshop on Deployable Decision Making in Embodied Systems (DDM), 2021Conference paper, Published paper (Other academic)
Abstract [en]

Reinforcement Learning (RL) has the potential of solving complex continuous control tasks, with direct applications to robotics. Nevertheless, current state-of-the-art methods are generally unsafe to learn directly on a physical robot as exploration by trial-and-error can cause harm to the real world systems. In this paper, we leverage a framework for learning latent action spaces for RL agents from demonstrated trajectories. We extend this framework by connecting it to a variable impedance Cartesian space controller, allowing us to learn contact-rich tasks safely and efficiently. Our method learns from trajectories that incorporate both positional, but also crucially impedance-space information. We evaluate our method on a number of peg-in-hole task variants with a Franka Panda arm and demonstrate that learning variable impedance actions for RL in Cartesian space can be safely deployed on the real robot directly, without resorting to learning in simulation and a subsequent policy transfer.

Place, publisher, year, edition, pages
2021.
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:oru:diva-95945OAI: oai:DiVA.org:oru-95945DiVA, id: diva2:1620121
Conference
NeurIPS 2021 Workshop on Deployable Decision Making in Embodied Systems (DDM), (Online conference), Sydney, Australia, December 6-14, 2021
Available from: 2021-12-14 Created: 2021-12-14 Last updated: 2022-09-12Bibliographically approved

Open Access in DiVA

Learning Impedance Actions for Safe Reinforcement Learning in Contact-Rich Tasks(455 kB)528 downloads
File information
File name FULLTEXT01.pdfFile size 455 kBChecksum SHA-512
8728bfa0fd8c09ccf93f7942f01bb67ecc3cbb11057b70469e59a31709e86202250582106b673721f8dfb34fee1337b86752cdb0b3cb08af680caa1c6893eb83
Type fulltextMimetype application/pdf

Authority records

Yang, QuantaoStork, Johannes AndreasStoyanov, Todor

Search in DiVA

By author/editor
Yang, QuantaoStork, Johannes AndreasStoyanov, Todor
By organisation
School of Science and Technology
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 543 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

urn-nbn

Altmetric score

urn-nbn
Total: 704 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf