To Örebro University

oru.seÖrebro University Publications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Variable Impedance Skill Learning for Contact-Rich Manipulation
Örebro University, School of Science and Technology. (Autonomous Mobile Manipulation Lab, Center for Applied Autonomous Sensor Systems (AASS))ORCID iD: 0000-0001-5655-0990
Department of Computer Science, Faculty of Engineering (LTH), Lund University, Lund, Sweden.
Department of Computer Science, Faculty of Engineering (LTH), Lund University, Lund, Sweden.
Örebro University, School of Science and Technology. (Autonomous Mobile Manipulation Lab, Center for Applied Autonomous Sensor Systems (AASS))ORCID iD: 0000-0003-3958-6179
Show others and affiliations
2022 (English)In: IEEE Robotics and Automation Letters, E-ISSN 2377-3766, Vol. 7, no 3, p. 8391-8398Article in journal, Letter (Refereed) Published
Abstract [en]

Contact-rich manipulation tasks remain a hard problem in robotics that requires interaction with unstructured environments. Reinforcement Learning (RL) is one potential solution to such problems, as it has been successfully demonstrated on complex continuous control tasks. Nevertheless, current state-of-the-art methods require policy training in simulation to prevent undesired behavior and later domain transfer even for simple skills involving contact. In this paper, we address the problem of learning contact-rich manipulation policies by extending an existing skill-based RL framework with a variable impedance action space. Our method leverages a small set of suboptimal demonstration trajectories and learns from both position, but also crucially impedance-space information. We evaluate our method on a number of peg-in-hole task variants with a Franka Panda arm and demonstrate that learning variable impedance actions for RL in Cartesian space can be deployed directly on the real robot, without resorting to learning in simulation.

Place, publisher, year, edition, pages
IEEE Press, 2022. Vol. 7, no 3, p. 8391-8398
Keywords [en]
Machine learning for robot control, reinforcement learning, variable impedance control
National Category
Robotics
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:oru:diva-100386DOI: 10.1109/LRA.2022.3187276ISI: 000838455200009Scopus ID: 2-s2.0-85133737407OAI: oai:DiVA.org:oru-100386DiVA, id: diva2:1685070
Funder
Knut and Alice Wallenberg FoundationAvailable from: 2022-08-01 Created: 2022-08-01 Last updated: 2024-01-17
In thesis
1. Robot Skill Acquisition through Prior-Conditioned Reinforcement Learning
Open this publication in new window or tab >>Robot Skill Acquisition through Prior-Conditioned Reinforcement Learning
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Advancements in robotics and artificial intelligence have paved the way for autonomous agents to perform complex tasks in various domains. A critical challenge in the field of robotics is enabling robots to acquire and refine skills efficiently, allowing them to adapt and excel in diverse environments. This thesis investigates the questions of how to acquire robot skills through priorconstrained machine learning and adapt these learned skills to novel environments safely and efficiently.

The thesis leverages the synergy between Reinforcement Learning (RL) and prior knowledge to facilitate skill acquisition in robots. It integrates existing task constraints, domain knowledge and contextual information into the learning process, enabling the robot to acquire new skills efficiently. The core idea behind our method is to exploit structured priors derived from both expert demonstrations and domain-specific information which guide the RL process to effectively explore and exploit the state-action space.

The first contribution lies in guaranteeing the execution of safe actions and preventing constraint violations during the exploration phase of RL. By incorporating task-specific constraints, the robot avoids entering into regions of the environment where potential risks or failures may occur. It allows for efficient exploration of the action space while maintaining safety, making it well-suited for scenarios where continuous actions need to adhere to specific constraints. The second contribution addresses the challenge of learning a policy on a real robot to accomplish contact-rich tasks by exploiting a set of pre-collected demonstrations. Specifically, a variable impedance action space is leveraged to enable the system to effectively adapt its interactions during contact-rich manipulation tasks. In the third contribution, the thesis explores the transferability of skills acquired across different tasks and domains, highlighting the framework’s potential for building a repository of reusable skills. By comparing the similarity between the target task and the prior tasks, prior knowledge is combined to guide the policy learning process for new tasks. In the fourth contribution of this thesis, we introduce a cycle generative model to transfer acquired skills across different robot platforms by learning from unstructured prior demonstrations. In summary, the thesis introduces a novel paradigm for advancing the field of robotic skill acquisition by synergizing prior knowledge with RL.

Place, publisher, year, edition, pages
Örebro: Örebro University, 2023. p. 66
Series
Örebro Studies in Technology, ISSN 1650-8580 ; 101
Keywords
Reinforcement Learning, Robot Manipulation, Transfer Learning, Safety Constraints, Prior Knowledge Learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:oru:diva-108230 (URN)9789175295251 (ISBN)
Public defence
2023-10-31, Örebro universitet, Forumhuset, Hörsal F, Fakultetsgatan 1, Örebro, 09:15 (English)
Opponent
Supervisors
Available from: 2023-09-12 Created: 2023-09-12 Last updated: 2023-10-17Bibliographically approved

Open Access in DiVA

Variable Impedance Skill Learning for Contact-Rich Manipulation(2394 kB)270 downloads
File information
File name FULLTEXT01.pdfFile size 2394 kBChecksum SHA-512
7b5f1cc52e9dba92656b5fa306743b1858c984ea7e4788f7b1240bf1ec71fc197c15f44e3ebd8e9f8fb9c29e62d64aae7a3ad485c6a00b8d2693f4ef8e60de36
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Yang, QuantaoStork, Johannes A.Stoyanov, Todor

Search in DiVA

By author/editor
Yang, QuantaoStork, Johannes A.Stoyanov, Todor
By organisation
School of Science and Technology
In the same journal
IEEE Robotics and Automation Letters
Robotics

Search outside of DiVA

GoogleGoogle Scholar
Total: 277 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 309 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf