To Örebro University

oru.seÖrebro University Publications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Null space based efficient reinforcement learning with hierarchical safety constraints
Örebro University, School of Science and Technology. (Autonomous Mobile Manipulation Lab (AMM))ORCID iD: 0000-0001-5655-0990
Örebro University, School of Science and Technology. (Autonomous Mobile Manipulation Lab (AMM))ORCID iD: 0000-0003-3958-6179
Örebro University, School of Science and Technology. (Autonomous Mobile Manipulation Lab (AMM))ORCID iD: 0000-0002-6013-4874
2021 (English)In: 2021 European Conference on Mobile Robots (ECMR), IEEE, 2021Conference paper, Published paper (Refereed)
Abstract [en]

Reinforcement learning is inherently unsafe for use in physical systems, as learning by trial-and-error can cause harm to the environment or the robot itself. One way to avoid unpredictable exploration is to add constraints in the action space to restrict the robot behavior. In this paper, we proposea null space based framework of integrating reinforcement learning methods in constrained continuous action spaces. We leverage a hierarchical control framework to decompose target robotic skills into higher ranked tasks (e. g., joint limits and obstacle avoidance) and lower ranked reinforcement learning task. Safe exploration is guaranteed by only learning policies in the null space of higher prioritized constraints. Meanwhile multiple constraint phases for different operational spaces are constructed to guide the robot exploration. Also, we add penalty loss for violating higher ranked constraints to accelerate the learning procedure. We have evaluated our method on different redundant robotic tasks in simulation and show that our null space based reinforcement learning method can explore and learn safely and efficiently.

Place, publisher, year, edition, pages
IEEE, 2021.
National Category
Robotics
Identifiers
URN: urn:nbn:se:oru:diva-95146DOI: 10.1109/ECMR50962.2021.9568848ISI: 000810510000061ISBN: 9781665412131 (print)OAI: oai:DiVA.org:oru-95146DiVA, id: diva2:1605135
Conference
European Conference on Mobile Robots (ECMR 2021), Virtual meeting, August 31 - September 3, 2021
Note

Funding agency:

Wallenberg Artificial Intelligence, Autonomous Systems and Software Program (WASP)

Available from: 2021-10-21 Created: 2021-10-21 Last updated: 2023-10-06Bibliographically approved
In thesis
1. Robot Skill Acquisition through Prior-Conditioned Reinforcement Learning
Open this publication in new window or tab >>Robot Skill Acquisition through Prior-Conditioned Reinforcement Learning
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Advancements in robotics and artificial intelligence have paved the way for autonomous agents to perform complex tasks in various domains. A critical challenge in the field of robotics is enabling robots to acquire and refine skills efficiently, allowing them to adapt and excel in diverse environments. This thesis investigates the questions of how to acquire robot skills through priorconstrained machine learning and adapt these learned skills to novel environments safely and efficiently.

The thesis leverages the synergy between Reinforcement Learning (RL) and prior knowledge to facilitate skill acquisition in robots. It integrates existing task constraints, domain knowledge and contextual information into the learning process, enabling the robot to acquire new skills efficiently. The core idea behind our method is to exploit structured priors derived from both expert demonstrations and domain-specific information which guide the RL process to effectively explore and exploit the state-action space.

The first contribution lies in guaranteeing the execution of safe actions and preventing constraint violations during the exploration phase of RL. By incorporating task-specific constraints, the robot avoids entering into regions of the environment where potential risks or failures may occur. It allows for efficient exploration of the action space while maintaining safety, making it well-suited for scenarios where continuous actions need to adhere to specific constraints. The second contribution addresses the challenge of learning a policy on a real robot to accomplish contact-rich tasks by exploiting a set of pre-collected demonstrations. Specifically, a variable impedance action space is leveraged to enable the system to effectively adapt its interactions during contact-rich manipulation tasks. In the third contribution, the thesis explores the transferability of skills acquired across different tasks and domains, highlighting the framework’s potential for building a repository of reusable skills. By comparing the similarity between the target task and the prior tasks, prior knowledge is combined to guide the policy learning process for new tasks. In the fourth contribution of this thesis, we introduce a cycle generative model to transfer acquired skills across different robot platforms by learning from unstructured prior demonstrations. In summary, the thesis introduces a novel paradigm for advancing the field of robotic skill acquisition by synergizing prior knowledge with RL.

Place, publisher, year, edition, pages
Örebro: Örebro University, 2023. p. 66
Series
Örebro Studies in Technology, ISSN 1650-8580 ; 101
Keywords
Reinforcement Learning, Robot Manipulation, Transfer Learning, Safety Constraints, Prior Knowledge Learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:oru:diva-108230 (URN)9789175295251 (ISBN)
Public defence
2023-10-31, Örebro universitet, Forumhuset, Hörsal F, Fakultetsgatan 1, Örebro, 09:15 (English)
Opponent
Supervisors
Available from: 2023-09-12 Created: 2023-09-12 Last updated: 2023-10-17Bibliographically approved

Open Access in DiVA

Null space based efficient reinforcement learning with hierarchical safety constraints(872 kB)414 downloads
File information
File name FULLTEXT01.pdfFile size 872 kBChecksum SHA-512
52f9a80e4988f79110247b1358d07f44b57c21dba53433e81d118faeba3a5739b756094fd7bcc2c8187603faefa5a98df931c32029bb180ace2b3fe86a0dac8e
Type fulltextMimetype application/pdf

Other links

Publisher's full text

Authority records

Yang, QuantaoStork, Johannes AndreasStoyanov, Todor

Search in DiVA

By author/editor
Yang, QuantaoStork, Johannes AndreasStoyanov, Todor
By organisation
School of Science and Technology
Robotics

Search outside of DiVA

GoogleGoogle Scholar
Total: 414 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 283 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf