Abstract
The rapid evolution of industrial robot hardware has created a technological gap with software, limiting its adoption. The software solutions proposed in recent years have yet to meet the industrial sector’s requirements, as they focus more on the definition of task structure than the definition and tuning of its execution parameters. A framework for task parameter optimization was developed to address this gap. It breaks down the task using a modular structure, allowing the task optimization piece by piece. The optimization is performed with a dedicated hill-climbing algorithm. This paper revisits the framework by proposing an alternative approach that replaces the algorithmic component with reinforcement learning (RL) models. Five RL models are proposed with increasing complexity and efficiency. A comparative analysis of the traditional algorithm and RL models is presented, highlighting efficiency, flexibility, and usability. The results demonstrate that although RL models improve task optimization efficiency by 95%, they still need more flexibility. However, the nature of these models provides significant opportunities for future advancements.
| Original language | English |
|---|---|
| Pages (from-to) | 173734-173748 |
| Number of pages | 15 |
| Journal | IEEE Access |
| Volume | 12 |
| DOIs | |
| Publication status | Published - 20 Nov 2024 |
Bibliographical note
Copyright © 2024 The Authors. This work is licensed under a Creative Commons Attribution 4.0 License:https://creativecommons.org/licenses/by/4.0/
Keywords
- intuitive robot programming
- Reinforcement learning
- robotic task optimization
- task-oriented programming