Abstract
The control of the interaction between the robot and environment, following a predefined geometric surface path with high accuracy, is a fundamental problem for contact-rich tasks such as machining, polishing, or grinding. Flexible path-following control presents numerous applications in emerging industry fields such as disassembly and recycling, where the control system must adapt to a range of dissimilar object classes, where the properties of the environment are uncertain. We present an end-to-end framework for trajectory-independent robotic path following for contact-rich tasks in the presence of parametric uncertainties. We formulate a combination of model predictive control with image-based path planning and real-time visual feedback, based on a learned state-space dynamic model. For modeling the dynamics of the robot-environment system during contact, we introduce the application of the differentiable neural computer, a type of memory augmented neural network (MANN). Although MANNs have been as yet unexplored in a control context, we demonstrate a reduction in RMS error of ∼ 21.0% compared with an equivalent Long Short-Term Memory (LSTM) architecture. Our framework was validated in simulation, demonstrating the ability to generalize to materials previously unseen in the training dataset.
| Original language | English |
|---|---|
| Article number | 688275 |
| Number of pages | 16 |
| Journal | Frontiers in Robotics and AI |
| Volume | 8 |
| DOIs | |
| Publication status | Published - 26 Jul 2021 |
Bibliographical note
Copyright © 2021 Rastegarpanah, Hathaway and Stolkin. This is an open-accessarticle distributed under the terms of the Creative Commons Attribution License (CC
BY). The use, distribution or reproduction in other forums is permitted, provided the
original author(s) and the copyright owner(s) are credited and that the original
publication in this journal is cited, in accordance with accepted academic practice. No
use, distribution or reproduction is permitted which does not comply with these terms.
Data Access Statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/Supplementary Material. The raw dataset used for the model training and validation and the raw data used to produce the graphs in Figure 7 can be found here: CrossRef Full Text.Keywords
- cutting
- dynamic modeling
- electric vehicles
- machine learning
- predictive control
- vision