University of Twente Student Theses

Login

Towards reducing the sample complexity of a model-free reinforcement learning agent controlling a single segment tendon-driven continuum manipulator

Hendriks, K.J.H. (2020) Towards reducing the sample complexity of a model-free reinforcement learning agent controlling a single segment tendon-driven continuum manipulator.

[img] PDF
15MB
Abstract:This work aims to outline an end-to-end process to develop a practically viable reinforcement learning controller based on the soft actor-critic algorithm by reducing its sample complexity. A tendon-driven continuum manipulator is fabricated and then modelled using a non-linear autoregressive exogenous neural network. This model is used to generate a student policy that imitates expert behaviour as well as the policy of a model-free agent trained in simulation. The simulated agent’s policy and the student policy are used to initialise a model-free learner, with the intent of reducing the sample complexity by allowing the agent to focus on fine-tuning an already competent policy. The effectiveness of these methods is evaluated by comparing the performance as a function of learning time with that of an agent that was trained without any prior knowledge. Results indicate that while the endoscope is able to learn a reaching task, the sparsity of information about the state-space in the student policy and the model inaccuracies used to develop the simulated agent lead to performances that were similar or worse for a given number of training steps.
Item Type:Essay (Master)
Faculty:EEMCS: Electrical Engineering, Mathematics and Computer Science
Subject:50 technical science in general
Programme:Systems and Control MSc (60359)
Link to this item:https://purl.utwente.nl/essays/84083
Export this item as:BibTeX
EndNote
HTML Citation
Reference Manager

 

Repository Staff Only: item control page