Vis enkel innførsel

dc.contributor.advisorLekkas, Anastasios
dc.contributor.authorLund, Emil Andreas
dc.date.accessioned2018-09-17T14:01:26Z
dc.date.available2018-09-17T14:01:26Z
dc.date.created2018-06-04
dc.date.issued2018
dc.identifierntnudaim:18660
dc.identifier.urihttp://hdl.handle.net/11250/2563040
dc.description.abstractUsing reinforcement learning as a part of a Guidance, Navigation and Control (GNC) system is a relatively unexplored field. This thesis explores the use of deep reinforcement learning in a path following control algorithm in a realistic simulated environment. A reinforcement learning method called the Asynchronous Advantage Actor-Critic (A3C) method has been implemented using artificial neural networks to model the learning, making it \textit{deep} reinforcement learning. In addition, the software needed to use the V-REP robot simulator together with the A3C algorithm was developed. This was used to successfully learn to control a vehicle model in a path following situation in the V-REP simulator. In addition, the algorithm was used to learn to control a much simpler ship model simulated without V-REP. It was found that the learned behavior from the ship model could be transferred and used to control the vehicle model in the more complex environment, substantially speeding up the total time of learning process for the vehicle.
dc.languageeng
dc.publisherNTNU
dc.subjectKybernetikk og robotikk, Autonome systemer
dc.titlePath Following in Simulated Environments using the A3C Reinforcement Learning Method
dc.typeMaster thesis


Tilhørende fil(er)

Thumbnail
Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel