Vis enkel innførsel

dc.contributor.authorKnudsen, Martinius
dc.contributor.authorHendseth, Sverre
dc.contributor.authorTufte, Gunnar
dc.contributor.authorSandvig, Axel
dc.date.accessioned2022-03-30T11:58:41Z
dc.date.available2022-03-30T11:58:41Z
dc.date.created2022-01-14T17:08:31Z
dc.date.issued2021
dc.identifier.citationModeling, Identification and Control. 2021, 42 (4), 197-204.en_US
dc.identifier.issn0332-7353
dc.identifier.urihttps://hdl.handle.net/11250/2988592
dc.description.abstractWe present here a model-free method for learning actions that lead to an all-source-all-destination shortest path solution. We motivate our approach in the context of biological learning for reactive control. Our method involves an agent exploring an unknown world with the objective of learning how to get from any starting state to any goal state in shortest time without having to run a path planning algorithm for each new goal selection. Using concepts of Lyapunov functions and Bellman's principle of optimality, our agent learns universal state-goal distances and best actions that solve this problem.en_US
dc.language.isoengen_US
dc.publisherNorwegian Society of Automatic Controlen_US
dc.rightsNavngivelse 4.0 Internasjonal*
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/deed.no*
dc.titleModel-Free All-Source-All-Destination Learning as a Model for Biological Reactive Controlen_US
dc.typeJournal articleen_US
dc.typePeer revieweden_US
dc.description.versionpublishedVersionen_US
dc.source.pagenumber197-204en_US
dc.source.volume42en_US
dc.source.journalModeling, Identification and Controlen_US
dc.source.issue4en_US
dc.identifier.doi10.4173/mic.2021.4.5
dc.identifier.cristin1981523
cristin.ispublishedtrue
cristin.fulltextoriginal
cristin.qualitycode1


Tilhørende fil(er)

Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel

Navngivelse 4.0 Internasjonal
Med mindre annet er angitt, så er denne innførselen lisensiert som Navngivelse 4.0 Internasjonal