Logo des Repositoriums
 

Accounting for Task-Difficulty in Active Multi-Task Robot Control Learning

dc.contributor.authorFabisch, Alexander
dc.contributor.authorMetzen, Jan Hendrik
dc.contributor.authorKrell, Mario Michael
dc.contributor.authorKirchner, Frank
dc.date.accessioned2018-01-08T09:18:05Z
dc.date.available2018-01-08T09:18:05Z
dc.date.issued2015
dc.description.abstractContextual policy search is a reinforcement learning approach for multi-task learning in the context of robot control learning. It can be used to learn versatilely applicable skills that generalize over a range of tasks specified by a context vector. In this work, we combine contextual policy search with ideas from active learning for selecting the task in which the next trial will be performed. Moreover, we use active training set selection for reducing detrimental effects of exploration in the sampling policy. A core challenge in this approach is that the distribution of the obtained rewards may not be directly comparable between different tasks. We propose the novel approach PUBSVE for estimating a reward baseline and investigate empirically on benchmark problems and simulated robotic tasks to which extent this method can remedy the issue of non-comparable reward.
dc.identifier.pissn1610-1987
dc.identifier.urihttps://dl.gi.de/handle/20.500.12116/11484
dc.publisherSpringer
dc.relation.ispartofKI - Künstliche Intelligenz: Vol. 29, No. 4
dc.relation.ispartofseriesKI - Künstliche Intelligenz
dc.subjectActive learning
dc.subjectContextual policy search
dc.subjectMulti-task learning
dc.titleAccounting for Task-Difficulty in Active Multi-Task Robot Control Learning
dc.typeText/Journal Article
gi.citation.endPage377
gi.citation.startPage369

Dateien