MODULE-BASED REINFORCEMENT LEARNING - EXPERIMENTS WITH A REAL ROBOT

Citation
Z. Kalmar et al., MODULE-BASED REINFORCEMENT LEARNING - EXPERIMENTS WITH A REAL ROBOT, Machine learning, 31(1-3), 1998, pp. 55-85
Citations number
72
Categorie Soggetti
Computer Science Artificial Intelligence","Computer Science Artificial Intelligence
Journal title
ISSN journal
08856125
Volume
31
Issue
1-3
Year of publication
1998
Pages
55 - 85
Database
ISI
SICI code
0885-6125(1998)31:1-3<55:MRL-EW>2.0.ZU;2-I
Abstract
The behavior of reinforcement learning (RL) algorithms is best underst ood in completely observable, discrete-rime controlled Markov chains w ith finite state and action spaces. In contrast, robot-learning domain s are inherently continuous both in time and space, and moreover are p artially observable. Here we suggest a systematic approach to solve su ch problems in which the available qualitative and quantitative knowle dge is used to reduce the complexity of learning task. The steps of th e design process are to: i) decompose the task into subtasks using the qualitative knowledge at hand; ii) design local controllers to solve the subtasks using the available quantitative knowledge and iii) learn a coordination of these controllers by means of reinforcement learnin g. It is argued that the approach enables fast, semi-automatic, but st ill high quality robot-control as no fine-tuning of the local controll ers is needed. The approach was verified on a non-trivial real-life ro bot task. Several RL algorithms were compared by ANOVA and it was foun d that the model-based approach worked significantly better than the m odel-free approach. The learnt switching strategy performed comparably to a handcrafted version. Moreover, the learnt strategy seemed to exp loit certain properties of the environment which were not foreseen in advance, thus supporting the view that adaptive algorithms are advanta geous to non-adaptive ones in complex environments.