Continuous Q-learning application


Abstract eng:
Standard algorithm of Q-Learning is limited by discrete states and actions and Qfunction is u sually represented as discrete table. To avoid this obstacle and extend the use of Q-learning for continuous states and actions the algorithm must be modified and such modification is presented in the paper. Straightforward way is to replace discrete table with suitable approximator. A number of approximators can be used, with respect to memory and computational requirements the local approximator is particularly favorable. We have used Locally Weighted Regression (LWR) algorithm. The paper discusses advantages and disadvantages of modified algorithm demonstrated on simple control task.

Contributors:
Publisher:
Institute of Thermomechanics AS CR, v.v.i., Prague
Conference Title:
Conference Title:
Engineering Mechanics 2004
Conference Venue:
Svratka (CZ)
Conference Dates:
2004-05-10 / 2004-05-13
Rights:
Text je chráněný podle autorského zákona č. 121/2000 Sb.



Record appears in:

 Record created 2014-11-14, last modified 2014-11-18


Original version of the author's contribution as presented on CD, . :
Download fulltext
PDF

Rate this document:

Rate this document:
1
2
3
 
(Not yet reviewed)