

Author: Hosokawa Shu
Publisher: Springer Publishing Company
ISSN: 1433-5298
Source: Artificial Life and Robotics, Vol.15, Iss.4, 2010-12, pp. : 560-564
Disclaimer: Any content in publications that violate the sovereignty, the constitution or regulations of the PRC is not accepted or approved by CNPIEC.
Abstract
Many reinforcement learning methods have been studied on the assumption that a state is discretized and the environment size is predetermined. However, an operating environment may have a continuous state and its size may not be known in advance, e.g., in robot navigation and control. When applying these methods to the environment described above, we may need a large amount of time for learning or failing to learn. In this study, we improve our previous human immunity-based reinforcement learning method so that it will work in continuous state space environments. Since our method selects an action based on the distance between the present state and the memorized action, information about the environment (e.g., environment size) is not required in advance. The validity of our method is demonstrated through simulations for the swingup control of an inverted pendulum.
Related content






Kernel-Based Reinforcement Learning
By Ormoneit D.
Machine Learning, Vol. 49, Iss. 2-3, 2002-11 ,pp. :




Explanation-Based Learning and Reinforcement Learning: A Unified View
Machine Learning, Vol. 28, Iss. 2-3, 1997-08 ,pp. :