TY - JOUR
T1 - Constructing action set from basis functions for reinforcement learning of robot control
AU - Yamaguchi, Akihiko
AU - Takamatsu, Jun
AU - Ogasawara, Tsukasa
N1 - Publisher Copyright:
© 2009 IEEE.
PY - 2009
Y1 - 2009
N2 - Continuous action sets are used in many reinforcement learning (RL) applications in robot control since the control input is continuous. However, discrete action sets also have the advantages of ease of implementation and compatibility with some sophisticated RL methods, such as the Dyna [1]. However, one of the problem is the absence of general principles on designing a discrete action set for robot control in higher dimensional input space. In this paper, we propose to construct a discrete action set given a set of basis functions (BFs). We designed the action set so that the size of the set is proportional to the number of the BFs. This method can exploit the function approximator's nature, that is, in practical RL applications, the number of BFs does not increase exponentially with the dimension of the state space (e.g. [2]). Thus, the size of the proposed action set does not increase exponentially with the dimension of the input space. We apply an RL with the proposed action set to a robot navigation task and a crawling and a jumping tasks. The simulation results demonstrate that the proposed action set has the advantages of improved learning speed, and better ability to acquire performance, compared to a conventional discrete action set.
AB - Continuous action sets are used in many reinforcement learning (RL) applications in robot control since the control input is continuous. However, discrete action sets also have the advantages of ease of implementation and compatibility with some sophisticated RL methods, such as the Dyna [1]. However, one of the problem is the absence of general principles on designing a discrete action set for robot control in higher dimensional input space. In this paper, we propose to construct a discrete action set given a set of basis functions (BFs). We designed the action set so that the size of the set is proportional to the number of the BFs. This method can exploit the function approximator's nature, that is, in practical RL applications, the number of BFs does not increase exponentially with the dimension of the state space (e.g. [2]). Thus, the size of the proposed action set does not increase exponentially with the dimension of the input space. We apply an RL with the proposed action set to a robot navigation task and a crawling and a jumping tasks. The simulation results demonstrate that the proposed action set has the advantages of improved learning speed, and better ability to acquire performance, compared to a conventional discrete action set.
KW - Crawling
KW - Discrete action set
KW - Jumping
KW - Motion learning
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=72849146580&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=72849146580&partnerID=8YFLogxK
U2 - 10.1109/ROBOT.2009.5152840
DO - 10.1109/ROBOT.2009.5152840
M3 - Conference article
AN - SCOPUS:72849146580
SN - 1050-4729
SP - 2525
EP - 2532
JO - Proceedings - IEEE International Conference on Robotics and Automation
JF - Proceedings - IEEE International Conference on Robotics and Automation
M1 - 5152840
T2 - 2009 IEEE International Conference on Robotics and Automation, ICRA '09
Y2 - 12 May 2009 through 17 May 2009
ER -