TY - JOUR
T1 - Learning strategies in table tennis using inverse reinforcement learning
AU - Muelling, Katharina
AU - Boularias, Abdeslam
AU - Mohler, Betty
AU - Schölkopf, Bernhard
AU - Peters, Jan
N1 - Publisher Copyright:
© 2014, Springer-Verlag Berlin Heidelberg.
PY - 2014/10
Y1 - 2014/10
N2 - Learning a complex task such as table tennis is a challenging problem for both robots and humans. Even after acquiring the necessary motor skills, a strategy is needed to choose where and how to return the ball to the opponent’s court in order to win the game. The data-driven identification of basic strategies in interactive tasks, such as table tennis, is a largely unexplored problem. In this paper, we suggest a computational model for representing and inferring strategies, based on a Markov decision problem, where the reward function models the goal of the task as well as the strategic information. We show how this reward function can be discovered from demonstrations of table tennis matches using model-free inverse reinforcement learning. The resulting framework allows to identify basic elements on which the selection of striking movements is based. We tested our approach on data collected from players with different playing styles and under different playing conditions. The estimated reward function was able to capture expert-specific strategic information that sufficed to distinguish the expert among players with different skill levels as well as different playing styles.
AB - Learning a complex task such as table tennis is a challenging problem for both robots and humans. Even after acquiring the necessary motor skills, a strategy is needed to choose where and how to return the ball to the opponent’s court in order to win the game. The data-driven identification of basic strategies in interactive tasks, such as table tennis, is a largely unexplored problem. In this paper, we suggest a computational model for representing and inferring strategies, based on a Markov decision problem, where the reward function models the goal of the task as well as the strategic information. We show how this reward function can be discovered from demonstrations of table tennis matches using model-free inverse reinforcement learning. The resulting framework allows to identify basic elements on which the selection of striking movements is based. We tested our approach on data collected from players with different playing styles and under different playing conditions. The estimated reward function was able to capture expert-specific strategic information that sufficed to distinguish the expert among players with different skill levels as well as different playing styles.
KW - Computational models of decision processes
KW - Inverse reinforcement learning
KW - Table tennis
UR - http://www.scopus.com/inward/record.url?scp=84919877285&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84919877285&partnerID=8YFLogxK
U2 - 10.1007/s00422-014-0599-1
DO - 10.1007/s00422-014-0599-1
M3 - Article
C2 - 24756167
AN - SCOPUS:84919877285
SN - 0340-1200
VL - 108
SP - 603
EP - 619
JO - Biological Cybernetics
JF - Biological Cybernetics
IS - 5
ER -