Dynamic choice of state abstraction in q-learning

Q-learning associates states and actions of a Markov Decision Process to expected future reward through online learning. In practice, however, when the state space is large and experience is still limited, the algorithm will not find a match between current state and experience unless some details describing states are ignored. On the other hand, reducing state information affects long term performance because decisions will need to be made on less informative inputs. We propose a variation of Q-learning that gradually enriches state descriptions, after enough experience is accumulated. This is coupled with an ad-hoc exploration strategy that aims at collecting key information that allows the algorithm to enrich state descriptions earlier. Experimental results obtained by applying our algorithm to the arcade game Pac-Man show that our approach significantly outperforms Q-learning during the learning process while not penalizing long-term performance.

Author: 
Marco Tamassia
Fabio Zambetta
William L. Raffe
Florian 'Floyd' Mueller
Xiaodong Li
Presented At: 
ECAI 2016: Volume 285 Frontiers in Artificial Intelligence and Applications (FAIA), pp. 46-54
Year: 
2016
Type: 
Conference Proceedings