Creators: | Faußer, Stefan A. and Schwenker, Friedhelm |
---|---|
Title: | Ensemble Methods for Reinforcement Learning with Function Approximation |
Item Type: | Conference or Workshop Item |
Event Title: | (Proceedings of the) 10th International Workshop on Multiple Classifier Systems (MCS) |
Event Location: | Naples, Italy |
Event Dates: | June, 15-17, 2011 |
Page Range: | pp. 56-65 |
Date: | 2011 |
Divisions: | Informationsmanagement |
Abstract (ENG): | Ensemble methods allow to combine multiple models to increase the predictive performances but mostly utilize labelled data. In this paper we propose several ensemble methods to learn a combined parameterized state-value function of multiple agents. For this purpose the Temporal-Difference (TD) and Residual-Gradient (RG) update methods as well as a policy function is adapted to learn from joint decisions. Such joint decisions include Majority Voting and Averaging of the state-values. We apply these ensemble methods to the simple pencil-and-paper game Tic-Tac-Toe and show that an ensemble of three agents outperforms a single agent in terms of the Mean-Squared Error (MSE) to the true values as well as in terms of the resulting policy. Further we apply the same methods to learn the shortest path in a 20 ×20 maze and empirically show that the learning speed is faster and the resulting policy, i.e. the number of correctly choosen actions is better in an ensemble of multiple agents than that of a single agent. |
Forthcoming: | No |
Language: | English |
Citation: | Faußer, Stefan A. and Schwenker, Friedhelm (2011) Ensemble Methods for Reinforcement Learning with Function Approximation. In: (Proceedings of the) 10th International Workshop on Multiple Classifier Systems (MCS), June, 15-17, 2011, Naples, Italy, pp. 56-65. ISBN 9783642215568 |
Actions for admins (login required)
View Item in edit mode |