Abstract
This paper addresses the decision-making mechanism, for controlling Markov Decision Processes (MDPs) with factored action-spaces. The paper proposes a simple decision-making approach, which decomposes a multidimensional action-space into its dimensions and distributes the dimensions amongst multiple autonomous agents. A distinct Q-function is defined for each agent, and a Reinforcement Learning (RL) algorithm, which uses the decision-making approach, is proposed. The algorithm maintains estimations that converge to the agents' Q-functions. It is shown that by distributing the decision-making, and due to the factored action-space, new opportunities for representation abstractions emerge. The paper's contribution is illustrated, using a navigation task that involves three robots.
Original language | English |
---|---|
Pages (from-to) | 188-193 |
Number of pages | 6 |
Journal | Proceedings of the IASTED International Conference on Intelligent Systems and Control |
State | Published - 2006 |
Event | 9th IASTED International Conference on Intelligent Systems and Control, ISC 2006 - Honolulu, HI, United States Duration: 14 Aug 2006 → 16 Aug 2006 |
Keywords
- Distributed reinforcement-learning
- Factored MDPs
- Representation abstraction