Rachelson, Emmanuel and Fabiani, Patrick and Garcia, Frédérick
Approximate Policy Iteration for Generalized Semi-Markov Decision Processes: an Improved Algorithm.
(2008)
In: 8th European Workshop on Reinforcement Learning (EWRL), 30 June 2008 - 3 July 2008 (Villeneuve d'Ascq, France).
|
(Document in English)
PDF (Author's version) - Requires a PDF viewer such as GSview, Xpdf or Adobe Acrobat Reader 365kB |
Abstract
In the context of time-dependent problems of planning under uncertainty, most of the problem's complexity comes from the concurrent interaction of simultaneous processes. Generalized Semi-Markov Decision Processes represent an efficient formalism to capture both concurrency of events and actions and uncertainty. We introduce GSMDP with observable time and hybrid state space and present an new algorithm based on Approximate Policy Iteration to generate efficient policies. This algorithm relies on simulation-based exploration and makes use of SVM regression. We experimentally illustrate the strengths and weaknesses of this algorithm and propose an improved version based on the weaknesses highlighted by the experiments.
Item Type: | Conference or Workshop Item (Paper) |
---|---|
Audience (conference): | International conference proceedings |
Uncontrolled Keywords: | |
Institution: | French research institutions > Institut National de la Recherche Agronomique - INRA (FRANCE) French research institutions > Office National d'Etudes et Recherches Aérospatiales - ONERA (FRANCE) |
Laboratory name: | |
Statistics: | download |
Deposited On: | 29 Nov 2017 16:00 |
Repository Staff Only: item control page