Multiscale Markov Decision Problems: Compression, Solution, and Transfer Learning
Abstract
Many problems in sequential decision making and stochastic control often have natural multiscale structure: sub-tasks are assembled together to accomplish complex goals. Systematically inferring and leveraging hierarchical structure, particularly beyond a single level of abstraction, has remained a longstanding challenge. We describe a fast multiscale procedure for repeatedly compressing, or homogenizing, Markov decision processes (MDPs), wherein a hierarchy of sub-problems at different scales is automatically determined. Coarsened MDPs are themselves independent, deterministic MDPs, and may be solved using existing algorithms. The multiscale representation delivered by this procedure decouples sub-tasks from each other and can lead to substantial improvements in convergence rates both locally within sub-problems and globally across sub-problems, yielding significant computational savings. A second fundamental aspect of this work is that these multiscale decompositions yield new transfer opportunities across different problems, where solutions of sub-tasks at different levels of the hierarchy may be amenable to transfer to new problems. Localized transfer of policies and potential operators at arbitrary scales is emphasized. Finally, we demonstrate compression and transfer in a collection of illustrative domains, including examples involving discrete and continuous statespaces.
- Publication:
-
arXiv e-prints
- Pub Date:
- December 2012
- DOI:
- 10.48550/arXiv.1212.1143
- arXiv:
- arXiv:1212.1143
- Bibcode:
- 2012arXiv1212.1143B
- Keywords:
-
- Computer Science - Artificial Intelligence;
- Computer Science - Systems and Control;
- Mathematics - Optimization and Control;
- Statistics - Machine Learning
- E-Print:
- 86 pages, 15 figures