Value Iteration with Options and State Aggregation

01/16/2015
by   Kamil Ciosek, et al.
0

This paper presents a way of solving Markov Decision Processes that combines state abstraction and temporal abstraction. Specifically, we combine state aggregation with the options framework and demonstrate that they work well together and indeed it is only after one combines the two that the full benefit of each is realized. We introduce a hierarchical value iteration algorithm where we first coarsely solve subgoals and then use these approximate solutions to exactly solve the MDP. This algorithm solved several problems faster than vanilla value iteration.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset