Timezone: »
Poster
Policy gradients in linearly-solvable MDPs
Emanuel Todorov
We present policy gradient results within the framework of linearly-solvable MDPs. For the first time, compatible function approximators and natural policy gradients are obtained by estimating the cost-to-go function, rather than the (much larger) state-action advantage function as is necessary in traditional MDPs. We also develop the first compatible function approximators and natural policy gradients for continuous-time stochastic systems.
Author Information
Emanuel Todorov (University of Washington)
More from the Same Authors
-
2018 : Talk 2: Emo Todorov - Physics-Based Control »
Emanuel Todorov -
2017 Poster: Towards Generalization and Simplicity in Continuous Control »
Aravind Rajeswaran · Kendall Lowrey · Emanuel Todorov · Sham Kakade -
2016 : Emo Todorov (University of Washington) »
Emanuel Todorov -
2015 Poster: Interactive Control of Diverse Complex Characters with Neural Networks »
Igor Mordatch · Kendall Lowrey · Galen Andrew · Zoran Popovic · Emanuel Todorov -
2015 Oral: Interactive Control of Diverse Complex Characters with Neural Networks »
Igor Mordatch · Kendall Lowrey · Galen Andrew · Zoran Popovic · Emanuel Todorov -
2009 Poster: Compositionality of optimal control laws »
Emanuel Todorov