Model-Based Meta Automatic Curriculum Learning (2022)
Zifan Xu, Yulin Zhang, Shahaf S. Shperberg, Reuth Mirsky, Yulin Zhan, Yuqian Jiang, Bo Liu, and Peter Stone
When an agent trains for one target task, its experience is expected to be useful for training on another target task. This paper formulates the meta curriculum learning problem that builds a sequence of intermediate training tasks, called a curriculum, which will assist the learner to train toward any given target task in general. We propose a model-based meta automatic curriculum learning algorithm (MM-ACL) that learns to predict the performance improvement on one task when the policy is trained on another, given contextual information such as the history of training tasks, loss functions, rollout state-action trajectories from the policy, etc. This predictor facilitates the generation of curricula that optimizes the performance of the learner on different target tasks. Our empirical results demonstrate that MM-ACL outperforms a random curriculum, a manually created curriculum, and a commonly used non-stationary bandit algorithm in a GridWorld domain.
In Decision Awareness in Reinforcement Learning (DARL) workshop t the +39th International Conference on Machine Learning (ICML), Baltimore, Maryland, USA, July 2022.

Yuqian Jiang Ph.D. Student
Peter Stone Faculty pstone [at] cs utexas edu