WIMLE: Uncertainty-Aware World Models with IMLE for Sample-Efficient Continuous Control

Abstract

Model-based reinforcement learning can improve sample efficiency by leveraging synthetic rollouts from learned world models, but policy learning often suffers from bias due to compounding model errors. We propose WIMLE, an uncertainty-aware model-based RL approach that uses Implicit Maximum Likelihood Estimation (IMLE) to learn multi-modal stochastic world models. WIMLE captures aleatoric uncertainty through latent-variable sampling, estimates epistemic uncertainty with deep ensembles, and applies uncertainty-weighted learning so each synthetic transition contributes proportionally to model confidence. Evaluated on over 40 continuous-control tasks across DeepMind Control Suite, HumanoidBench, and MyoSuite, WIMLE delivers substantial gains in sample efficiency and asymptotic performance, with state-of-the-art results on challenging humanoid and dog tasks.

Publication
The Fourteenth International Conference on Learning Representations
@inproceedings{aghabozorgi2026wimle,
  title={{WIMLE}: Uncertainty-Aware World Models with {IMLE} for Sample-Efficient Continuous Control},
  author={Mehran Aghabozorgi and Yanshu Zhang and Alireza Moazeni and Ke Li},
  booktitle={The Fourteenth International Conference on Learning Representations},
  year={2026},
  url={https://openreview.net/forum?id=mzLOnTb3WH}
}
Mehran Aghabozorgi
Mehran Aghabozorgi
Ph.D. Student

My research interests include Generative modeling and Reinforcement Learning.

Yanshu Zhang
Yanshu Zhang
Ph.D. Student

My research interests include 3D vision and Reinforcement Learning.

Alireza Moazeni
Alireza Moazeni
Ph.D. Student

My research interests include Neural Rendering and Generative Models.

Ke Li
Ke Li
Professor of Computer Science

My research interests include Machine Learning, Computer Vision and Algorithms.