TempoRL - Learning When to Act Permalink
TL;DR: Jointly learning when and how to act improves sample efficiency of RL agents through better exploration and improved exploitation.
TL;DR: Jointly learning when and how to act improves sample efficiency of RL agents through better exploration and improved exploitation.
TL;DR: New toy benchmarks enable better study of RL agents performance and allows us to compare against ground truth optimal policies.
TL;DR: We propose SPaCE, a general approach that can be deployed for any agent with a value function and task instances identified by their context without a...
TL;DR: An overview of research on AutoRL in 2021 by the AutoML Freiburg group.