The second RL theory workshop (co-located with COLT 2024)
Our second* RL theory workshop took place 27-29 June 2024 at the University of Alberta. Project Eva, Jack Mayo and Daniel Abrahamian’s startup, sponsored the event.
David Janz, Alex Ayoub & Csaba Szepesvári
*Recordings from the previous workshop are available here.
Talks list
June 28th am:
- Truncated Variance Reduced Value Iteration, 1hr, Ishani Aniruddha Karmarkar, video
- Optimistic Q-learning for average reward and episodic RL, 1hr, Shipra Agrawal, video
- Uncertainty-Aware Reward-Free Exploration with General Function Approximation, 30min, Dongruo Zhou, video
- On the Instance-dependent Sample Complexity of Tabular RL, 30min Kevin Jamieson, video
June 28th pm:
- Scalable Online Exploration via Coverability, 30min Philip Amortila, paper link, video
- Statistical and Algorithmic Reductions for Reinforcement Learning From Rich Observations, 30min, Philip Amortila, video
- Bisimulation Metrics are Optimal Transport Distances, and Can be Computed Efficiently, 45min Gergely Neu, video
- Towards Principled, Practical Policy Gradient for Bandits and Tabular MDPs, 30min Sharan Vaswani, video
- Self-Play Preference Optimization for Language Model Alignment, 30min Quanquan Gu, video
June 29th am:
- A Computationally Efficient Algorithm for Infinite-Horizon Average Reward Reinforcement Learning with Linear MDPs, 1hr, Ki Hong, video
- Offline Reinforcement Learning: Role of State Aggregation and Trajectory Data, 1hr, Zeyu Jia, paper link, video
- Efficient exploration in deep RL via utility theory, 45min, Brendan O’Donoghue, video
June 29th pm:
- Robust Online Learning in the Presence of Outliers, 30min, Tim van Erven, paper link, (video recording failed)
- Reinforcement Learning in Mean Field Games: the pitfalls and promises, 30min, Niao He, video
- When are Offline Multi-Agent Games Solvable?, 30min, Simon Du, video