The second RL theory workshop (co-located with COLT 2024)
We would like to invite you to the second* RL theory workshop, taking place 27-29 June 2024 at the University of Alberta (Co-located with COLT; COLT commences on the 30th of June). Schedule:
- Evening of June 27th: dinner and drinks (starting, say, 19:30)
- June 28th: Full day of talks and discussions, some fun activity in the evening
- June 29th: First half dedicated to talks, second half to recreation and
winding-down before COLT
Let us know if you intend to come by filling out this Google form (please fill this out even if you have confirmed your attendence otherwise, say, through having emailed with us about this). The workshop is free to attend, but we cannot provide funding or accomodation.
To get a VISA invitation letter, please use this form on the COLT website, noting in the “additional input” box that whether you will attend COLT, or just the workshop. We will generate the COLT invitation letter for you and if you need any modification to it, please let us know after receiving the letter.
We look forward to seeing you there,
David Janz (david.janz93 AT gmail.com), Alex Ayoub & Csaba Szepesvári
*Recordings from last year’s workshop are available here.
Organisational details
The workshop will take place at or nearby the Computing Science Centre of the University of Alberta; this is the same location as COLT. Regarding accommodation, two reasonable choices are:
- Guest rooms at the university (Lister dorms), website
- Campus tower hotel, website
The first is on campus, the second a three minute walk away.
Tentative talks list
Core RL theory
- Optimistic Q-learning for weakly communicating MDPs, Shipra Agrawal
- Truncated Variance Reduced Value Iteration, Ishani Aniruddha Karmarkar
- A Computationally Efficient Algorithm for Infinite-Horizon Average Reward Reinforcement Learning with Linear MDPs, Ki Hong
- Offline Reinforcement Learning: Role of State Aggregation and Trajectory Data, Zeyu Jia, paper link
- Statistical and Algorithmic Reductions for Reinforcement Learning From Rich Observations, Philip Amortila
- Scalable Online Exploration via Coverability, Dylan Foster, paper link
- On the Instance-dependent Sample Complexity of Tabular RL, Kevin Jamieson
- Uncertainty-Aware Reward-Free Exploration with General Function Approximation, Zhou, Dongruo
- Towards Principled, Practical Policy Gradient for Bandits and Tabular MDPs, Sharan Vaswani
Bandits
- The Dissimilarity Dimension: Sharper Bounds for Optimistic Algorithms, Aldo Pacchiano, paper link
- Achieving Optimal Static Regret and Dynamic Regret Simultaneously in Bandits, Jian Qian
Game theory
- Reinforcement Learning in Mean Field Games: the pitfalls and promises, Niao He
- When are Offline Multi-Agent Games Solvable?, Simon Du
- Towards Characterizing the First-order Query Complexity of Learning
(Approximate) Nash Equilibria in Zero-sum Matrix Games, Tim van Erven, paper link
Other fun topics
- Reinforcement Learning from Human Feedback with Active Queries, Quanquan Gu
- Efficient exploration in deep RL via utility theory, Brendan O’Donoghue
- Bisimulation Metrics are Optimal Transport Distances,
and Can be Computed Efficiently, Gergely Neu