Social Foundations of Computation Talk Biography
24 June 2024 at 10:30

The Unreasonable Effectiveness of Distributional Reinforcement Learning

ORGANIZERS
Thumb ticker sm 20241104 hardt moritz 12 cleaned kleiner
Social Foundations of Computation
  • Director
Thumb ticker xxl nathan

Distributional Reinforcement Learning (RL) learns the whole conditional distribution of rewards-to-go, given current state and action, but then only ever looks at the mean (e.g., C51, IQN). While this appears inefficient on its face, empirically it often improves on analogous approaches (e.g., DQN) that directly learn just the conditional mean (i.e., the Q-function). A principled understanding as to why and when this happens has been elusive.

Speaker Biography

Nathan Kallus (Cornell Tech, Cornell University)

Associate Professor

Nathan Kallus is an Associate Professor at the Cornell Tech campus of Cornell University in NYC and a Research Director at Netflix. Nathan's research interests include the statistics of optimization under uncertainty, causal inference especially when combined with machine learning, sequential and dynamic decision making, and algorithmic fairness.