Sham Kakade, University of Washington
A fundamental question in the theory of reinforcement learning is what properties govern our ability to generalize and avoid the curse of dimensionality. With regards to supervised learning, these questions are well understood theoretically, and, practically speaking, we have overwhelming evidence on the value of representational learning (say through modern deep networks) as a means for sample efficient learning. Providing an analogous theory for reinforcement learning is far more challenging, where even characterizing the representational conditions which support sample efficient generalization is far less well understood.
This talk will highlight recent advances towards characterizing when generalization is possible in reinforcement learning, focusing on both lower bounds (addressing issues of what constitutes a good representation) along with upper bounds (where we consider a broad set of sufficient conditions).
You are invited to a scheduled Zoom meeting. Zoom is Yale’s audio and visual conferencing platform.
Topic: Yale S&DS Department Seminar
Time: 4:00pm - 5:00pm
Join from PC, Mac, Linux, iOS or Android: https://yale.zoom.us/j/99169700816?pwd=SWEvWHI5d3dPNVdHMkZMZURMWWJPUT09
Or Telephone：203-432-9666 (2-ZOOM if on-campus) or 646 568 7788
Meeting ID: 991 6970 0816
International numbers available: https://yale.zoom.us/u/acBOaD1ic6
For H.323 and SIP information for video conferencing units please click here: https://yale.service-now.com/it?id=support_article&sys_id=434b72d3db9e8fc83514b1c0ef961924