BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//Penn Engineering Events - ECPv6.15.18//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-WR-CALNAME:Penn Engineering Events
X-ORIGINAL-URL:https://seasevents.nmsdev7.com
X-WR-CALDESC:Events for Penn Engineering Events
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/New_York
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20210314T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20211107T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20220313T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20221106T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20230312T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20231105T060000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20220331T153000
DTEND;TZID=America/New_York:20220331T163000
DTSTAMP:20260406T012500
CREATED:20220324T132300Z
LAST-MODIFIED:20220324T132300Z
UID:6597-1648740600-1648744200@seasevents.nmsdev7.com
SUMMARY:CIS Seminar: “Towards a Foundation for Reinforcement Learning”
DESCRIPTION:In recent years\, reinforcement learning algorithms have achieved strong empirical success on a wide variety of real-world problems. However\, these algorithms usually require a huge number of samples even just for solving simple tasks. It is unclear if there are fundamental statistical limits on such methods\, or such sample complexity burden can be alleviated by a better algorithm. In this talk\, I will give an overview of my research efforts towards bridging the gap between the theory and the practice of reinforcement learning. \nIn the first part of the talk\, I will show that under conditions that permit sample-efficient supervised learning\, any offline reinforcement learning algorithm still requires exponential number of samples information-theoretically\, due to a geometric amplification of the estimation error. Moreover\, through extensive experiments on a range of tasks\, I will show that substantial error amplification does occur in practical scenarios. Our results highlight a crucial difference between offline reinforcement learning and supervised learning. I will conclude this part by suggesting possible ways to improve the performance of practical reinforcement learning systems based on our new insights. \nIn the second part of the talk\, I will focus on the horizon-dependence of the sample complexity of tabular reinforcement learning. I will show the first tabular reinforcement learning algorithm whose sample complexity is completely independent of the horizon length. Our result resolves a fundamental open problem in reinforcement learning theory.
URL:https://seasevents.nmsdev7.com/event/6597/
LOCATION:Zoom – Email CIS for link\, cherylh@cis.upenn.edu
ORGANIZER;CN="Computer and Information Science":MAILTO:cherylh@cis.upenn.edu
END:VEVENT
END:VCALENDAR