BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//Penn Engineering Events - ECPv6.15.18//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-WR-CALNAME:Penn Engineering Events
X-ORIGINAL-URL:https://seasevents.nmsdev7.com
X-WR-CALDESC:Events for Penn Engineering Events
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/New_York
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20210314T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20211107T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20220313T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20221106T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20230312T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20231105T060000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20220406T150000
DTEND;TZID=America/New_York:20220406T160000
DTSTAMP:20260406T013305
CREATED:20220401T170128Z
LAST-MODIFIED:20220401T170128Z
UID:6657-1649257200-1649260800@seasevents.nmsdev7.com
SUMMARY:Spring 2022 GRASP SFI: Jason Ma\, University of Pennsylvania\, “Beyond Expected Reward in Offline Reinforcement Learning”
DESCRIPTION:*This will be a HYBRID Event with in-person attendance in Levine 512 and Virtual attendance via Zoom \nOffline reinforcement learning (RL)\, which uses pre-collected\, reusable offline data without further environment interactions\, permits sample-efficient\, scalable and practical decision-making; however\, most of the existing literature (1) focuses on improving algorithms for maximizing the expected cumulative reward\, and (2) assumes the reward function to be given. This limits the applicability of offline RL in many realistic settings — for instance\, there are often safety or risk constraints that need to be satisfied\, and the reward function is often difficult to specify. In this talk\, we will explore how we can (1) train a broad class of risk-sensitive agents using purely risk-neutral offline data and provably prevent out-of-distribution extrapolations\, and (2) bootstrap offline RL through flexible forms of expert demonstrations\, significantly expanding the scope of valid supervision for offline policy learning. With these advances\, we aim to bring offline RL closer to real-world applications.
URL:https://seasevents.nmsdev7.com/event/spring-2022-grasp-sfi-jason-ma-university-of-pennsylvania-beyond-expected-reward-in-offline-reinforcement-learning/
LOCATION:Levine 512
ORGANIZER;CN="General Robotics%2C Automation%2C Sensing and Perception (GRASP) Lab":MAILTO:grasplab@seas.upenn.edu
END:VEVENT
END:VCALENDAR