BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//Penn Engineering Events - ECPv6.15.18//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-WR-CALNAME:Penn Engineering Events
X-ORIGINAL-URL:https://seasevents.nmsdev7.com
X-WR-CALDESC:Events for Penn Engineering Events
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/New_York
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20220313T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20221106T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20230312T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20231105T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20240310T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20241103T060000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20230313T110000
DTEND;TZID=America/New_York:20230313T120000
DTSTAMP:20260404T154059
CREATED:20230227T190515Z
LAST-MODIFIED:20230227T190515Z
UID:8615-1678705200-1678708800@seasevents.nmsdev7.com
SUMMARY:ESE Spring Seminar - "Architecting High Performance Silicon Systems for Accurate and Efficient On-Chip Deep Learning"
DESCRIPTION:The unabated pursuit for omniscient and omnipotent AI is levying hefty latency\, memory\, and energy taxes at all computing scales. At the same time\, the end of Dennard scaling is sunsetting traditional performance gains commonly attained with reduction in transistor feature size. Faced with these challenges\, my research is building a heterogeneity of solutions co-optimized across the algorithm\, memory subsystem\, hardware architecture\, and silicon stack to generate breakthrough advances in arithmetic performance\, compute density and flexibility\, and energy efficiency for on-chip machine learning\, and natural language processing (NLP) in particular. I will start\, in the algorithm front\, by discussing award-winning work on developing a novel floating-point based data type\, AdaptivFloat\, which enables resilient quantized AI computations; and is particularly suitable for NLP networks with very large parameter distribution. Then\, I will describe a 16nm chip prototype that adopts AdaptivFloat in the acceleration of noise-robust AI speech and machine translation tasks – and whose fidelity to the front-end application is verified via a formal hardware/software compiler interface. Towards the goal of lowering the prohibitive energy cost of inferencing large language models on TinyML devices\, I will describe a principled algorithm-hardware co-design solution\, validated in a 12nm chip tapeout\, that accelerates Transformer workloads by tailoring the accelerator’s latency and energy expenditures according to the complexity of the input query it processes. Finally\, I will conclude with some of my current and future research efforts on further pushing the on-chip energy-efficiency frontiers by leveraging specialized non-conventional dynamic memory structures for on-device training — and recently prototyped in a 16nm tapeout.
URL:https://seasevents.nmsdev7.com/event/ese-spring-seminar-architecting-high-performance-silicon-systems-for-accurate-and-efficient-on-chip-deep-learning/
LOCATION:Raisler Lounge (Room 225)\, Towne Building\, 220 South 33rd Street\, Philadelphia\, PA\, 19104\, United States
CATEGORIES:Colloquium
ORGANIZER;CN="Electrical and Systems Engineering":MAILTO:eseevents@seas.upenn.edu
END:VEVENT
END:VCALENDAR