ASSET Seminar: “Beyond Scaling: Frontiers of Retrieval-Augmented Language Models”
February 12, 2025 at 12:00 PM - 1:15 PM
Details
Abstract:
Large Language Models (LMs) have demonstrated remarkable capabilities by scaling up training data and model sizes. However, they continue to face critical challenges, including hallucinations and outdated knowledge, which particularly limit their reliability in expert domains such as scientific research and software development. In this talk, I will urge the necessity of moving beyond the traditional scaling of monolithic LMs and advocate for Augmented LMs—a new AI paradigm that designs, trains, and deploys LMs alongside complementary modules to address these limitations. Focusing on my research on Retrieval-Augmented LMs, one of the most impactful and widely adopted forms of Augmented LMs today, I will begin by presenting our systematic analyses of current LM shortcomings and demonstrate how Retrieval-Augmented LMs offer a more effective and efficient path forward. I will then discuss my work to establish new foundations for further reliability and efficiency by designing and training new LMs and retrieval systems to dynamically adapt to diverse inputs. Finally, I will demonstrate the real-world impact of such Retrieval-Augmented LMs through OpenScholar, our fully open Retrieval-Augmented LM designed to assist scientists in synthesizing scientific literature, now used by more than 25,000 researchers and practitioners worldwide. I will conclude by outlining my vision for the future of Augmented LMs, emphasizing advancements in their abilities to handle heterogeneous and diverse modalities, more efficient and effective integration with diverse components, and advancing evaluations with interdisciplinary collaboration.
Zoom Link (if unable to attend in-person): https://upenn.zoom.us/j/95663463468

