Scaling Multi-Modal Generative AI with Luke Zettlemoyer
EPISODE 650
|
OCTOBER
9,
2023
Watch
Follow
Share
About this Episode
Today we’re joined by Luke Zettlemoyer, professor at University of Washington and a research manager at Meta. In our conversation with Luke, we cover multimodal generative AI, the effect of data on models, and the significance of open source and open science. We explore the grounding problem, the need for visual grounding and embodiment in text-based models, the advantages of discretization tokenization in image generation, and his paper Scaling Laws for Generative Mixed-Modal Language Models, which focuses on simultaneously training LLMs on various modalities. Additionally, we cover his papers on Self-Alignment with Instruction Backtranslation, and LIMA: Less Is More for Alignment.
About the Guest
Luke Zettlemoyer
University of Washington, Meta
Resources
- Paper: Scaling Laws for Generative Mixed-Modal Language Models
- Paper: Scaling Autoregressive Multi-modal models (CM3Leon)
- Paper: Self-Alignment with Instruction Backtranslation
- Paper: LIMA: Less Is More for Alignment
- Paper: Language Contamination Helps Explain the Cross-lingual Capabilities of English Pretrained Models
- Paper: Retrieval-Augmented Multimodal Language Modeling
- Unifying Vision and Language Models with Mohit Bansal - #636
- Is ChatGPT Getting Worse? with James Zou - #645

