Mixture-of-Experts and Trends in Large-Scale Language Modeling with Irwan Bello
EPISODE 569
|
APRIL
25,
2022
Watch
Follow
Share
About this Episode
Today we’re joined by Irwan Bello, formerly a research scientist at Google Brain, and now on the founding team at a stealth AI startup. We begin our conversation with an exploration of Irwan’s recent paper, Designing Effective Sparse Expert Models, which acts as a design guide for building sparse large language model architectures. We discuss mixture of experts as a technique, the scalability of this method, and it's applicability beyond NLP tasks the data sets this experiment was benchmarked against. We also explore Irwan’s interest in the research areas of alignment and retrieval, talking through interesting lines of work for each area including instruction tuning and direct alignment.
About the Guest
Irwan Bello
Google Brain
Resources
- Paper: Designing Effective Sparse Expert Models
- Paper: Improving language models by retrieving from trillions of tokens
- Blog: WebGPT: Improving the Factual Accuracy of Language Models through Web Browsing
- 100x Improvements in Deep Learning Performance with Sparsity, w/ Subutai Ahmad - #562
- Hierarchical and Continual RL with Doina Precup - #567