Multilingual LLMs and the Values Divide in AI with Sara Hooker
EPISODE 651
|
OCTOBER
16,
2023
Watch
Follow
Share
About this Episode
Today we’re joined by Sara Hooker, director at Cohere and head of Cohere For AI, Cohere’s research lab. In our conversation with Sara, we explore some of the challenges with multilingual models like poor data quality and tokenization, and how they rely on data augmentation and preference training to address these bottlenecks. We also discuss the disadvantages and the motivating factors behind the Mixture of Experts technique, and the importance of common language between ML researchers and hardware architects to address the pain points in frameworks and create a better cohesion between the distinct communities. Sara also highlights the impact and the emotional connection that language models have created in society, the benefits and the current safety concerns of universal models, and the significance of having grounded conversations to characterize and mitigate the risk and development of AI models. Along the way, we also dive deep into Cohere and Cohere for AI, along with their Aya project, an open science project that aims to build a state-of-the-art multilingual generative language model as well as some of their recent research papers.
About the Guest
Sara Hooker
Cohere For AI, Cohere
Resources
- Cohere for AI website
- The Aya Project
- Aya open science initiative
- Learning sparse networks using targeted dropout
- Attention is all you need
- BLOOM
- When Less is More: Investigating Data Pruning for Pretraining LLMs at Scale
- Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning
- The Grand Illusion: The Myth of Software Portability and Implications for ML Progress
- The Hardware Lottery
- Model Cards for Model Reporting
- Datasheets for Datasets
- Frontier AI Regulation: Managing Emerging Risks to Public Safety
- Evaluating Model Explainability Methods with Sara Hooker - #189
- Ensuring LLM Safety for Production Applications with Shreya Rajpal - #647
- Big Science and Embodied Learning at Hugging Face with Thomas Wolf - #564
