Big Science and Embodied Learning at Hugging Face š¤ with Thomas Wolf
EPISODE 564
|
MARCH
21,
2022
Watch
Follow
Share
About this Episode
Today weāre joined by Thomas Wolf, co-founder and chief science officer at Hugging Face š¤. We cover a ton of ground In our conversation, starting with Thomasā interesting backstory as a quantum physicist and patent lawyer, and how that lead him to a career in machine learning. We explore how Hugging Face began, what the current direction is for the company, and how much of their focus is NLP and language models versus other disciplines. We also discuss the BigScience project, a year-long research workshop where 1000+ researchers of all backgrounds and disciplines have come together to create a 800GB multilingual dataset and model. We talk through their approach to curating the dataset, model evaluation at this scale, and how they differentiate their work from projects like Eluther AI. Finally, we dig into Thomasā work on multimodality, his thoughts on the metaverse, his new book NLP with Transformers, and much more!
About the Guest
Thomas Wolf
HuggingFace š¤
Resources
- Book: Natural Language Processing with Transformers: Building Language Applications with Hugging Face
- Paper: Multitask Prompted Training Enables Zero-Shot Task Generalization
- Course: Hugging Face Course - Transformer models
- BigScience
- CodeParrot Generation
- DataMeasurementsTool - a Hugging Face Space by huggingface
- ElutherAI
- DeepMoji
- Can Language Models Be Too Big? w/ Emily Bender and Margaret Mitchell - #467
- What's Next for Fast.ai with Jeremy Howard - #421
