Snorkel: A System for Fast Training Data Creation with Alex Jason Ratner
EPISODE 270
|
MAY
30,
2019
Watch
Follow
Share
About this Episode
Today we're joined by Alex Ratner, Ph.D. student at Stanford, to discuss his work on Snorkel, a framework for creating training data with weak supervised learning techniques.
With Snorkel, Alex and his team hope to tackle the ever-present issue of having large data sets available by having users instead write a set of labeling functions, or scripts that programmatically label data. In our conversation, we discuss the original inspiration for Snorkel and some of the projects they've undertaken since it's inception. We also discuss some of the papers that have been presented at various conferences, that used Snorkel for training data, including Kunle Olokotun's "Software 2.0" presentation that we broke down in our 2018 NeurIPS series.
About the Guest
Alex Jason Ratner
Snorkel AI; University of Washington
Resources
"
- Snorkel
- Paper: Data Programming: Creating Large Training Sets, Quickly
- Babble Labble: Learning from Natural Language Explanations
- Paper: Training Classifiers with Natural Language Explanations
- Podcast: Approaches to Fairness in Machine Learning with Richard Zemel
- Podcast: Designing Computer Systems for Software with Kunle Olukotun
- Deep Dive
- DARPA Memex
- TWIML Presents: AI Platforms Vol. 1
