Fairness in Machine Learning with Hanna Wallach

EPISODE 232
|
FEBRUARY 20, 2019
Watch
Banner Image: Hanna Wallach - Podcast Interview
Don't Miss an Episode!  Join our mailing list for episode summaries and other updates.

About this Episode

Today we're joined by Hanna Wallach, a Principal Researcher at Microsoft Research. Hanna and I really dig into how bias and a lack of interpretability and transparency show up across machine learning. We discuss the role that human biases, even those that are inadvertent, play in tainting data, and whether deployment of "fair" ML models can actually be achieved in practice, and much more. Along the way, Hanna points us to a TON of papers and resources to further explore the topic of fairness in ML. You'll definitely want to check out the notes page for this episode, which you'll find at twimlai.com/talk/232.

About the Guest

Hanna Wallach

Microsoft Research

Connect with Hanna

Resources

Related Topics