SLIDE: Smart Algorithms over Hardware Acceleration for Large-Scale Deep Learning with Beidi Chen

EPISODE 356
|
MARCH 12, 2020
Watch
Banner Image: Beidi Chen - Podcast Interview
Don't Miss an Episode!  Join our mailing list for episode summaries and other updates.

About this Episode

Today we're joined by Beidi Chen, a PhD student at Rice University, who along with her co-contributors, developed a cheaper, algorithmic, CPU alternative to state-of-the-art GPU machines in their paper, SLIDE: In Defense of Smart Algorithms Over Hardware Acceleration for Large-Scale Deep Learning Systems. The team aimed to solve the problem of computational power that often creates a bottleneck for training large-scale neural networks. "We usually have a brute force way of doing things...or the time to achieve it...But the problem is we usually don't have that time and computational power." Their solution is an algorithm called SLIDE (Sub-LInear Deep learning Engine) which combines a few elements: Randomized Algorithms (applies randomness in its logic); Multicore Parallelism (performing several computations at the same time); and Workload Optimization (maximum performance for data processing). Check out our full write-up on this interview here.

About the Guest

Beidi Chen

Rice University

Connect with Beidi

Resources

Related Topics