Dissertation Talk: System Design for Large Scale Machine Learning
Seminar | August 28 | 11 a.m.-12 p.m. | 420 Soda Hall
Shivaram Venkataraman, UC Berkeley
Machine learning models trained on massive datasets power a number of applications; from machine translation to detecting supernovae in astrophysics. However the end of Moores law and the shift towards distributed computing architectures presents many new challenges for building and executing such applications in a scalable fashion.
In this talk I will present my research on systems that make it easier to develop new machine learning applications and scale them while achieving high performance. I will first present programming models that let users easily build distributed machine learning applications. Next, I will show how we can exploit the structure of machine learning workloads to build low-overhead performance models that can help users understand scalability and simplify large scale deployments. Finally, I will describe scheduling techniques that can improve scalability and achieve high performance when using distributed data processing frameworks.