Update JULAIN Talk Thijs Vogels 9 June 2022 authored by Susanne Wenzel's avatar Susanne Wenzel
[[Home](Home)]
---
**Thijs Vogels**
Machine Learning & Optimization Laboratory, École polytechnique fédérale de Lausanne (EPFL)
**Communication-efficient distributed learning and PowerSGD**
* When: 9 June 2022, 4pm<br>
* Where: virtual event
* video conf link:
https://us06web.zoom.us/j/84697003031?pwd=clpXUUtTUWtlYkI0MUE2bnJPTDl3Zz09
Meeting ID: 846 9700 3031 Passcode: 843njb
**Invitation and moderation:** Hanno Scharr, IAS-8
### Abstract
In data-parallel optimization of machine learning models, workers collaborate to speed up the training.
By averaging their model updates with each other, the updates become more informative, resulting in faster convergence.
For today’s deep learning models, model updates can be gigabytes large, and averaging them between all workers can be a bottleneck in the scalability of distributed learning.
In this talk, we explore two approaches to alleviating communication bottlenecks: lossy communication compression and sparse (decentralized) communication.
We focus on the PowerSGD communication compression algorithm which approximates gradient updates as low-rank matrices.
PowerSGD can yield communication savings of > 100x and was used successfully to speed up the training of OpenAI’s DALL-E, RoBERTa, and Meta’s XLM-R.
### Short CV
Thijs is a PhD student at EPFL’s Machine Learning & Optimization Laboratory under Martin Jaggi.
He works on developing and understanding practical optimization algorithms for large-scale distributed training of deep learning models.
https://arxiv.org/pdf/1905.13727.pdf
https://arxiv.org/pdf/2008.01425.pdf
https://arxiv.org/pdf/2110.04175.pdf
---
[[Home](Home)]
\ No newline at end of file