Update JULAIN Talk Thijs Vogels 9 June 2022 authored by Susanne Wenzel's avatar Susanne Wenzel
...@@ -31,8 +31,8 @@ PowerSGD can yield communication savings of > 100x and was used successfully to ...@@ -31,8 +31,8 @@ PowerSGD can yield communication savings of > 100x and was used successfully to
Thijs is a PhD student at EPFL’s Machine Learning & Optimization Laboratory under Martin Jaggi. Thijs is a PhD student at EPFL’s Machine Learning & Optimization Laboratory under Martin Jaggi.
He works on developing and understanding practical optimization algorithms for large-scale distributed training of deep learning models. He works on developing and understanding practical optimization algorithms for large-scale distributed training of deep learning models.
*Readings:* ###Readings:
* [RelaySum for Decentralized Deep Learning on Heterogeneous Data](https://arxiv.org/pdf/2110.04175.pdf), NeurIPS 2020 * [RelaySum for Decentralized Deep Learning on Heterogeneous Data](https://arxiv.org/pdf/2110.04175.pdf), NeurIPS 2021
* [Practical Low-Rank Communication Compression in Decentralized Deep Learning](https://arxiv.org/pdf/2008.01425.pdf ), NeurIPS 2020 * [Practical Low-Rank Communication Compression in Decentralized Deep Learning](https://arxiv.org/pdf/2008.01425.pdf ), NeurIPS 2020
* [PowerSGD: Practical Low-Rank Gradient Compression for Distributed Optimization](https://arxiv.org/pdf/1905.13727.pdf), NeurIPS 2019 * [PowerSGD: Practical Low-Rank Gradient Compression for Distributed Optimization](https://arxiv.org/pdf/1905.13727.pdf), NeurIPS 2019
... ...
......