Those two papers are about learning discrete representations from data by taking inspiration from vector quantization. Learning discrete representations using neural networks is challenging and can helpful for tasks such as compression, planning, reasoning, and can be potentially more interpretable than continuous ones. In the two papers use those learned discrete representations to build autoregressive generative models on image, sound, and video. The second paper (Generating Diverse High-Fidelity Images with VQ-VAE-2) is basically a sequel of the first (Neural Discrete Representation Learning) where they scale the models to bigger datasets and images (up to 1024x1024 resolution).
Those two papers are about learning discrete representations from data by taking inspiration from vector quantization. Learning discrete representations using neural networks is challenging and can helpful for tasks such as compression, planning, reasoning, and can be potentially more interpretable than continuous ones. In the two papers use those learned discrete representations to build autoregressive generative models on image, sound, and video. The second paper (Generating Diverse High-Fidelity Images with VQ-VAE-2) is basically a sequel of the first (Neural Discrete Representation Learning) where they scale the models to bigger datasets and images (up to 1024x1024 resolution).
### Monday 30 March 10-11:30am - Full-Resolution Residual Networks for Semantic Segmentation
### Monday 30 March - Full-Resolution Residual Networks for Semantic Segmentation
**Replacement date for canceled meeting at Monday 16 March 10-11:30am**
virtual meeting via dfnconf: https://conf.dfn.de/webapp/conference/97977564<br>
**Replacement date for canceled meeting at Monday 16 March**
alternative link, if dfn is down: https://us04web.zoom.us/j/433015211
Venue: **INM-1 Seminar room**, building 15.9, room 4001b
* Full-Resolution Residual Networks for Semantic Segmentation in Street Scenes (CVPR'17 Oral)<br>
* Full-Resolution Residual Networks for Semantic Segmentation in Street Scenes (CVPR'17 Oral)<br>
T. Pohlen, A. Hermans, M. Mathias, and B. Leibe<br>
T. Pohlen, A. Hermans, M. Mathias, and B. Leibe<br>
Most common architectures for semantic segmentation consisting of an encoder and decoder part (e.g. U-Net) heavily reduce the spatial dimension of input images and may loose important details or fail to localize precisely. The proposed papers present full-resolution networks, which try to preserve high-resolution features throughout the network and improve localization accuracy.
Most common architectures for semantic segmentation consisting of an encoder and decoder part (e.g. U-Net) heavily reduce the spatial dimension of input images and may loose important details or fail to localize precisely. The proposed papers present full-resolution networks, which try to preserve high-resolution features throughout the network and improve localization accuracy.
### Monday 17 February 10-11:30am - Speech Recognition
### Monday 17 February - Speech Recognition
Venue: **INM-1 Seminar room**, building 15.9, room 4001b
* Deep Speech 2: End-to-End Speech Recognition in English and Mandarin, Amodei et al., 2015<br>
* Deep Speech 2: End-to-End Speech Recognition in English and Mandarin, Amodei et al., 2015<br>
https://arxiv.org/abs/1512.02595<br>
https://arxiv.org/abs/1512.02595<br>
...
@@ -84,20 +125,16 @@ Speech Recognition with Deep Recurrent Neural Networks<br>
...
@@ -84,20 +125,16 @@ Speech Recognition with Deep Recurrent Neural Networks<br>
Alex Graves, Abdel-rahman Mohamed, Geoffrey Hinton, 2013<br>
Alex Graves, Abdel-rahman Mohamed, Geoffrey Hinton, 2013<br>
https://arxiv.org/abs/1303.5778
https://arxiv.org/abs/1303.5778
### Monday 20 January 10-11:30am
### Monday 20 January
Venue: **INM-1 Seminar room**, building 15.9, room 4001b
* Multi-Context Recurrent Neural Networks for Time Series Applications <br>
* Multi-Context Recurrent Neural Networks for Time Series Applications <br>
https://publications.waset.org/3524/pdf
https://publications.waset.org/3524/pdf
* Global Sparse Momentum SGD for Pruning Very Deep Neural Networks <br>
* Global Sparse Momentum SGD for Pruning Very Deep Neural Networks <br>
https://arxiv.org/pdf/1909.12778v3.pdf
https://arxiv.org/pdf/1909.12778v3.pdf
### Monday 16 December 10-11am
### Monday 16 December
Venue: **JSC, Rotunda**, building 16.4, room 301
Our first Journal Club will cover two papers from ICCV 2019 about GANs.
Our first Journal Club will cover two papers from ICCV 2019 about GANs.
* SinGAN: Learning a Generative Model from a Single Natural Image (Best Paper Award) <br> http://openaccess.thecvf.com/content_ICCV_2019/papers/Shaham_SinGAN_Learning_a_Generative_Model_From_a_Single_Natural_Image_ICCV_2019_paper.pdf
* SinGAN: Learning a Generative Model from a Single Natural Image (Best Paper Award) <br> http://openaccess.thecvf.com/content_ICCV_2019/papers/Shaham_SinGAN_Learning_a_Generative_Model_From_a_Single_Natural_Image_ICCV_2019_paper.pdf
* COCO-GAN: Generation by Parts via Conditional Coordinating <br>
* COCO-GAN: Generation by Parts via Conditional Coordinating <br>