stat940F21: Difference between revisions
Jump to navigation
Jump to search
No edit summary |
|||
Line 19: | Line 19: | ||
|Week of Nov 11 || || || || || | |Week of Nov 11 || || || || || | ||
|- | |- | ||
|Week of Nov 11 || | |Week of Nov 11 || Benyamin Jamialahmad || || Perceiver: General Perception with Iterative Attention || [https://arxiv.org/abs/2103.03206] ||[https://www.youtube.com/watch?v=NmzsUgfzor0] || | ||
|- | |- | ||
|Week of Nov 11 || || || || || | |Week of Nov 11 || || || || || | ||
Line 48: | Line 48: | ||
|Week of Nov 25 || Alice Leung || ||ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators || [https://openreview.net/pdf?id=r1xMH1BtvB Paper]|| | |Week of Nov 25 || Alice Leung || ||ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators || [https://openreview.net/pdf?id=r1xMH1BtvB Paper]|| | ||
|- | |- | ||
|Week of Nov 25 || Maryam Yalsavar || || Knowledge Extraction with No Observable Data || http://papers.neurips.cc/paper/8538-knowledge-extraction-with-no-observable-data.pdf || | |Week of Nov 25 || Maryam Yalsavar || || Knowledge Extraction with No Observable Data || [http://papers.neurips.cc/paper/8538-knowledge-extraction-with-no-observable-data.pdf] || | ||
|- | |- | ||
|Week of Nov 25 || Shervin Hakimi || || What Do Neural Networks Learn When Trained With Random Labels? || [https://proceedings.neurips.cc/paper/2020/file/e4191d610537305de1d294adb121b513-Paper.pdf] || | |Week of Nov 25 || Shervin Hakimi || || What Do Neural Networks Learn When Trained With Random Labels? || [https://proceedings.neurips.cc/paper/2020/file/e4191d610537305de1d294adb121b513-Paper.pdf] || |
Revision as of 17:38, 15 November 2021
Project Proposal
Paper presentation
Date | Name | Paper number | Title | Link to the paper | Link to the video | |
Week of Nov 8 | Abhinav Chanana (Example) | 1 | AUGMIX: A Simple Data Procession method to Improve Robustness And Uncertainity | Paper | Presentation | |
Week of Nov 11 | ||||||
Week of Nov 11 | Benyamin Jamialahmad | Perceiver: General Perception with Iterative Attention | [1] | [2] | ||
Week of Nov 11 | ||||||
Week of Nov 11 | ||||||
Week of Nov 11 | ||||||
Week of Nov 11 | ||||||
Week of Nov 18 | Yuliang Shi | |||||
Week of Nov 18 | Veronica Salm | StructBERT: Incorporating Language Structures Into Pre-Training for Deep Language Understanding | Paper | |||
Week of Nov 18 | Youssef Fathi | NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis | Paper | |||
Week of Nov 18 | Wei Liang | Deep Stable Learning for Out-Of-Distribution Generalization (Received by CVPR2021) | [3] | |||
Week of Nov 18 | Taj Jones-McCormick | Watch out! Motion is Blurring the Vision of Your Deep Neural Networks | https://papers.nips.cc/paper/2020/file/0a73de68f10e15626eb98701ecf03adb-Paper.pdf | [4] | ||
Week of Nov 18 | Muhammad Maruf Sazed | An unsupervised deep learning approach for real-world image denoising | https://openreview.net/pdf?id=tIjRAiFmU3y | |||
Week of Nov 25 | Yuxiang Huang | |||||
Week of Nov 25 | Varnan Sarangian | Big Bird: Transformers for Longer Sequences | [5] | |||
Week of Nov 25 | Alice Leung | ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators | Paper | |||
Week of Nov 25 | Maryam Yalsavar | Knowledge Extraction with No Observable Data | [6] | |||
Week of Nov 25 | Shervin Hakimi | What Do Neural Networks Learn When Trained With Random Labels? | [7] | |||
Week of Nov 25 | Islam Nasr | |||||
Week of Nov 25 | Jared Feng | The Autoencoding Variational Autoencoder | https://nips.cc/virtual/2020/public/poster_ac10ff1941c540cd87c107330996f4f6.html | |||
Week of Nov 25 | Mina Kebriaee | |||||
Week of Nov 25 | Mehrshad Sadria | scGen predicts single-cell perturbation responses | https://www.nature.com/articles/s41592-019-0494-8 | |||
Week of Nov 25 | Xuanzhi Huang |