- Adam Paszke
- Aditya Ramesh
- Agata Mosinska
- Alec Radford
- Aleksander Madry
- Alexandre Alahi
- Alexei A. Efros
- Alexey Dosovitskiy
- Alexis Conneau
- Alfredo Canziani
- Ali Farhadi
- Alimohammad Beigi
- Anant Jain
- Andreas Mayr
- Andrew G. Howard
- Andrew Ilyas
- Anselm Levskaya
- Antoine Bosselut
- Ari S. Morcos
- Ashish Vaswani
- Asim Kadav
- Augustus Odena
- Bharath Hariharan
- Boaz Barak
- Carroll L. Wainwright
- Chandra Bhagavatula
- Christian Buck
- Christian Szegedy
- Colin Raffel
- Danqi Chen
- Dario Amodei
- David Berthelot
- David J. Schwab
- David Silver
- Davis Wertheimer
- Dedy Kredo
- Demis Hassabis
- Denny Zhou
- Dian Yu
- Diederik P. Kingma
- Dimitris Metaxas
- Dimitris Tsipras
- Diogo Almeida
- Diyi Yang
- Dmytro Mishkin
- Eugenio Culurciello
- Fangxiaoyu Feng
- Forrest N. Iandola
- Gabriel Goh
- Gal Kaplun
- Geoffrey E Hinton
- Greg Corrado
- Guillaume Lample
- Günter Klambauer
- Han Zhang
- Hannah Rashkin
- Hao Li
- Hao Tan
- Hattie Zhou
- Hyung Won Chung
- Ian Goodfellow
- Iftekhar Naim
- Illia Polosukhin
- Ilya Sutskever
- Ioannis Antonoglou
- Itamar Friedman
- Izhak Shafran
- Jaakko Lehtinen
- Jack Hessel
- Jacob Devlin
- Jae Sung Park
- Jaegul Choo
- Jaemin Cho
- Jakob Uszkoreit
- James J. Little
- Jan Leike
- Janice Lan
- Jared Kaplan
- Jason Wei
- Jason Weston
- Jason Yosinski
- Javier Romero
- Jeff Wu
- Jeffrey Dean
- Jeffrey Wu
- Jeffrey Zhao
- Jessica B. Hamrick
- Jiakai Zhang
- Jie Le
- Jimmy Ba
- Jiri Matas
- Jitendra Malik
- Jonathan Frankle
- Jong Wook Kim
- Jordan Hoffmann
- Julian Schrittwiese
- Julieta Martinez
- Jun Huang
- Jun-Yan Zhu
- Jung-Woo Ha
- Justin Johnson
- Kai Chen
- Kaiming He
- Karthik Narasimhan
- Karthik Raman
- Ke Li
- Kenton Lee
- Krishna Srinivasan
- Kristen Grauman
- Kristina Toutanova
- Larry Davis
- Laurent Sifre
- Le Hou
- Leonard Adolphs
- Li Fei-Fei
- Li Yang
- Long Ouyang
- Lucas Beyer
- Ludovic Denoyer
- Lukasz Kaiser
- Luke Metz
- Luming Tang
- Léon Bottou
- Maarten Sap
- Mandar Joshi
- Marc'Aurelio Ranzato
- Mario Fritz
- Martin Arjovsky
- Masanori Koyama
- Mateusz Koziński
- Matthew D. Zeiler
- Matthias Bethge
- Mengli Cheng
- Menglin Jia
- Menglong Zhu
- Michael Carbin
- Michelle Chen Huebscher
- Mike Lewis
- Mikhail Pavlov
- Ming-Wei Chang
- Minghui Qiu
- Minje Choi
- Minjun Li
- Mohit Bansal
- Munyoung Kim
- Myle Ott
- Nan Du
- Neil Houlsby
- Nicholas Frosst
- Nikita Kitaev
- Ning Yu
- Noah A. Smith
- Noam Shazeer
- Oriol Vinyals
- Other
- Pablo Márquez-Neila
- Pamela Mishkin
- Pascal Fua
- Paul Christiano
- Peng Zhou
- Peter J. Liu
- Peter W. Battaglia
- Peter West
- Peter Wonka
- Phillip Isola
- Ping Yu
- Preetum Nakkiran
- Qifan Wang
- Quoc V. Le
- Rameen Abdal
- Rayat Hossain
- Rob Fergus
- Ronan Le Bras
- Rosanne Liu
- Ryan Lowe
- Samuli Laine
- Santhosh K. Ramakrishnan
- Sara Sabour
- Scott Gray
- Sebastian Borgeaud
- Serge Belongie
- Shayne Longpre
- Shibani Santurkar
- Shiyu Chang
- Shunyu Yao
- Song Han
- Soumith Chintala
- Sowmya Yellapragada
- Stanislas Polu
- Stephen Merity
- Sudharshan Chandra Babu
- Sunghun Kim
- Taesung Park
- Takeru Miyato
- Takeshi Kojima
- Tal Ridnik
- Tero Karras
- Thomas Schumm
- Thomas Unterthiner
- Tim Salimans
- Timo Aila
- Timothy Dozat
- Tinghui Zhou
- Tomas Mikolov
- Tong He
- Toshiki Kataoka
- Tristan Yang
- Vicki Cheung
- Wei Lin
- Wei Liu
- Wieland Brendel
- Wojciech Zaremba
- Xi Chen
- Xian Li
- Xiangyu Zhang
- Xiaofei Sun
- Xing Shi
- Xu Jiang
- Xuezhi Wang
- Xun Huang
- Yamini Bansal
- Yanghua Jin
- Yanqi Zhou
- Yejin Choi
- Ygor Rebouças Serpa
- Yifan Jiang
- Yin Cui
- Yinfei Yang
- Yipeng Qin
- Yuan Cao
- Yuichi Yoshida
- Yunjey Choi
- Yuntao Ba
- Yusuke Iwasawa
- Yuxin Wu
- Zhangyang Wang
- Zhen Tan
- Zhi Zhang
- Ziad Al-Halah
- Łukasz Kaiser
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Evaluation of neural network architectures for embedded systems | Read | CNNs, CV , Image | Comparison | 2017 | IEEE ISCAS | Adam Paszke, Alfredo Canziani, Eugenio Culurciello | Compare CNN classification architectures on accuracy, memory footprint, parameters, operations count, inference time and power consumption. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | DALL·E: Creating Images from Text | Pending | Image , Text , Transformers | 2021 | Blog | Aditya Ramesh, Gabriel Goh, Ilya Sutskever, Mikhail Pavlov, Scott Gray | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Topological Loss: Beyond the Pixel-Wise Loss for Topology-Aware Delineation | Pending | Image , Loss Function, Segmentation | 2018 | CVPR | Agata Mosinska, Mateusz Koziński, Pablo Márquez-Neila, Pascal Fua | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | GPT-2 (Language Models are Unsupervised Multitask Learners) | Pending | Attention, Text , Transformers | 2019 | Alec Radford, Dario Amodei, Ilya Sutskever, Jeffrey Wu | link | |||
1 | Improved Techniques for Training GANs | Pending | GANs, Image | Semi-Supervised | 2016 | NIPS | Alec Radford, Ian Goodfellow, Tim Salimans, Vicki Cheung, Wojciech Zaremba, Xi Chen | link | |
2 | CLIP: Connecting Text and Images | Pending | Image , Text , Transformers | Multimodal, Pre-Training | 2021 | arXiv | Alec Radford, Ilya Sutskever, Jong Wook Kim | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | How Does Batch Normalization Help Optimization? | Pending | NNs, Normalization | Optimizations | 2018 | arXiv | Aleksander Madry, Andrew Ilyas, Dimitris Tsipras, Shibani Santurkar | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Perceptual Losses for Real-Time Style Transfer and Super-Resolution | Pending | Loss Function, NNs | 2016 | ECCV | Alexandre Alahi, Justin Johnson, Li Fei-Fei | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Pix2Pix: Image-to-Image Translation with Conditional Adversarial Nets | Read | GANs, Image | 2017 | CVPR | Alexei A. Efros, Jun-Yan Zhu, Phillip Isola, Tinghui Zhou | Image to image translation using Conditional GANs and dataset of image pairs from one domain to another. | link | |
1 | CycleGAN: Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks | Pending | GANs, Image | Architecture | 2017 | ICCV | Alexei A. Efros, Jun-Yan Zhu, Phillip Isola, Taesung Park | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Vision Transformer: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale | Pending | Attention, Image , Transformers | 2021 | ICLR | Alexey Dosovitskiy, Jakob Uszkoreit, Lucas Beyer, Neil Houlsby | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Phrase-Based & Neural Unsupervised Machine Translation | Pending | NMT, Text , Transformers | Unsupervised | 2018 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link | |
1 | Unsupervised Machine Translation Using Monolingual Corpora Only | Pending | GANs, NMT, Text , Transformers | Unsupervised | 2017 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link | |
2 | Cross-lingual Language Model Pretraining | Pending | NMT, Text , Transformers | Unsupervised | 2019 | arXiv | Alexis Conneau, Guillaume Lample | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Evaluation of neural network architectures for embedded systems | Read | CNNs, CV , Image | Comparison | 2017 | IEEE ISCAS | Adam Paszke, Alfredo Canziani, Eugenio Culurciello | Compare CNN classification architectures on accuracy, memory footprint, parameters, operations count, inference time and power consumption. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | VisualCOMET: Reasoning about the Dynamic Context of a Still Image | Pending | AGI, Dataset, Image , Text , Transformers | 2020 | ECCV | Ali Farhadi, Chandra Bhagavatula, Jae Sung Park, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Large Language Models for Data Annotation: A Survey | This week | Dataset, Generative, Large-Language-Models | Prompting, Tips & Tricks | 2024 | arXiv | Alimohammad Beigi, Zhen Tan | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Breaking neural networks with adversarial attacks | Pending | CNNs, Image | Adversarial | 2019 | Blog | Anant Jain | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Self-Normalizing Neural Networks | Pending | Activation Function, Tabular | Optimizations, Tips & Tricks | 2017 | NIPS | Andreas Mayr, Günter Klambauer, Thomas Unterthiner | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | MobileNet (Efficient Convolutional Neural Networks for Mobile Vision Applications) | Pending | CNNs, CV , Image | Architecture, Optimization-No. of params | 2017 | arXiv | Andrew G. Howard, Menglong Zhu | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | How Does Batch Normalization Help Optimization? | Pending | NNs, Normalization | Optimizations | 2018 | arXiv | Aleksander Madry, Andrew Ilyas, Dimitris Tsipras, Shibani Santurkar | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Reformer: The Efficient Transformer | Read | Attention, Text , Transformers | Architecture, Optimization-Memory, Optimization-No. of params | 2020 | arXiv | Anselm Levskaya, Lukasz Kaiser, Nikita Kitaev | Overcome time and memory complexity of Transformers by bucketing Query, Keys and using Reversible residual connections. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | COMET: Commonsense Transformers for Automatic Knowledge Graph Construction | Pending | AGI, Text , Transformers | 2019 | ACL | Antoine Bosselut, Hannah Rashkin, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training BatchNorm and Only BatchNorm: On the Expressive Power of Random Features in CNNs | Pending | CNNs, Image | 2020 | arXiv | Ari S. Morcos, David J. Schwab, Jonathan Frankle | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Attention is All you Need | Read | Attention, Text , Transformers | Architecture | 2017 | NIPS | Ashish Vaswani, Illia Polosukhin, Noam Shazeer, Łukasz Kaiser | Talks about Transformer architecture which brings SOTA performance for different tasks in NLP | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Pruning Filters for Efficient ConvNets | Pending | CNNs, CV , Image | Optimization-No. of params | 2017 | arXiv | Asim Kadav, Hao Li | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | SAGAN: Self-Attention Generative Adversarial Networks | Pending | Attention, GANs, Image | Architecture | 2018 | arXiv | Augustus Odena, Dimitris Metaxas, Han Zhang, Ian Goodfellow | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Few-Shot Learning with Localization in Realistic Settings | Pending | CNNs, Image | Few-shot-learning | 2019 | CVPR | Bharath Hariharan, Davis Wertheimer | link | |
1 | Revisiting Pose-Normalization for Fine-Grained Few-Shot Recognition | Pending | CNNs, Image | Few-shot-learning | 2020 | CVPR | Bharath Hariharan, Davis Wertheimer, Luming Tang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Deep Double Descent: Where Bigger Models and More Data Hurt | Pending | NNs | 2019 | arXiv | Boaz Barak, Gal Kaplun, Ilya Sutskever, Preetum Nakkiran, Tristan Yang, Yamini Bansal | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training language models to follow instructions with human feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Semi-Supervised | 2022 | arXiv | Carroll L. Wainwright, Diogo Almeida, Jan Leike, Jeff Wu, Long Ouyang, Pamela Mishkin, Paul Christiano, Ryan Lowe, Xu Jiang | This paper presents InstructGPT, a model fine-tuned with human feedback to better align with user intent across various tasks. Despite having significantly fewer parameters than larger models, InstructGPT outperforms them in human evaluations, demonstrating improved truthfulness, reduced toxicity, and minimal performance regressions on public NLP datasets, highlighting the potential of fine-tuning with human feedback for enhancing language model alignment with human intent. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | VisualCOMET: Reasoning about the Dynamic Context of a Still Image | Pending | AGI, Dataset, Image , Text , Transformers | 2020 | ECCV | Ali Farhadi, Chandra Bhagavatula, Jae Sung Park, Yejin Choi | link | ||
1 | Symbolic Knowledge Distillation: from General Language Models to Commonsense Models | Pending | Dataset, Text , Transformers | Optimizations, Tips & Tricks | 2021 | arXiv | Chandra Bhagavatula, Jack Hessel, Peter West, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Decoding a Neural Retriever’s Latent Space for Query Suggestion | Pending | Text | Embeddings, Latent space | 2022 | arXiv | Christian Buck, Leonard Adolphs, Michelle Chen Huebscher | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Inception-v1 (Going Deeper With Convolutions) | Read | CNNs, CV , Image | Architecture | 2015 | CVPR | Christian Szegedy, Wei Liu | Propose the use of 1x1 conv operations to reduce the number of parameters in a deep and wide CNN | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer | Read | Attention, Text , Transformers | 2020 | JMLR | Colin Raffel, Noam Shazeer, Peter J. Liu, Wei Liu, Yanqi Zhou | Presents a Text-to-Text transformer model with multi-task learning capabilities, simultaneously solving problems such as machine translation, document summarization, question answering, and classification tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | SpanBERT: Improving Pre-training by Representing and Predicting Spans | Read | Question-Answering, Text , Transformers | Pre-Training | 2020 | TACL | Danqi Chen, Mandar Joshi | A different pre-training strategy for BERT model to improve performance for Question Answering task. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | GPT-2 (Language Models are Unsupervised Multitask Learners) | Pending | Attention, Text , Transformers | 2019 | Alec Radford, Dario Amodei, Ilya Sutskever, Jeffrey Wu | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | BEGAN: Boundary Equilibrium Generative Adversarial Networks | Pending | GANs, Image | 2017 | arXiv | David Berthelot, Luke Metz, Thomas Schumm | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training BatchNorm and Only BatchNorm: On the Expressive Power of Random Features in CNNs | Pending | CNNs, Image | 2020 | arXiv | Ari S. Morcos, David J. Schwab, Jonathan Frankle | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | MuZero: Mastering Go, chess, shogi and Atari without rules | Pending | Reinforcement-Learning | 2020 | Nature | David Silver, Demis Hassabis, Ioannis Antonoglou, Julian Schrittwiese | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Few-Shot Learning with Localization in Realistic Settings | Pending | CNNs, Image | Few-shot-learning | 2019 | CVPR | Bharath Hariharan, Davis Wertheimer | link | |
1 | Revisiting Pose-Normalization for Fine-Grained Few-Shot Recognition | Pending | CNNs, Image | Few-shot-learning | 2020 | CVPR | Bharath Hariharan, Davis Wertheimer, Luming Tang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering | Pending | Large-Language-Models | Prompting, Tips & Tricks | 2024 | arXiv | Dedy Kredo, Itamar Friedman, Tal Ridnik | This paper introduces AlphaCodium, a novel test-based, multi-stage, code-oriented iterative approach for improving the performance of Language Model Models (LLMs) on code generation tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | MuZero: Mastering Go, chess, shogi and Atari without rules | Pending | Reinforcement-Learning | 2020 | Nature | David Silver, Demis Hassabis, Ioannis Antonoglou, Julian Schrittwiese | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Chain of Thought Prompting Elicits Reasoning in Large Language Models | Pending | Question-Answering, Text , Transformers | 2022 | arXiv | Denny Zhou, Jason Wei, Xuezhi Wang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ReAct: Synergizing Reasoning and Acting in Language Models | Pending | Generative, Large-Language-Models, Text | Optimizations, Tips & Tricks | 2023 | ICLR | Dian Yu, Izhak Shafran, Jeffrey Zhao, Karthik Narasimhan, Nan Du, Shunyu Yao, Yuan Cao | This paper introduces ReAct, a novel approach that leverages Large Language Models (LLMs) to interleave reasoning traces and task-specific actions. ReAct outperforms existing methods on various language and decision-making tasks, addressing issues like hallucination, error propagation, and improving human interpretability and trustworthiness. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Adam: A Method for Stochastic Optimization | Pending | NNs, Optimizers | 2015 | ICLR | Diederik P. Kingma, Jimmy Ba | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | SAGAN: Self-Attention Generative Adversarial Networks | Pending | Attention, GANs, Image | Architecture | 2018 | arXiv | Augustus Odena, Dimitris Metaxas, Han Zhang, Ian Goodfellow | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | How Does Batch Normalization Help Optimization? | Pending | NNs, Normalization | Optimizations | 2018 | arXiv | Aleksander Madry, Andrew Ilyas, Dimitris Tsipras, Shibani Santurkar | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training language models to follow instructions with human feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Semi-Supervised | 2022 | arXiv | Carroll L. Wainwright, Diogo Almeida, Jan Leike, Jeff Wu, Long Ouyang, Pamela Mishkin, Paul Christiano, Ryan Lowe, Xu Jiang | This paper presents InstructGPT, a model fine-tuned with human feedback to better align with user intent across various tasks. Despite having significantly fewer parameters than larger models, InstructGPT outperforms them in human evaluations, demonstrating improved truthfulness, reduced toxicity, and minimal performance regressions on public NLP datasets, highlighting the potential of fine-tuning with human feedback for enhancing language model alignment with human intent. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Interpreting Deep Learning Models in Natural Language Processing: A Review | Pending | Text | Comparison, Visualization | 2021 | arXiv | Diyi Yang, Xiaofei Sun | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | All you need is a good init | Pending | NN Initialization | Tips & Tricks | 2015 | arXiv | Dmytro Mishkin, Jiri Matas | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Evaluation of neural network architectures for embedded systems | Read | CNNs, CV , Image | Comparison | 2017 | IEEE ISCAS | Adam Paszke, Alfredo Canziani, Eugenio Culurciello | Compare CNN classification architectures on accuracy, memory footprint, parameters, operations count, inference time and power consumption. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Language-Agnostic BERT Sentence Embedding | Read | Attention, Siamese Network, Text , Transformers | Embeddings | 2020 | arXiv | Fangxiaoyu Feng, Yinfei Yang | A BERT model with multilingual sentence embeddings learned over 112 languages and Zero-shot learning over unseen languages. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | SqueezeNet | Read | CNNs, CV , Image | Architecture, Optimization-No. of params | 2016 | arXiv | Forrest N. Iandola, Song Han | Explores model compression by using 1x1 convolutions called fire modules. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | DALL·E: Creating Images from Text | Pending | Image , Text , Transformers | 2021 | Blog | Aditya Ramesh, Gabriel Goh, Ilya Sutskever, Mikhail Pavlov, Scott Gray | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Deep Double Descent: Where Bigger Models and More Data Hurt | Pending | NNs | 2019 | arXiv | Boaz Barak, Gal Kaplun, Ilya Sutskever, Preetum Nakkiran, Tristan Yang, Yamini Bansal | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Capsule Networks: Dynamic Routing Between Capsules | Pending | CV , Image | Architecture | 2017 | arXiv | Geoffrey E Hinton, Nicholas Frosst, Sara Sabour | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Word2Vec: Efficient Estimation of Word Representations in Vector Space | Pending | Text | Embeddings, Tips & Tricks | 2013 | arXiv | Greg Corrado, Jeffrey Dean, Kai Chen, Tomas Mikolov | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Phrase-Based & Neural Unsupervised Machine Translation | Pending | NMT, Text , Transformers | Unsupervised | 2018 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link | |
1 | Unsupervised Machine Translation Using Monolingual Corpora Only | Pending | GANs, NMT, Text , Transformers | Unsupervised | 2017 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link | |
2 | Cross-lingual Language Model Pretraining | Pending | NMT, Text , Transformers | Unsupervised | 2019 | arXiv | Alexis Conneau, Guillaume Lample | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Self-Normalizing Neural Networks | Pending | Activation Function, Tabular | Optimizations, Tips & Tricks | 2017 | NIPS | Andreas Mayr, Günter Klambauer, Thomas Unterthiner | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | SAGAN: Self-Attention Generative Adversarial Networks | Pending | Attention, GANs, Image | Architecture | 2018 | arXiv | Augustus Odena, Dimitris Metaxas, Han Zhang, Ian Goodfellow | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | COMET: Commonsense Transformers for Automatic Knowledge Graph Construction | Pending | AGI, Text , Transformers | 2019 | ACL | Antoine Bosselut, Hannah Rashkin, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Pruning Filters for Efficient ConvNets | Pending | CNNs, CV , Image | Optimization-No. of params | 2017 | arXiv | Asim Kadav, Hao Li | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Vokenization: Improving Language Understanding with Contextualized, Visual-Grounded Supervision | This week | Image , Text , Transformers | Multimodal | 2020 | EMNLP | Hao Tan, Mohit Bansal | link | |
1 | VL-T5: Unifying Vision-and-Language Tasks via Text Generation | Read | CNNs, CV , Generative, Image , Large-Language-Models, Question-Answering, Text , Transformers | Architecture, Embeddings, Multimodal, Pre-Training | 2021 | arXiv | Hao Tan, Jaemin Cho, Jie Le, Mohit Bansal | Unifying two modalities (image and text) together in a single transformer model to solve multiple tasks in a single architecture using text prefixes similar to T5. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask | Read | NN Initialization, NNs | Comparison, Optimization-No. of params, Tips & Tricks | 2019 | NeurIPS | Hattie Zhou, Janice Lan, Jason Yosinski, Rosanne Liu | Follow up on Lottery Ticket Hypothesis exploring the effects of different Masking criteria as well as Mask-1 and Mask-0 actions. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Flan-T5: Scaling Instruction-Finetuned Language Models | Pending | Generative, Text , Transformers | Architecture, Pre-Training | 2022 | arXiv | Hyung Won Chung, Le Hou | link | |
1 | Scaling Instruction-Finetuned Language Models (FLAN) | Pending | Generative, Large-Language-Models, Question-Answering, Text , Transformers | Instruction-Finetuning | 2022 | arXiv | Hyung Won Chung, Jason Wei, Jeffrey Dean, Le Hou, Quoc V. Le, Shayne Longpre | https://arxiv.org/abs/2210.11416 introduces FLAN (Fine-tuned LAnguage Net), an instruction finetuning method, and presents the results of its application. The study demonstrates that by fine-tuning the 540B PaLM model on 1836 tasks while incorporating Chain-of-Thought Reasoning data, FLAN achieves improvements in generalization, human usability, and zero-shot reasoning over the base model. The paper also provides detailed information on how each these aspects was evaluated. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | SAGAN: Self-Attention Generative Adversarial Networks | Pending | Attention, GANs, Image | Architecture | 2018 | arXiv | Augustus Odena, Dimitris Metaxas, Han Zhang, Ian Goodfellow | link | |
1 | Improved Techniques for Training GANs | Pending | GANs, Image | Semi-Supervised | 2016 | NIPS | Alec Radford, Ian Goodfellow, Tim Salimans, Vicki Cheung, Wojciech Zaremba, Xi Chen | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Transforming Sequence Tagging Into A Seq2Seq Task | Pending | Generative, Text | Comparison, Tips & Tricks | 2022 | arXiv | Iftekhar Naim, Karthik Raman, Krishna Srinivasan | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Attention is All you Need | Read | Attention, Text , Transformers | Architecture | 2017 | NIPS | Ashish Vaswani, Illia Polosukhin, Noam Shazeer, Łukasz Kaiser | Talks about Transformer architecture which brings SOTA performance for different tasks in NLP | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | GPT-2 (Language Models are Unsupervised Multitask Learners) | Pending | Attention, Text , Transformers | 2019 | Alec Radford, Dario Amodei, Ilya Sutskever, Jeffrey Wu | link | |||
1 | Deep Double Descent: Where Bigger Models and More Data Hurt | Pending | NNs | 2019 | arXiv | Boaz Barak, Gal Kaplun, Ilya Sutskever, Preetum Nakkiran, Tristan Yang, Yamini Bansal | link | ||
2 | GPT-f: Generative Language Modeling for Automated Theorem Proving | Pending | Attention, Transformers | 2020 | arXiv | Ilya Sutskever, Stanislas Polu | link | ||
3 | DALL·E: Creating Images from Text | Pending | Image , Text , Transformers | 2021 | Blog | Aditya Ramesh, Gabriel Goh, Ilya Sutskever, Mikhail Pavlov, Scott Gray | link | ||
4 | CLIP: Connecting Text and Images | Pending | Image , Text , Transformers | Multimodal, Pre-Training | 2021 | arXiv | Alec Radford, Ilya Sutskever, Jong Wook Kim | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | MuZero: Mastering Go, chess, shogi and Atari without rules | Pending | Reinforcement-Learning | 2020 | Nature | David Silver, Demis Hassabis, Ioannis Antonoglou, Julian Schrittwiese | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering | Pending | Large-Language-Models | Prompting, Tips & Tricks | 2024 | arXiv | Dedy Kredo, Itamar Friedman, Tal Ridnik | This paper introduces AlphaCodium, a novel test-based, multi-stage, code-oriented iterative approach for improving the performance of Language Model Models (LLMs) on code generation tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ReAct: Synergizing Reasoning and Acting in Language Models | Pending | Generative, Large-Language-Models, Text | Optimizations, Tips & Tricks | 2023 | ICLR | Dian Yu, Izhak Shafran, Jeffrey Zhao, Karthik Narasimhan, Nan Du, Shunyu Yao, Yuan Cao | This paper introduces ReAct, a novel approach that leverages Large Language Models (LLMs) to interleave reasoning traces and task-specific actions. ReAct outperforms existing methods on various language and decision-making tasks, addressing issues like hallucination, error propagation, and improving human interpretability and trustworthiness. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Progressive Growing of GANs for Improved Quality, Stability, and Variation | Pending | GANs, Image | Tips & Tricks | 2018 | ICLR | Jaakko Lehtinen, Samuli Laine, Tero Karras, Timo Aila | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Symbolic Knowledge Distillation: from General Language Models to Commonsense Models | Pending | Dataset, Text , Transformers | Optimizations, Tips & Tricks | 2021 | arXiv | Chandra Bhagavatula, Jack Hessel, Peter West, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding | Read | Attention, Text , Transformers | Embeddings | 2018 | NAACL | Jacob Devlin, Kenton Lee, Kristina Toutanova, Ming-Wei Chang | BERT is an extension to Transformer based architecture which introduces a masked word pretraining and next sentence prediction task to pretrain the model for a wide variety of tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | VisualCOMET: Reasoning about the Dynamic Context of a Still Image | Pending | AGI, Dataset, Image , Text , Transformers | 2020 | ECCV | Ali Farhadi, Chandra Bhagavatula, Jae Sung Park, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation | Pending | GANs, Image | 2018 | CVPR | Jaegul Choo, Jung-Woo Ha, Minje Choi, Munyoung Kim, Sunghun Kim, Yunjey Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | VL-T5: Unifying Vision-and-Language Tasks via Text Generation | Read | CNNs, CV , Generative, Image , Large-Language-Models, Question-Answering, Text , Transformers | Architecture, Embeddings, Multimodal, Pre-Training | 2021 | arXiv | Hao Tan, Jaemin Cho, Jie Le, Mohit Bansal | Unifying two modalities (image and text) together in a single transformer model to solve multiple tasks in a single architecture using text prefixes similar to T5. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Vision Transformer: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale | Pending | Attention, Image , Transformers | 2021 | ICLR | Alexey Dosovitskiy, Jakob Uszkoreit, Lucas Beyer, Neil Houlsby | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | A Simple yet Effective Baseline for 3D Human Pose Estimation | Pending | CV , Pose Estimation | 2017 | ICCV | James J. Little, Javier Romero, Julieta Martinez, Rayat Hossain | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training language models to follow instructions with human feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Semi-Supervised | 2022 | arXiv | Carroll L. Wainwright, Diogo Almeida, Jan Leike, Jeff Wu, Long Ouyang, Pamela Mishkin, Paul Christiano, Ryan Lowe, Xu Jiang | This paper presents InstructGPT, a model fine-tuned with human feedback to better align with user intent across various tasks. Despite having significantly fewer parameters than larger models, InstructGPT outperforms them in human evaluations, demonstrating improved truthfulness, reduced toxicity, and minimal performance regressions on public NLP datasets, highlighting the potential of fine-tuning with human feedback for enhancing language model alignment with human intent. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask | Read | NN Initialization, NNs | Comparison, Optimization-No. of params, Tips & Tricks | 2019 | NeurIPS | Hattie Zhou, Janice Lan, Jason Yosinski, Rosanne Liu | Follow up on Lottery Ticket Hypothesis exploring the effects of different Masking criteria as well as Mask-1 and Mask-0 actions. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Constitutional AI: Harmlessness from AI Feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Unsupervised | 2022 | arXiv | Jared Kaplan, Yuntao Ba | The paper introduces Constitutional AI, a method for training a safe AI assistant without human-labeled data on harmful outputs. It combines supervised learning and reinforcement learning phases, enabling the AI to engage with harmful queries by explaining its objections, thus improving control, transparency, and human-judged performance with minimal human oversight. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Chain of Thought Prompting Elicits Reasoning in Large Language Models | Pending | Question-Answering, Text , Transformers | 2022 | arXiv | Denny Zhou, Jason Wei, Xuezhi Wang | link | ||
1 | Scaling Instruction-Finetuned Language Models (FLAN) | Pending | Generative, Large-Language-Models, Question-Answering, Text , Transformers | Instruction-Finetuning | 2022 | arXiv | Hyung Won Chung, Jason Wei, Jeffrey Dean, Le Hou, Quoc V. Le, Shayne Longpre | https://arxiv.org/abs/2210.11416 introduces FLAN (Fine-tuned LAnguage Net), an instruction finetuning method, and presents the results of its application. The study demonstrates that by fine-tuning the 540B PaLM model on 1836 tasks while incorporating Chain-of-Thought Reasoning data, FLAN achieves improvements in generalization, human usability, and zero-shot reasoning over the base model. The paper also provides detailed information on how each these aspects was evaluated. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Self-Alignment with Instruction Backtranslation | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning | 2023 | arXiv | Jason Weston, Mike Lewis, Ping Yu, Xian Li | The paper introduces a scalable method called "instruction backtranslation" to create a high-quality instruction-following language model. This method involves self-augmentation and self-curation of training examples generated from web documents, resulting in a model that outperforms others in its category without relying on distillation data, showcasing its effective self-alignment capability. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask | Read | NN Initialization, NNs | Comparison, Optimization-No. of params, Tips & Tricks | 2019 | NeurIPS | Hattie Zhou, Janice Lan, Jason Yosinski, Rosanne Liu | Follow up on Lottery Ticket Hypothesis exploring the effects of different Masking criteria as well as Mask-1 and Mask-0 actions. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | A Simple yet Effective Baseline for 3D Human Pose Estimation | Pending | CV , Pose Estimation | 2017 | ICCV | James J. Little, Javier Romero, Julieta Martinez, Rayat Hossain | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training language models to follow instructions with human feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Semi-Supervised | 2022 | arXiv | Carroll L. Wainwright, Diogo Almeida, Jan Leike, Jeff Wu, Long Ouyang, Pamela Mishkin, Paul Christiano, Ryan Lowe, Xu Jiang | This paper presents InstructGPT, a model fine-tuned with human feedback to better align with user intent across various tasks. Despite having significantly fewer parameters than larger models, InstructGPT outperforms them in human evaluations, demonstrating improved truthfulness, reduced toxicity, and minimal performance regressions on public NLP datasets, highlighting the potential of fine-tuning with human feedback for enhancing language model alignment with human intent. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Word2Vec: Efficient Estimation of Word Representations in Vector Space | Pending | Text | Embeddings, Tips & Tricks | 2013 | arXiv | Greg Corrado, Jeffrey Dean, Kai Chen, Tomas Mikolov | link | |
1 | Scaling Instruction-Finetuned Language Models (FLAN) | Pending | Generative, Large-Language-Models, Question-Answering, Text , Transformers | Instruction-Finetuning | 2022 | arXiv | Hyung Won Chung, Jason Wei, Jeffrey Dean, Le Hou, Quoc V. Le, Shayne Longpre | https://arxiv.org/abs/2210.11416 introduces FLAN (Fine-tuned LAnguage Net), an instruction finetuning method, and presents the results of its application. The study demonstrates that by fine-tuning the 540B PaLM model on 1836 tasks while incorporating Chain-of-Thought Reasoning data, FLAN achieves improvements in generalization, human usability, and zero-shot reasoning over the base model. The paper also provides detailed information on how each these aspects was evaluated. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | GPT-2 (Language Models are Unsupervised Multitask Learners) | Pending | Attention, Text , Transformers | 2019 | Alec Radford, Dario Amodei, Ilya Sutskever, Jeffrey Wu | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ReAct: Synergizing Reasoning and Acting in Language Models | Pending | Generative, Large-Language-Models, Text | Optimizations, Tips & Tricks | 2023 | ICLR | Dian Yu, Izhak Shafran, Jeffrey Zhao, Karthik Narasimhan, Nan Du, Shunyu Yao, Yuan Cao | This paper introduces ReAct, a novel approach that leverages Large Language Models (LLMs) to interleave reasoning traces and task-specific actions. ReAct outperforms existing methods on various language and decision-making tasks, addressing issues like hallucination, error propagation, and improving human interpretability and trustworthiness. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Graph Neural Network: Relational inductive biases, deep learning, and graph networks | Pending | GraphNN | Architecture | 2018 | arXiv | Jessica B. Hamrick, Oriol Vinyals, Peter W. Battaglia | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | AnimeGAN: Towards the Automatic Anime Characters Creation with Generative Adversarial Networks | Pending | GANs, Image | 2017 | NIPS | Jiakai Zhang, Minjun Li, Yanghua Jin | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | VL-T5: Unifying Vision-and-Language Tasks via Text Generation | Read | CNNs, CV , Generative, Image , Large-Language-Models, Question-Answering, Text , Transformers | Architecture, Embeddings, Multimodal, Pre-Training | 2021 | arXiv | Hao Tan, Jaemin Cho, Jie Le, Mohit Bansal | Unifying two modalities (image and text) together in a single transformer model to solve multiple tasks in a single architecture using text prefixes similar to T5. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Adam: A Method for Stochastic Optimization | Pending | NNs, Optimizers | 2015 | ICLR | Diederik P. Kingma, Jimmy Ba | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | All you need is a good init | Pending | NN Initialization | Tips & Tricks | 2015 | arXiv | Dmytro Mishkin, Jiri Matas | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | IMLE-GAN: Inclusive GAN: Improving Data and Minority Coverage in Generative Models | Pending | GANs | 2020 | arXiv | Jitendra Malik, Ke Li, Larry Davis, Mario Fritz, Ning Yu, Peng Zhou | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks | Read | NN Initialization, NNs | Optimization-No. of params, Tips & Tricks | 2019 | ICLR | Jonathan Frankle, Michael Carbin | Lottery ticket hypothesis: dense, randomly-initialized, feed-forward networks contain subnetworks (winning tickets) that—when trained in isolation— reach test accuracy comparable to the original network in a similar number of iterations. | link |
1 | Training BatchNorm and Only BatchNorm: On the Expressive Power of Random Features in CNNs | Pending | CNNs, Image | 2020 | arXiv | Ari S. Morcos, David J. Schwab, Jonathan Frankle | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | CLIP: Connecting Text and Images | Pending | Image , Text , Transformers | Multimodal, Pre-Training | 2021 | arXiv | Alec Radford, Ilya Sutskever, Jong Wook Kim | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training Compute-Optimal Large Language Models | Pending | Large-Language-Models, Transformers | Architecture, Optimization-No. of params, Pre-Training, Tips & Tricks | 2022 | arXiv | Jordan Hoffmann, Laurent Sifre, Oriol Vinyals, Sebastian Borgeaud | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | MuZero: Mastering Go, chess, shogi and Atari without rules | Pending | Reinforcement-Learning | 2020 | Nature | David Silver, Demis Hassabis, Ioannis Antonoglou, Julian Schrittwiese | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | A Simple yet Effective Baseline for 3D Human Pose Estimation | Pending | CV , Pose Estimation | 2017 | ICCV | James J. Little, Javier Romero, Julieta Martinez, Rayat Hossain | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | One-shot Text Field Labeling using Attention and Belief Propagation for Structure Information Extraction | Pending | Image , Text | 2020 | arXiv | Jun Huang, Mengli Cheng, Minghui Qiu, Wei Lin, Xing Shi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Pix2Pix: Image-to-Image Translation with Conditional Adversarial Nets | Read | GANs, Image | 2017 | CVPR | Alexei A. Efros, Jun-Yan Zhu, Phillip Isola, Tinghui Zhou | Image to image translation using Conditional GANs and dataset of image pairs from one domain to another. | link | |
1 | CycleGAN: Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks | Pending | GANs, Image | Architecture | 2017 | ICCV | Alexei A. Efros, Jun-Yan Zhu, Phillip Isola, Taesung Park | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation | Pending | GANs, Image | 2018 | CVPR | Jaegul Choo, Jung-Woo Ha, Minje Choi, Munyoung Kim, Sunghun Kim, Yunjey Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Perceptual Losses for Real-Time Style Transfer and Super-Resolution | Pending | Loss Function, NNs | 2016 | ECCV | Alexandre Alahi, Justin Johnson, Li Fei-Fei | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Word2Vec: Efficient Estimation of Word Representations in Vector Space | Pending | Text | Embeddings, Tips & Tricks | 2013 | arXiv | Greg Corrado, Jeffrey Dean, Kai Chen, Tomas Mikolov | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ResNet (Deep Residual Learning for Image Recognition) | Read | CNNs, CV , Image | Architecture | 2016 | CVPR | Kaiming He, Xiangyu Zhang | Introduces Residual or Skip Connections to allow increase in the depth of a DNN | link |
1 | Group Normalization | Pending | NNs, Normalization | Optimizations | 2018 | arXiv | Kaiming He, Yuxin Wu | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ReAct: Synergizing Reasoning and Acting in Language Models | Pending | Generative, Large-Language-Models, Text | Optimizations, Tips & Tricks | 2023 | ICLR | Dian Yu, Izhak Shafran, Jeffrey Zhao, Karthik Narasimhan, Nan Du, Shunyu Yao, Yuan Cao | This paper introduces ReAct, a novel approach that leverages Large Language Models (LLMs) to interleave reasoning traces and task-specific actions. ReAct outperforms existing methods on various language and decision-making tasks, addressing issues like hallucination, error propagation, and improving human interpretability and trustworthiness. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Transforming Sequence Tagging Into A Seq2Seq Task | Pending | Generative, Text | Comparison, Tips & Tricks | 2022 | arXiv | Iftekhar Naim, Karthik Raman, Krishna Srinivasan | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | IMLE-GAN: Inclusive GAN: Improving Data and Minority Coverage in Generative Models | Pending | GANs | 2020 | arXiv | Jitendra Malik, Ke Li, Larry Davis, Mario Fritz, Ning Yu, Peng Zhou | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding | Read | Attention, Text , Transformers | Embeddings | 2018 | NAACL | Jacob Devlin, Kenton Lee, Kristina Toutanova, Ming-Wei Chang | BERT is an extension to Transformer based architecture which introduces a masked word pretraining and next sentence prediction task to pretrain the model for a wide variety of tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Transforming Sequence Tagging Into A Seq2Seq Task | Pending | Generative, Text | Comparison, Tips & Tricks | 2022 | arXiv | Iftekhar Naim, Karthik Raman, Krishna Srinivasan | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Occupancy Anticipation for Efficient Exploration and Navigation | Pending | CNNs, Image | Reinforcement-Learning | 2020 | ECCV | Kristen Grauman, Santhosh K. Ramakrishnan, Ziad Al-Halah | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding | Read | Attention, Text , Transformers | Embeddings | 2018 | NAACL | Jacob Devlin, Kenton Lee, Kristina Toutanova, Ming-Wei Chang | BERT is an extension to Transformer based architecture which introduces a masked word pretraining and next sentence prediction task to pretrain the model for a wide variety of tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | IMLE-GAN: Inclusive GAN: Improving Data and Minority Coverage in Generative Models | Pending | GANs | 2020 | arXiv | Jitendra Malik, Ke Li, Larry Davis, Mario Fritz, Ning Yu, Peng Zhou | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training Compute-Optimal Large Language Models | Pending | Large-Language-Models, Transformers | Architecture, Optimization-No. of params, Pre-Training, Tips & Tricks | 2022 | arXiv | Jordan Hoffmann, Laurent Sifre, Oriol Vinyals, Sebastian Borgeaud | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Flan-T5: Scaling Instruction-Finetuned Language Models | Pending | Generative, Text , Transformers | Architecture, Pre-Training | 2022 | arXiv | Hyung Won Chung, Le Hou | link | |
1 | Scaling Instruction-Finetuned Language Models (FLAN) | Pending | Generative, Large-Language-Models, Question-Answering, Text , Transformers | Instruction-Finetuning | 2022 | arXiv | Hyung Won Chung, Jason Wei, Jeffrey Dean, Le Hou, Quoc V. Le, Shayne Longpre | https://arxiv.org/abs/2210.11416 introduces FLAN (Fine-tuned LAnguage Net), an instruction finetuning method, and presents the results of its application. The study demonstrates that by fine-tuning the 540B PaLM model on 1836 tasks while incorporating Chain-of-Thought Reasoning data, FLAN achieves improvements in generalization, human usability, and zero-shot reasoning over the base model. The paper also provides detailed information on how each these aspects was evaluated. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Decoding a Neural Retriever’s Latent Space for Query Suggestion | Pending | Text | Embeddings, Latent space | 2022 | arXiv | Christian Buck, Leonard Adolphs, Michelle Chen Huebscher | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Perceptual Losses for Real-Time Style Transfer and Super-Resolution | Pending | Loss Function, NNs | 2016 | ECCV | Alexandre Alahi, Justin Johnson, Li Fei-Fei | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Learning to Extract Attribute Value from Product via Question Answering: A Multi-task Approach | Read | Question-Answering, Text , Transformers | Zero-shot-learning | 2020 | KDD | Li Yang, Qifan Wang | Question Answering BERT model used to extract attributes from products. Introduce further No Answer loss and distillation to promote zero shot learning. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training language models to follow instructions with human feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Semi-Supervised | 2022 | arXiv | Carroll L. Wainwright, Diogo Almeida, Jan Leike, Jeff Wu, Long Ouyang, Pamela Mishkin, Paul Christiano, Ryan Lowe, Xu Jiang | This paper presents InstructGPT, a model fine-tuned with human feedback to better align with user intent across various tasks. Despite having significantly fewer parameters than larger models, InstructGPT outperforms them in human evaluations, demonstrating improved truthfulness, reduced toxicity, and minimal performance regressions on public NLP datasets, highlighting the potential of fine-tuning with human feedback for enhancing language model alignment with human intent. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Vision Transformer: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale | Pending | Attention, Image , Transformers | 2021 | ICLR | Alexey Dosovitskiy, Jakob Uszkoreit, Lucas Beyer, Neil Houlsby | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Phrase-Based & Neural Unsupervised Machine Translation | Pending | NMT, Text , Transformers | Unsupervised | 2018 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link | |
1 | Unsupervised Machine Translation Using Monolingual Corpora Only | Pending | GANs, NMT, Text , Transformers | Unsupervised | 2017 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Reformer: The Efficient Transformer | Read | Attention, Text , Transformers | Architecture, Optimization-Memory, Optimization-No. of params | 2020 | arXiv | Anselm Levskaya, Lukasz Kaiser, Nikita Kitaev | Overcome time and memory complexity of Transformers by bucketing Query, Keys and using Reversible residual connections. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | BEGAN: Boundary Equilibrium Generative Adversarial Networks | Pending | GANs, Image | 2017 | arXiv | David Berthelot, Luke Metz, Thomas Schumm | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Revisiting Pose-Normalization for Fine-Grained Few-Shot Recognition | Pending | CNNs, Image | Few-shot-learning | 2020 | CVPR | Bharath Hariharan, Davis Wertheimer, Luming Tang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | WGAN: Wasserstein GAN | Pending | GANs, Loss Function | 2017 | arXiv | Léon Bottou, Martin Arjovsky, Soumith Chintala | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ATOMIC: An Atlas of Machine Commonsense for If-Then Reasoning | Pending | AGI, Dataset, Text | 2019 | AAAI | Maarten Sap, Noah A. Smith, Ronan Le Bras, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | SpanBERT: Improving Pre-training by Representing and Predicting Spans | Read | Question-Answering, Text , Transformers | Pre-Training | 2020 | TACL | Danqi Chen, Mandar Joshi | A different pre-training strategy for BERT model to improve performance for Question Answering task. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Phrase-Based & Neural Unsupervised Machine Translation | Pending | NMT, Text , Transformers | Unsupervised | 2018 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link | |
1 | Unsupervised Machine Translation Using Monolingual Corpora Only | Pending | GANs, NMT, Text , Transformers | Unsupervised | 2017 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | IMLE-GAN: Inclusive GAN: Improving Data and Minority Coverage in Generative Models | Pending | GANs | 2020 | arXiv | Jitendra Malik, Ke Li, Larry Davis, Mario Fritz, Ning Yu, Peng Zhou | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | WGAN: Wasserstein GAN | Pending | GANs, Loss Function | 2017 | arXiv | Léon Bottou, Martin Arjovsky, Soumith Chintala | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Spectral Normalization for GANs | Pending | GANs, Normalization | Optimizations | 2018 | arXiv | Masanori Koyama, Takeru Miyato, Toshiki Kataoka, Yuichi Yoshida | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Topological Loss: Beyond the Pixel-Wise Loss for Topology-Aware Delineation | Pending | Image , Loss Function, Segmentation | 2018 | CVPR | Agata Mosinska, Mateusz Koziński, Pablo Márquez-Neila, Pascal Fua | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ZF Net (Visualizing and Understanding Convolutional Networks) | Read | CNNs, CV , Image | Visualization | 2014 | ECCV | Matthew D. Zeiler, Rob Fergus | Visualize CNN Filters / Kernels using De-Convolutions on CNN filter activations. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Approximating CNNs with Bag-of-local-Features models works surprisingly well on ImageNet | Reading | CNNs, CV , Image | 2019 | arXiv | Matthias Bethge, Wieland Brendel | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | One-shot Text Field Labeling using Attention and Belief Propagation for Structure Information Extraction | Pending | Image , Text | 2020 | arXiv | Jun Huang, Mengli Cheng, Minghui Qiu, Wei Lin, Xing Shi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Class-Balanced Loss Based on Effective Number of Samples | Pending | Loss Function | Tips & Tricks | 2019 | CVPR | Menglin Jia, Yin Cui | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | MobileNet (Efficient Convolutional Neural Networks for Mobile Vision Applications) | Pending | CNNs, CV , Image | Architecture, Optimization-No. of params | 2017 | arXiv | Andrew G. Howard, Menglong Zhu | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks | Read | NN Initialization, NNs | Optimization-No. of params, Tips & Tricks | 2019 | ICLR | Jonathan Frankle, Michael Carbin | Lottery ticket hypothesis: dense, randomly-initialized, feed-forward networks contain subnetworks (winning tickets) that—when trained in isolation— reach test accuracy comparable to the original network in a similar number of iterations. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Decoding a Neural Retriever’s Latent Space for Query Suggestion | Pending | Text | Embeddings, Latent space | 2022 | arXiv | Christian Buck, Leonard Adolphs, Michelle Chen Huebscher | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Self-Alignment with Instruction Backtranslation | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning | 2023 | arXiv | Jason Weston, Mike Lewis, Ping Yu, Xian Li | The paper introduces a scalable method called "instruction backtranslation" to create a high-quality instruction-following language model. This method involves self-augmentation and self-curation of training examples generated from web documents, resulting in a model that outperforms others in its category without relying on distillation data, showcasing its effective self-alignment capability. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | DALL·E: Creating Images from Text | Pending | Image , Text , Transformers | 2021 | Blog | Aditya Ramesh, Gabriel Goh, Ilya Sutskever, Mikhail Pavlov, Scott Gray | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding | Read | Attention, Text , Transformers | Embeddings | 2018 | NAACL | Jacob Devlin, Kenton Lee, Kristina Toutanova, Ming-Wei Chang | BERT is an extension to Transformer based architecture which introduces a masked word pretraining and next sentence prediction task to pretrain the model for a wide variety of tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | One-shot Text Field Labeling using Attention and Belief Propagation for Structure Information Extraction | Pending | Image , Text | 2020 | arXiv | Jun Huang, Mengli Cheng, Minghui Qiu, Wei Lin, Xing Shi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation | Pending | GANs, Image | 2018 | CVPR | Jaegul Choo, Jung-Woo Ha, Minje Choi, Munyoung Kim, Sunghun Kim, Yunjey Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | AnimeGAN: Towards the Automatic Anime Characters Creation with Generative Adversarial Networks | Pending | GANs, Image | 2017 | NIPS | Jiakai Zhang, Minjun Li, Yanghua Jin | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Vokenization: Improving Language Understanding with Contextualized, Visual-Grounded Supervision | This week | Image , Text , Transformers | Multimodal | 2020 | EMNLP | Hao Tan, Mohit Bansal | link | |
1 | VL-T5: Unifying Vision-and-Language Tasks via Text Generation | Read | CNNs, CV , Generative, Image , Large-Language-Models, Question-Answering, Text , Transformers | Architecture, Embeddings, Multimodal, Pre-Training | 2021 | arXiv | Hao Tan, Jaemin Cho, Jie Le, Mohit Bansal | Unifying two modalities (image and text) together in a single transformer model to solve multiple tasks in a single architecture using text prefixes similar to T5. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation | Pending | GANs, Image | 2018 | CVPR | Jaegul Choo, Jung-Woo Ha, Minje Choi, Munyoung Kim, Sunghun Kim, Yunjey Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Phrase-Based & Neural Unsupervised Machine Translation | Pending | NMT, Text , Transformers | Unsupervised | 2018 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link | |
1 | Unsupervised Machine Translation Using Monolingual Corpora Only | Pending | GANs, NMT, Text , Transformers | Unsupervised | 2017 | arXiv | Alexis Conneau, Guillaume Lample, Ludovic Denoyer, Marc'Aurelio Ranzato, Myle Ott | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ReAct: Synergizing Reasoning and Acting in Language Models | Pending | Generative, Large-Language-Models, Text | Optimizations, Tips & Tricks | 2023 | ICLR | Dian Yu, Izhak Shafran, Jeffrey Zhao, Karthik Narasimhan, Nan Du, Shunyu Yao, Yuan Cao | This paper introduces ReAct, a novel approach that leverages Large Language Models (LLMs) to interleave reasoning traces and task-specific actions. ReAct outperforms existing methods on various language and decision-making tasks, addressing issues like hallucination, error propagation, and improving human interpretability and trustworthiness. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Vision Transformer: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale | Pending | Attention, Image , Transformers | 2021 | ICLR | Alexey Dosovitskiy, Jakob Uszkoreit, Lucas Beyer, Neil Houlsby | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Capsule Networks: Dynamic Routing Between Capsules | Pending | CV , Image | Architecture | 2017 | arXiv | Geoffrey E Hinton, Nicholas Frosst, Sara Sabour | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Reformer: The Efficient Transformer | Read | Attention, Text , Transformers | Architecture, Optimization-Memory, Optimization-No. of params | 2020 | arXiv | Anselm Levskaya, Lukasz Kaiser, Nikita Kitaev | Overcome time and memory complexity of Transformers by bucketing Query, Keys and using Reversible residual connections. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | IMLE-GAN: Inclusive GAN: Improving Data and Minority Coverage in Generative Models | Pending | GANs | 2020 | arXiv | Jitendra Malik, Ke Li, Larry Davis, Mario Fritz, Ning Yu, Peng Zhou | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ATOMIC: An Atlas of Machine Commonsense for If-Then Reasoning | Pending | AGI, Dataset, Text | 2019 | AAAI | Maarten Sap, Noah A. Smith, Ronan Le Bras, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Attention is All you Need | Read | Attention, Text , Transformers | Architecture | 2017 | NIPS | Ashish Vaswani, Illia Polosukhin, Noam Shazeer, Łukasz Kaiser | Talks about Transformer architecture which brings SOTA performance for different tasks in NLP | link |
1 | T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer | Read | Attention, Text , Transformers | 2020 | JMLR | Colin Raffel, Noam Shazeer, Peter J. Liu, Wei Liu, Yanqi Zhou | Presents a Text-to-Text transformer model with multi-task learning capabilities, simultaneously solving problems such as machine translation, document summarization, question answering, and classification tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Graph Neural Network: Relational inductive biases, deep learning, and graph networks | Pending | GraphNN | Architecture | 2018 | arXiv | Jessica B. Hamrick, Oriol Vinyals, Peter W. Battaglia | link | |
1 | Training Compute-Optimal Large Language Models | Pending | Large-Language-Models, Transformers | Architecture, Optimization-No. of params, Pre-Training, Tips & Tricks | 2022 | arXiv | Jordan Hoffmann, Laurent Sifre, Oriol Vinyals, Sebastian Borgeaud | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Table-GPT: Table-tuned GPT for Diverse Table Tasks | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning | 2023 | arXiv | Other | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Topological Loss: Beyond the Pixel-Wise Loss for Topology-Aware Delineation | Pending | Image , Loss Function, Segmentation | 2018 | CVPR | Agata Mosinska, Mateusz Koziński, Pablo Márquez-Neila, Pascal Fua | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training language models to follow instructions with human feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Semi-Supervised | 2022 | arXiv | Carroll L. Wainwright, Diogo Almeida, Jan Leike, Jeff Wu, Long Ouyang, Pamela Mishkin, Paul Christiano, Ryan Lowe, Xu Jiang | This paper presents InstructGPT, a model fine-tuned with human feedback to better align with user intent across various tasks. Despite having significantly fewer parameters than larger models, InstructGPT outperforms them in human evaluations, demonstrating improved truthfulness, reduced toxicity, and minimal performance regressions on public NLP datasets, highlighting the potential of fine-tuning with human feedback for enhancing language model alignment with human intent. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Topological Loss: Beyond the Pixel-Wise Loss for Topology-Aware Delineation | Pending | Image , Loss Function, Segmentation | 2018 | CVPR | Agata Mosinska, Mateusz Koziński, Pablo Márquez-Neila, Pascal Fua | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training language models to follow instructions with human feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Semi-Supervised | 2022 | arXiv | Carroll L. Wainwright, Diogo Almeida, Jan Leike, Jeff Wu, Long Ouyang, Pamela Mishkin, Paul Christiano, Ryan Lowe, Xu Jiang | This paper presents InstructGPT, a model fine-tuned with human feedback to better align with user intent across various tasks. Despite having significantly fewer parameters than larger models, InstructGPT outperforms them in human evaluations, demonstrating improved truthfulness, reduced toxicity, and minimal performance regressions on public NLP datasets, highlighting the potential of fine-tuning with human feedback for enhancing language model alignment with human intent. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | IMLE-GAN: Inclusive GAN: Improving Data and Minority Coverage in Generative Models | Pending | GANs | 2020 | arXiv | Jitendra Malik, Ke Li, Larry Davis, Mario Fritz, Ning Yu, Peng Zhou | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer | Read | Attention, Text , Transformers | 2020 | JMLR | Colin Raffel, Noam Shazeer, Peter J. Liu, Wei Liu, Yanqi Zhou | Presents a Text-to-Text transformer model with multi-task learning capabilities, simultaneously solving problems such as machine translation, document summarization, question answering, and classification tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Graph Neural Network: Relational inductive biases, deep learning, and graph networks | Pending | GraphNN | Architecture | 2018 | arXiv | Jessica B. Hamrick, Oriol Vinyals, Peter W. Battaglia | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Symbolic Knowledge Distillation: from General Language Models to Commonsense Models | Pending | Dataset, Text , Transformers | Optimizations, Tips & Tricks | 2021 | arXiv | Chandra Bhagavatula, Jack Hessel, Peter West, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space? | Pending | GANs, Image | 2019 | ICCV | Peter Wonka, Rameen Abdal, Yipeng Qin | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Pix2Pix: Image-to-Image Translation with Conditional Adversarial Nets | Read | GANs, Image | 2017 | CVPR | Alexei A. Efros, Jun-Yan Zhu, Phillip Isola, Tinghui Zhou | Image to image translation using Conditional GANs and dataset of image pairs from one domain to another. | link | |
1 | CycleGAN: Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks | Pending | GANs, Image | Architecture | 2017 | ICCV | Alexei A. Efros, Jun-Yan Zhu, Phillip Isola, Taesung Park | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Self-Alignment with Instruction Backtranslation | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning | 2023 | arXiv | Jason Weston, Mike Lewis, Ping Yu, Xian Li | The paper introduces a scalable method called "instruction backtranslation" to create a high-quality instruction-following language model. This method involves self-augmentation and self-curation of training examples generated from web documents, resulting in a model that outperforms others in its category without relying on distillation data, showcasing its effective self-alignment capability. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Deep Double Descent: Where Bigger Models and More Data Hurt | Pending | NNs | 2019 | arXiv | Boaz Barak, Gal Kaplun, Ilya Sutskever, Preetum Nakkiran, Tristan Yang, Yamini Bansal | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Learning to Extract Attribute Value from Product via Question Answering: A Multi-task Approach | Read | Question-Answering, Text , Transformers | Zero-shot-learning | 2020 | KDD | Li Yang, Qifan Wang | Question Answering BERT model used to extract attributes from products. Introduce further No Answer loss and distillation to promote zero shot learning. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Scaling Instruction-Finetuned Language Models (FLAN) | Pending | Generative, Large-Language-Models, Question-Answering, Text , Transformers | Instruction-Finetuning | 2022 | arXiv | Hyung Won Chung, Jason Wei, Jeffrey Dean, Le Hou, Quoc V. Le, Shayne Longpre | https://arxiv.org/abs/2210.11416 introduces FLAN (Fine-tuned LAnguage Net), an instruction finetuning method, and presents the results of its application. The study demonstrates that by fine-tuning the 540B PaLM model on 1836 tasks while incorporating Chain-of-Thought Reasoning data, FLAN achieves improvements in generalization, human usability, and zero-shot reasoning over the base model. The paper also provides detailed information on how each these aspects was evaluated. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space? | Pending | GANs, Image | 2019 | ICCV | Peter Wonka, Rameen Abdal, Yipeng Qin | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | A Simple yet Effective Baseline for 3D Human Pose Estimation | Pending | CV , Pose Estimation | 2017 | ICCV | James J. Little, Javier Romero, Julieta Martinez, Rayat Hossain | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ZF Net (Visualizing and Understanding Convolutional Networks) | Read | CNNs, CV , Image | Visualization | 2014 | ECCV | Matthew D. Zeiler, Rob Fergus | Visualize CNN Filters / Kernels using De-Convolutions on CNN filter activations. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ATOMIC: An Atlas of Machine Commonsense for If-Then Reasoning | Pending | AGI, Dataset, Text | 2019 | AAAI | Maarten Sap, Noah A. Smith, Ronan Le Bras, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask | Read | NN Initialization, NNs | Comparison, Optimization-No. of params, Tips & Tricks | 2019 | NeurIPS | Hattie Zhou, Janice Lan, Jason Yosinski, Rosanne Liu | Follow up on Lottery Ticket Hypothesis exploring the effects of different Masking criteria as well as Mask-1 and Mask-0 actions. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training language models to follow instructions with human feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Semi-Supervised | 2022 | arXiv | Carroll L. Wainwright, Diogo Almeida, Jan Leike, Jeff Wu, Long Ouyang, Pamela Mishkin, Paul Christiano, Ryan Lowe, Xu Jiang | This paper presents InstructGPT, a model fine-tuned with human feedback to better align with user intent across various tasks. Despite having significantly fewer parameters than larger models, InstructGPT outperforms them in human evaluations, demonstrating improved truthfulness, reduced toxicity, and minimal performance regressions on public NLP datasets, highlighting the potential of fine-tuning with human feedback for enhancing language model alignment with human intent. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | StyleGAN: A Style-Based Generator Architecture for Generative Adversarial Networks | Pending | GANs, Image | 2019 | CVPR | Samuli Laine, Tero Karras, Timo Aila | link | ||
1 | Progressive Growing of GANs for Improved Quality, Stability, and Variation | Pending | GANs, Image | Tips & Tricks | 2018 | ICLR | Jaakko Lehtinen, Samuli Laine, Tero Karras, Timo Aila | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Occupancy Anticipation for Efficient Exploration and Navigation | Pending | CNNs, Image | Reinforcement-Learning | 2020 | ECCV | Kristen Grauman, Santhosh K. Ramakrishnan, Ziad Al-Halah | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Capsule Networks: Dynamic Routing Between Capsules | Pending | CV , Image | Architecture | 2017 | arXiv | Geoffrey E Hinton, Nicholas Frosst, Sara Sabour | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | DALL·E: Creating Images from Text | Pending | Image , Text , Transformers | 2021 | Blog | Aditya Ramesh, Gabriel Goh, Ilya Sutskever, Mikhail Pavlov, Scott Gray | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training Compute-Optimal Large Language Models | Pending | Large-Language-Models, Transformers | Architecture, Optimization-No. of params, Pre-Training, Tips & Tricks | 2022 | arXiv | Jordan Hoffmann, Laurent Sifre, Oriol Vinyals, Sebastian Borgeaud | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Arbitrary Style Transfer in Real-Time With Adaptive Instance Normalization | Pending | CNNs, Image | 2017 | ICCV | Serge Belongie, Xun Huang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Scaling Instruction-Finetuned Language Models (FLAN) | Pending | Generative, Large-Language-Models, Question-Answering, Text , Transformers | Instruction-Finetuning | 2022 | arXiv | Hyung Won Chung, Jason Wei, Jeffrey Dean, Le Hou, Quoc V. Le, Shayne Longpre | https://arxiv.org/abs/2210.11416 introduces FLAN (Fine-tuned LAnguage Net), an instruction finetuning method, and presents the results of its application. The study demonstrates that by fine-tuning the 540B PaLM model on 1836 tasks while incorporating Chain-of-Thought Reasoning data, FLAN achieves improvements in generalization, human usability, and zero-shot reasoning over the base model. The paper also provides detailed information on how each these aspects was evaluated. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | How Does Batch Normalization Help Optimization? | Pending | NNs, Normalization | Optimizations | 2018 | arXiv | Aleksander Madry, Andrew Ilyas, Dimitris Tsipras, Shibani Santurkar | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | TransGAN: Two Transformers Can Make One Strong GAN | Pending | GANs, Image , Transformers | Architecture | 2021 | arXiv | Shiyu Chang, Yifan Jiang, Zhangyang Wang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ReAct: Synergizing Reasoning and Acting in Language Models | Pending | Generative, Large-Language-Models, Text | Optimizations, Tips & Tricks | 2023 | ICLR | Dian Yu, Izhak Shafran, Jeffrey Zhao, Karthik Narasimhan, Nan Du, Shunyu Yao, Yuan Cao | This paper introduces ReAct, a novel approach that leverages Large Language Models (LLMs) to interleave reasoning traces and task-specific actions. ReAct outperforms existing methods on various language and decision-making tasks, addressing issues like hallucination, error propagation, and improving human interpretability and trustworthiness. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | SqueezeNet | Read | CNNs, CV , Image | Architecture, Optimization-No. of params | 2016 | arXiv | Forrest N. Iandola, Song Han | Explores model compression by using 1x1 convolutions called fire modules. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | WGAN: Wasserstein GAN | Pending | GANs, Loss Function | 2017 | arXiv | Léon Bottou, Martin Arjovsky, Soumith Chintala | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Understanding Loss Functions in Computer Vision | Pending | CV , GANs, Image , Loss Function | Comparison, Tips & Tricks | 2020 | Blog | Sowmya Yellapragada | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | GPT-f: Generative Language Modeling for Automated Theorem Proving | Pending | Attention, Transformers | 2020 | arXiv | Ilya Sutskever, Stanislas Polu | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Single Headed Attention RNN: Stop Thinking With Your Head | Pending | Attention, LSTMs, Text | Optimization-No. of params | 2019 | arXiv | Stephen Merity | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | A 2019 guide to Human Pose Estimation with Deep Learning | Pending | CV , Pose Estimation | Comparison | 2019 | Blog | Sudharshan Chandra Babu | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation | Pending | GANs, Image | 2018 | CVPR | Jaegul Choo, Jung-Woo Ha, Minje Choi, Munyoung Kim, Sunghun Kim, Yunjey Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | CycleGAN: Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks | Pending | GANs, Image | Architecture | 2017 | ICCV | Alexei A. Efros, Jun-Yan Zhu, Phillip Isola, Taesung Park | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Spectral Normalization for GANs | Pending | GANs, Normalization | Optimizations | 2018 | arXiv | Masanori Koyama, Takeru Miyato, Toshiki Kataoka, Yuichi Yoshida | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Large Language Models are Zero-Shot Reasoners | Pending | Generative, Question-Answering, Text | Tips & Tricks, Zero-shot-learning | 2022 | arXiv | Takeshi Kojima, Yusuke Iwasawa | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering | Pending | Large-Language-Models | Prompting, Tips & Tricks | 2024 | arXiv | Dedy Kredo, Itamar Friedman, Tal Ridnik | This paper introduces AlphaCodium, a novel test-based, multi-stage, code-oriented iterative approach for improving the performance of Language Model Models (LLMs) on code generation tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | StyleGAN: A Style-Based Generator Architecture for Generative Adversarial Networks | Pending | GANs, Image | 2019 | CVPR | Samuli Laine, Tero Karras, Timo Aila | link | ||
1 | Progressive Growing of GANs for Improved Quality, Stability, and Variation | Pending | GANs, Image | Tips & Tricks | 2018 | ICLR | Jaakko Lehtinen, Samuli Laine, Tero Karras, Timo Aila | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | BEGAN: Boundary Equilibrium Generative Adversarial Networks | Pending | GANs, Image | 2017 | arXiv | David Berthelot, Luke Metz, Thomas Schumm | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Self-Normalizing Neural Networks | Pending | Activation Function, Tabular | Optimizations, Tips & Tricks | 2017 | NIPS | Andreas Mayr, Günter Klambauer, Thomas Unterthiner | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Improved Techniques for Training GANs | Pending | GANs, Image | Semi-Supervised | 2016 | NIPS | Alec Radford, Ian Goodfellow, Tim Salimans, Vicki Cheung, Wojciech Zaremba, Xi Chen | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | StyleGAN: A Style-Based Generator Architecture for Generative Adversarial Networks | Pending | GANs, Image | 2019 | CVPR | Samuli Laine, Tero Karras, Timo Aila | link | ||
1 | Progressive Growing of GANs for Improved Quality, Stability, and Variation | Pending | GANs, Image | Tips & Tricks | 2018 | ICLR | Jaakko Lehtinen, Samuli Laine, Tero Karras, Timo Aila | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | NADAM: Incorporating Nesterov Momentum into Adam | Pending | NNs, Optimizers | Comparison | 2016 | Timothy Dozat | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Pix2Pix: Image-to-Image Translation with Conditional Adversarial Nets | Read | GANs, Image | 2017 | CVPR | Alexei A. Efros, Jun-Yan Zhu, Phillip Isola, Tinghui Zhou | Image to image translation using Conditional GANs and dataset of image pairs from one domain to another. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Word2Vec: Efficient Estimation of Word Representations in Vector Space | Pending | Text | Embeddings, Tips & Tricks | 2013 | arXiv | Greg Corrado, Jeffrey Dean, Kai Chen, Tomas Mikolov | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Bag of Tricks for Image Classification with Convolutional Neural Networks | Read | CV , Image | Optimizations, Tips & Tricks | 2018 | arXiv | Tong He, Zhi Zhang | Shows a dozen tricks (mixup, label smoothing, etc.) to improve CNN accuracy and training time. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Spectral Normalization for GANs | Pending | GANs, Normalization | Optimizations | 2018 | arXiv | Masanori Koyama, Takeru Miyato, Toshiki Kataoka, Yuichi Yoshida | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Deep Double Descent: Where Bigger Models and More Data Hurt | Pending | NNs | 2019 | arXiv | Boaz Barak, Gal Kaplun, Ilya Sutskever, Preetum Nakkiran, Tristan Yang, Yamini Bansal | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Improved Techniques for Training GANs | Pending | GANs, Image | Semi-Supervised | 2016 | NIPS | Alec Radford, Ian Goodfellow, Tim Salimans, Vicki Cheung, Wojciech Zaremba, Xi Chen | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | One-shot Text Field Labeling using Attention and Belief Propagation for Structure Information Extraction | Pending | Image , Text | 2020 | arXiv | Jun Huang, Mengli Cheng, Minghui Qiu, Wei Lin, Xing Shi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Inception-v1 (Going Deeper With Convolutions) | Read | CNNs, CV , Image | Architecture | 2015 | CVPR | Christian Szegedy, Wei Liu | Propose the use of 1x1 conv operations to reduce the number of parameters in a deep and wide CNN | link |
1 | T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer | Read | Attention, Text , Transformers | 2020 | JMLR | Colin Raffel, Noam Shazeer, Peter J. Liu, Wei Liu, Yanqi Zhou | Presents a Text-to-Text transformer model with multi-task learning capabilities, simultaneously solving problems such as machine translation, document summarization, question answering, and classification tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Approximating CNNs with Bag-of-local-Features models works surprisingly well on ImageNet | Reading | CNNs, CV , Image | 2019 | arXiv | Matthias Bethge, Wieland Brendel | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Improved Techniques for Training GANs | Pending | GANs, Image | Semi-Supervised | 2016 | NIPS | Alec Radford, Ian Goodfellow, Tim Salimans, Vicki Cheung, Wojciech Zaremba, Xi Chen | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Improved Techniques for Training GANs | Pending | GANs, Image | Semi-Supervised | 2016 | NIPS | Alec Radford, Ian Goodfellow, Tim Salimans, Vicki Cheung, Wojciech Zaremba, Xi Chen | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Self-Alignment with Instruction Backtranslation | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning | 2023 | arXiv | Jason Weston, Mike Lewis, Ping Yu, Xian Li | The paper introduces a scalable method called "instruction backtranslation" to create a high-quality instruction-following language model. This method involves self-augmentation and self-curation of training examples generated from web documents, resulting in a model that outperforms others in its category without relying on distillation data, showcasing its effective self-alignment capability. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ResNet (Deep Residual Learning for Image Recognition) | Read | CNNs, CV , Image | Architecture | 2016 | CVPR | Kaiming He, Xiangyu Zhang | Introduces Residual or Skip Connections to allow increase in the depth of a DNN | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Interpreting Deep Learning Models in Natural Language Processing: A Review | Pending | Text | Comparison, Visualization | 2021 | arXiv | Diyi Yang, Xiaofei Sun | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | One-shot Text Field Labeling using Attention and Belief Propagation for Structure Information Extraction | Pending | Image , Text | 2020 | arXiv | Jun Huang, Mengli Cheng, Minghui Qiu, Wei Lin, Xing Shi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Training language models to follow instructions with human feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Semi-Supervised | 2022 | arXiv | Carroll L. Wainwright, Diogo Almeida, Jan Leike, Jeff Wu, Long Ouyang, Pamela Mishkin, Paul Christiano, Ryan Lowe, Xu Jiang | This paper presents InstructGPT, a model fine-tuned with human feedback to better align with user intent across various tasks. Despite having significantly fewer parameters than larger models, InstructGPT outperforms them in human evaluations, demonstrating improved truthfulness, reduced toxicity, and minimal performance regressions on public NLP datasets, highlighting the potential of fine-tuning with human feedback for enhancing language model alignment with human intent. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Chain of Thought Prompting Elicits Reasoning in Large Language Models | Pending | Question-Answering, Text , Transformers | 2022 | arXiv | Denny Zhou, Jason Wei, Xuezhi Wang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Arbitrary Style Transfer in Real-Time With Adaptive Instance Normalization | Pending | CNNs, Image | 2017 | ICCV | Serge Belongie, Xun Huang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Deep Double Descent: Where Bigger Models and More Data Hurt | Pending | NNs | 2019 | arXiv | Boaz Barak, Gal Kaplun, Ilya Sutskever, Preetum Nakkiran, Tristan Yang, Yamini Bansal | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | AnimeGAN: Towards the Automatic Anime Characters Creation with Generative Adversarial Networks | Pending | GANs, Image | 2017 | NIPS | Jiakai Zhang, Minjun Li, Yanghua Jin | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer | Read | Attention, Text , Transformers | 2020 | JMLR | Colin Raffel, Noam Shazeer, Peter J. Liu, Wei Liu, Yanqi Zhou | Presents a Text-to-Text transformer model with multi-task learning capabilities, simultaneously solving problems such as machine translation, document summarization, question answering, and classification tasks. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ATOMIC: An Atlas of Machine Commonsense for If-Then Reasoning | Pending | AGI, Dataset, Text | 2019 | AAAI | Maarten Sap, Noah A. Smith, Ronan Le Bras, Yejin Choi | link | ||
1 | COMET: Commonsense Transformers for Automatic Knowledge Graph Construction | Pending | AGI, Text , Transformers | 2019 | ACL | Antoine Bosselut, Hannah Rashkin, Yejin Choi | link | ||
2 | VisualCOMET: Reasoning about the Dynamic Context of a Still Image | Pending | AGI, Dataset, Image , Text , Transformers | 2020 | ECCV | Ali Farhadi, Chandra Bhagavatula, Jae Sung Park, Yejin Choi | link | ||
3 | Symbolic Knowledge Distillation: from General Language Models to Commonsense Models | Pending | Dataset, Text , Transformers | Optimizations, Tips & Tricks | 2021 | arXiv | Chandra Bhagavatula, Jack Hessel, Peter West, Yejin Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | A Comprehensive Guide on Activation Functions | This week | Activation Function | 2020 | Blog | Ygor Rebouças Serpa | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | TransGAN: Two Transformers Can Make One Strong GAN | Pending | GANs, Image , Transformers | Architecture | 2021 | arXiv | Shiyu Chang, Yifan Jiang, Zhangyang Wang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Class-Balanced Loss Based on Effective Number of Samples | Pending | Loss Function | Tips & Tricks | 2019 | CVPR | Menglin Jia, Yin Cui | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Language-Agnostic BERT Sentence Embedding | Read | Attention, Siamese Network, Text , Transformers | Embeddings | 2020 | arXiv | Fangxiaoyu Feng, Yinfei Yang | A BERT model with multilingual sentence embeddings learned over 112 languages and Zero-shot learning over unseen languages. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space? | Pending | GANs, Image | 2019 | ICCV | Peter Wonka, Rameen Abdal, Yipeng Qin | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | ReAct: Synergizing Reasoning and Acting in Language Models | Pending | Generative, Large-Language-Models, Text | Optimizations, Tips & Tricks | 2023 | ICLR | Dian Yu, Izhak Shafran, Jeffrey Zhao, Karthik Narasimhan, Nan Du, Shunyu Yao, Yuan Cao | This paper introduces ReAct, a novel approach that leverages Large Language Models (LLMs) to interleave reasoning traces and task-specific actions. ReAct outperforms existing methods on various language and decision-making tasks, addressing issues like hallucination, error propagation, and improving human interpretability and trustworthiness. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Spectral Normalization for GANs | Pending | GANs, Normalization | Optimizations | 2018 | arXiv | Masanori Koyama, Takeru Miyato, Toshiki Kataoka, Yuichi Yoshida | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation | Pending | GANs, Image | 2018 | CVPR | Jaegul Choo, Jung-Woo Ha, Minje Choi, Munyoung Kim, Sunghun Kim, Yunjey Choi | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Constitutional AI: Harmlessness from AI Feedback | Pending | Generative, Large-Language-Models, Training Method | Instruction-Finetuning, Reinforcement-Learning, Unsupervised | 2022 | arXiv | Jared Kaplan, Yuntao Ba | The paper introduces Constitutional AI, a method for training a safe AI assistant without human-labeled data on harmful outputs. It combines supervised learning and reinforcement learning phases, enabling the AI to engage with harmful queries by explaining its objections, thus improving control, transparency, and human-judged performance with minimal human oversight. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Large Language Models are Zero-Shot Reasoners | Pending | Generative, Question-Answering, Text | Tips & Tricks, Zero-shot-learning | 2022 | arXiv | Takeshi Kojima, Yusuke Iwasawa | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Group Normalization | Pending | NNs, Normalization | Optimizations | 2018 | arXiv | Kaiming He, Yuxin Wu | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | TransGAN: Two Transformers Can Make One Strong GAN | Pending | GANs, Image , Transformers | Architecture | 2021 | arXiv | Shiyu Chang, Yifan Jiang, Zhangyang Wang | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Large Language Models for Data Annotation: A Survey | This week | Dataset, Generative, Large-Language-Models | Prompting, Tips & Tricks | 2024 | arXiv | Alimohammad Beigi, Zhen Tan | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Bag of Tricks for Image Classification with Convolutional Neural Networks | Read | CV , Image | Optimizations, Tips & Tricks | 2018 | arXiv | Tong He, Zhi Zhang | Shows a dozen tricks (mixup, label smoothing, etc.) to improve CNN accuracy and training time. | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Occupancy Anticipation for Efficient Exploration and Navigation | Pending | CNNs, Image | Reinforcement-Learning | 2020 | ECCV | Kristen Grauman, Santhosh K. Ramakrishnan, Ziad Al-Halah | link |
Paper Name | Status | Topic | Category | Year | Conference | Author | Summary | Link | |
---|---|---|---|---|---|---|---|---|---|
0 | Attention is All you Need | Read | Attention, Text , Transformers | Architecture | 2017 | NIPS | Ashish Vaswani, Illia Polosukhin, Noam Shazeer, Łukasz Kaiser | Talks about Transformer architecture which brings SOTA performance for different tasks in NLP | link |