Papers: Difference between revisions
No edit summary |
No edit summary |
||
Line 40: | Line 40: | ||
|[[BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding]] || 2018/10/11 || [[arxiv:1810.04805]] || [[NLP]] || [[BERT]] | |[[BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding]] || 2018/10/11 || [[arxiv:1810.04805]] || [[NLP]] || [[BERT]] | ||
|- | |- | ||
|[[Transformer-XL]] || 2019/01/09 || [[arxiv:1901.02860]] || || | |[[Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context]] || 2019/01/09 || [[arxiv:1901.02860]] || || [[Transformer-XL]] | ||
|- | |- | ||
|[[Language Models are Few-Shot Learners (GPT-3)]] || 2020/05/28 || [[arxiv:2005.14165]] || [[NLP]] || [[GPT-3]] | |[[Language Models are Few-Shot Learners (GPT-3)]] || 2020/05/28 || [[arxiv:2005.14165]] || [[NLP]] || [[GPT-3]] | ||
Line 46: | Line 46: | ||
|[[An Image is Worth 16x16 Words]] || 2020/10/22 || [[arxiv:2010.11929]] || || Transformers for Image Recognition at Scale - [[Vision Transformer]] ([[ViT]]) | |[[An Image is Worth 16x16 Words]] || 2020/10/22 || [[arxiv:2010.11929]] || || Transformers for Image Recognition at Scale - [[Vision Transformer]] ([[ViT]]) | ||
|- | |- | ||
|[[ | |[[Learning Transferable Visual Models From Natural Language Supervision (CLIP)]] || 2021/02/26 || [[arxiv:2103.00020]]<br>[https://openai.com/blog/clip/ OpenAI Blog] || || | ||
|- | |- | ||
|[[MobileViT]] || 2021/10/05 || [[arxiv:2110.02178]] || || | |[[MobileViT: Light-weight, General-purpose, and Mobile-friendly Vision Transformer]] || 2021/10/05 || [[arxiv:2110.02178]] || || [[MobileViT]] | ||
|- | |- | ||
|[[Block-Recurrent Transformers]] || 2022/03/11 || [[arxiv:2203.07852]] || || | |[[Block-Recurrent Transformers]] || 2022/03/11 || [[arxiv:2203.07852]] || || | ||
Line 54: | Line 54: | ||
|[[Memorizing Transformers]] || 2022/03/16 ||[[arxiv:2203.08913]] || || | |[[Memorizing Transformers]] || 2022/03/16 ||[[arxiv:2203.08913]] || || | ||
|- | |- | ||
|[[STaR]] || 2022/03/28 || [[arxiv:2203.14465]] || || | |[[STaR: Bootstrapping Reasoning With Reasoning]] || 2022/03/28 || [[arxiv:2203.14465]] || || [[STaR]] | ||
|- | |- | ||
|} | |} |
Revision as of 02:46, 6 February 2023
Important Papers
Other Papers
https://arxiv.org/abs/2301.13779 (FLAME: A small language model for spreadsheet formulas) - Small model specifically for spreadsheets by Miscrofot