Jump to content

Papers: Difference between revisions

95 bytes added ,  5 February 2023
no edit summary
No edit summary
No edit summary
Line 3: Line 3:
'''[[Attention Is All You Need]]''' - https://arxiv.org/abs/1706.03762 - - influential paper that introduced [[Transformer]]
'''[[Attention Is All You Need]]''' - https://arxiv.org/abs/1706.03762 - - influential paper that introduced [[Transformer]]


'''[[An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale]]''' - https://arxiv.org/abs/2010.11929 - [[Vision Transformer]] ([[ViT]])
'''[[An Image is Worth 16x16 Words]]''' - https://arxiv.org/abs/2010.11929 - Transformers for Image Recognition at Scale - [[Vision Transformer]] ([[ViT]])
 
'''[[Language Models are Few-Shot Learners]]''' - https://arxiv.org/abs/2005.14165 - [[GPT]]


'''[[Memorizing Transformers]]''' - https://arxiv.org/abs/2203.08913 -  
'''[[Memorizing Transformers]]''' - https://arxiv.org/abs/2203.08913 -