Papers: Difference between revisions
No edit summary |
No edit summary |
||
Line 15: | Line 15: | ||
|- | |- | ||
|[[Very Deep Convolutional Networks for Large-Scale Image Recognition (VGGNet)]] || 2014/09/04 || [[arxiv:409.1556]] || || | |[[Very Deep Convolutional Networks for Large-Scale Image Recognition (VGGNet)]] || 2014/09/04 || [[arxiv:409.1556]] || || | ||
|- | |||
|[[Sequence to Sequence Learning with Neural Networks (Seq2Seq)]] || 2014/09/10 || [[arxiv:1409.3215]] || || | |||
|- | |- | ||
|[[Deep Residual Learning for Image Recognition (ResNet)]] || 2015/12/10 || [[arxiv:409.1556]] || || | |[[Deep Residual Learning for Image Recognition (ResNet)]] || 2015/12/10 || [[arxiv:409.1556]] || || | ||
Line 21: | Line 23: | ||
|- | |- | ||
|[[Asynchronous Methods for Deep Reinforcement Learning (A3C)]] || 2016/02/04 || [[arxiv:1602.01783]] || || | |[[Asynchronous Methods for Deep Reinforcement Learning (A3C)]] || 2016/02/04 || [[arxiv:1602.01783]] || || | ||
|- | |||
|[[WaveNet: A Generative Model for Raw Audio]] || 2016/09/12 || [[arxiv:1609.03499]] || [[Audio]] || | |||
|- | |- | ||
|[[Attention Is All You Need (Transformer)]] || 2017/06/12 || [[arxiv:1706.03762]] || || influential paper that introduced [[Transformer]] | |[[Attention Is All You Need (Transformer)]] || 2017/06/12 || [[arxiv:1706.03762]] || || influential paper that introduced [[Transformer]] |
Revision as of 02:41, 6 February 2023
Important Papers
Other Papers
https://arxiv.org/abs/2301.13779 (FLAME: A small language model for spreadsheet formulas) - Small model specifically for spreadsheets by Miscrofot