Improving Language Understanding by Generative Pre-Training (GPT): Revision history

Diff selection: Mark the radio buttons of the revisions to compare and hit enter or the button at the bottom.
Legend: (cur) = difference with latest revision, (prev) = difference with preceding revision, m = minor edit.

2 March 2023

23 February 2023

  • curprev 16:0316:03, 23 February 2023Daikon Radish talk contribs 3,374 bytes +2,343 No edit summary
  • curprev 16:0116:01, 23 February 2023Daikon Radish talk contribs 1,031 bytes +1,031 Created page with "===Introduction=== In June 2018, OpenAI introduced GPT-1, a language model that combined unsupervised pre-training with the transformer architecture to achieve significant progress in natural language understanding. The team fine-tuned the model for specific tasks and found that pre-training helped it perform well on various NLP tasks with minimal fine-tuning. GPT-1 used the BooksCorpus dataset and self-attention in the transformer's decoder with 117 million parameters,..."