Search results
- ...11, 2020 || 175 Billion || 2,048 || 1,433 || 3 pages || 570 GB || [[Common Crawl]], [[WebText2]], [[Books1]], [[Books2]], [[Wikipedia]] || No || [[Language ...5, 2022 || 175 Billion || 4,000 || 2,800 || 6 pages || 570 GB || [[Common Crawl]], [[WebText2]], [[Books1]], [[Books2]], [[Wikipedia]] || No ||4 KB (493 words) - 18:00, 15 July 2023
- '''[[Common Crawl]]''' - open repository of web crawl data4 KB (550 words) - 09:53, 14 May 2023
- ...are both trained on unlabeled text datasets, such as Wikipedia and Common Crawl. Specifically, LaMDA is trained on dialogue training sets to generate non-g12 KB (1,749 words) - 14:03, 3 May 2023