2021
DOI: 10.5281/zenodo.5297715
|View full text |Cite
|
Sign up to set email alerts
|

GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
142
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 134 publications
(142 citation statements)
references
References 0 publications
0
142
0
Order By: Relevance
“…The recent emergence of grassroots based open-sourcing initiatives can be attributed to an increasing adoption of the closed-source commercial API access mode of dissemination being used for projects such as GPT-3 [34], CLIP and DALL-E 13 . EleutherAI 14 achieved success by replicating both the WebText dataset (on which GPT-3 was trained) and the GPT-3 model itself by unveiling the Pile dataset [42] and the GPT-Neo [43]/GPT-NeoX [44] models. As indicated in the README section of the LAION Github repository 15 , the primal motivation behind the LAION-400M undertaking was to produce open-source variants of the opaque WIT (WebImageText) dataset, and the CLIP [2] and DALL-E [45] models.…”
Section: Motivational Drive: Open-sourcing the Closed-sourcementioning
confidence: 99%
“…The recent emergence of grassroots based open-sourcing initiatives can be attributed to an increasing adoption of the closed-source commercial API access mode of dissemination being used for projects such as GPT-3 [34], CLIP and DALL-E 13 . EleutherAI 14 achieved success by replicating both the WebText dataset (on which GPT-3 was trained) and the GPT-3 model itself by unveiling the Pile dataset [42] and the GPT-Neo [43]/GPT-NeoX [44] models. As indicated in the README section of the LAION Github repository 15 , the primal motivation behind the LAION-400M undertaking was to produce open-source variants of the opaque WIT (WebImageText) dataset, and the CLIP [2] and DALL-E [45] models.…”
Section: Motivational Drive: Open-sourcing the Closed-sourcementioning
confidence: 99%
“…We used GitHub's code review tool 7 to manually classify errors in the code translations. Two authors examined each of the visual diffs produced within GitHub and made comments to label errors and explain the reason for why the code contained that error.…”
Section: Code Quality Measuresmentioning
confidence: 99%
“…For cases in which participants omitted any implementation, we corrected their code by adding the two-method implementation, as it was closer in spirit to the original Java. 7 https://github.com/features/code-review/ 8 https://docs.github.com/en/rest/ 9 Souce lines of code (SLOC) is a metric of the number of source lines of code; it does not include blank or commented lines. We used the cloc utility to compute SLOC for all code artifacts in our study, available at https://github.com/AlDanial/cloc.…”
Section: Code Quality Measuresmentioning
confidence: 99%
See 1 more Smart Citation
“…Additionally, we conduct comparative experiments to verify whether open-source alternatives to GPT-3 could still provide comparable performance gains through data augmentation. As opensource alternatives, GPT-2 (Radford et al) and GPT-neo (Black et al, 2021) were chosen. The latter is a popular alternative to the commercial GPT-3, performing competitively with the smaller versions (ada and babbage) of the counterpart.…”
Section: Language Model Capacitymentioning
confidence: 99%