Leo Gao
Leo Gao
Verified email at - Homepage
Cited by
Cited by
Multitask prompted training enables zero-shot task generalization
V Sanh, A Webson, C Raffel, SH Bach, L Sutawika, Z Alyafeai, A Chaffin, ...
arXiv preprint arXiv:2110.08207, 2021
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
L Gao, S Biderman, S Black, L Golding, T Hoppe, C Foster, J Phang, H He, ...
arXiv preprint arXiv:2101.00027, 2020
Bloom: A 176b-parameter open-access multilingual language model
T Le Scao, A Fan, C Akiki, E Pavlick, S Ilić, D Hesslow, R Castagné, ...
Gpt-4 technical report
J Achiam, S Adler, S Agarwal, L Ahmad, I Akkaya, FL Aleman, D Almeida, ...
arXiv preprint arXiv:2303.08774, 2023
Beyond the imitation game: Quantifying and extrapolating the capabilities of language models
A Srivastava, A Rastogi, A Rao, AAM Shoeb, A Abid, A Fisch, AR Brown, ...
arXiv preprint arXiv:2206.04615, 2022
GPT-NeoX-20B: An Open-Source Autoregressive Language Model
S Black, S Biderman, E Hallahan, Q Anthony, L Gao, L Golding, H He, ...
ACL Workshop on Challenges & Perspectives in Creating Large Language Models, 2022
GPT-Neo: Large scale autoregressive language modeling with mesh-tensorflow
S Black, L Gao, P Wang, C Leahy, S Biderman
A framework for few-shot language model evaluation
L Gao, J Tow, S Biderman, S Black, A DiPofi, C Foster, L Golding, J Hsu, ...
Scaling Laws for Reward Model Overoptimization
L Gao, J Schulman, J Hilton
International Conference on Machine Learning, 2022
Language models can explain neurons in language models
S Bills, N Cammarata, D Mossing, H Tillman, L Gao, G Goh, I Sutskever, ...
OpenAI blog, 2023
Weak-to-Strong Generalization: Eliciting Strong Capabilities with Weak Supervision
C Burns, P Izmailov, JH Kirchner, B Baker, L Gao, L Aschenbrenner, ...
arXiv preprint arXiv:2312.09390, 2023
Datasheet for the pile
S Biderman, K Bicheno, L Gao
arXiv preprint arXiv:2201.07311, 2022
Collaborative Storytelling with Large-scale Neural Language Models
E Nichols, L Gao, R Gomez
ACM SIGGRAPH Conference on Motion, Interaction and Games, 2020
On the sizes of openai api models
L Gao
EleutherAI Blog, 2021
Collaborative storytelling with social robots
E Nichols, L Gao, Y Vasylkiv, R Gomez
2021 IEEE/RSJ International Conference on Intelligent Robots and Systems …, 2021
Cut the carp: Fishing for zero-shot story evaluation
S Matiana, JR Smith, R Teehan, L Castricato, S Biderman, L Gao, ...
arXiv preprint arXiv:2110.03111, 2021
Multitask prompted training enables zero-shot task generalization. arXiv
V Sanh, A Webson, C Raffel, SH Bach, L Sutawika, Z Alyafeai, A Chaffin, ...
arXiv preprint arXiv:2110.08207, 2021
EleutherAI: Going Beyond "Open Science" to "Science in the Open"
J Phang, H Bradley, L Gao, L Castricato, S Biderman
arXiv preprint arXiv:2210.06413, 2022
Multitask prompted training enables zero-shot task generalization, 2021
V Sanh, A Webson, C Raffel, SH Bach, L Sutawika, Z Alyafeai, A Chaffin, ...
URL https://arxiv. org/abs/2110.08207, 2022
Rotary embeddings: A relative revolution
S Biderman, S Black, C Foster, L Gao, E Hallahan, H He, B Wang, ..., 2021
The system can't perform the operation now. Try again later.
Articles 1–20