2021
DOI: 10.48550/arxiv.2108.07258
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the Opportunities and Risks of Foundation Models

Abstract: AI is undergoing a paradigm shift with the rise of models (e.g., BERT, DALL-E, GPT-3) that are trained on broad data at scale and are adaptable to a wide range of downstream tasks. We call these models foundation models to underscore their critically central yet incomplete character. This report provides a thorough account of the opportunities and risks of foundation models, ranging from their capabilities (e.g., language, vision, robotics, reasoning, human interaction) and technical principles (e.g., model ar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
652
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 649 publications
(847 citation statements)
references
References 867 publications
(659 reference statements)
5
652
0
Order By: Relevance
“…Like other large machine learning systems pretrained on web data, our system may reproduce harmful social biases present in its training data. While a variety of past work has studied risks of language-only pretraining [125, 14,8,60], the video-centric pretraining that we explore in our work might have different benefits and risks. We discuss these below, along with how we worked to mitigate them through our work.…”
Section: A Broader Impact Statementmentioning
confidence: 99%
“…Like other large machine learning systems pretrained on web data, our system may reproduce harmful social biases present in its training data. While a variety of past work has studied risks of language-only pretraining [125, 14,8,60], the video-centric pretraining that we explore in our work might have different benefits and risks. We discuss these below, along with how we worked to mitigate them through our work.…”
Section: A Broader Impact Statementmentioning
confidence: 99%
“…Due to the superiority of Transformers in the computation complexity and flexibility over prior models, they are popular, especially in the field of NLP. Besides, its capability of computation parallelization allows more intense use of massive data, which led to the development of large-scale pretrained models, namely foundation models [6]. In this paper, we consider two examples of Transformers in particular: BERT [7] for discriminative tasks, and GPT [8] for generative tasks.…”
Section: A Transformer Modelsmentioning
confidence: 99%
“…Based on transfer learning, models for various tasks could origin from the same pretrained language model. As an extended definition of pretrained language model, the model that is "trained on broad data at scale and can be adapted to a wide range of downstream tasks" is named "foundation model" [6]. In NLP, representative examples include BERT and GPT, which are trained on large corpora of text and then adapted to a wide range of downstream tasks, e.g., machine translation, question answering, and sentiment analysis.…”
Section: B Transfer Learning and Adaptive Schemesmentioning
confidence: 99%
See 1 more Smart Citation
“…While general, these tabula rasa architectures rely on large amounts of experience to learn. Data efficiency is especially important in motor control because, unlike computer vision and natural language processing which have greatly benefited from the availability of large datasets (Bommasani et al, 2021), gathering large-scale data for robotics is challenging. Experience must be gathered through interaction with the environment, which is difficult for reasons including time, human labor, safety, maintenance, and reproducibility (Kroemer et al, 2020).…”
Section: Introductionmentioning
confidence: 99%