Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.404
|View full text |Cite
|
Sign up to set email alerts
|

Probing Simile Knowledge from Pre-trained Language Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(31 citation statements)
references
References 0 publications
0
31
0
Order By: Relevance
“…(2) optimizing the step-1 model using the proximal policy optimization (PPO) (Schulman et al, 2017) method, the authors first built a comparison dataset by collecting responses from GPT-4, InstructGPT (Ouyang et al, 2022), and OPT-IML (Iyer et al, 2022) to a collection of instructions and then asked GPT-4 to rate each response from 1 to 10. Using the ratings, a reward model is trained based on OPT (Zhang et al, 2022a). The fine-tuned model from Step 1 is optimized by using the reward model to compute the policy gradient.…”
Section: Gpt-4-llmmentioning
confidence: 99%
See 1 more Smart Citation
“…(2) optimizing the step-1 model using the proximal policy optimization (PPO) (Schulman et al, 2017) method, the authors first built a comparison dataset by collecting responses from GPT-4, InstructGPT (Ouyang et al, 2022), and OPT-IML (Iyer et al, 2022) to a collection of instructions and then asked GPT-4 to rate each response from 1 to 10. Using the ratings, a reward model is trained based on OPT (Zhang et al, 2022a). The fine-tuned model from Step 1 is optimized by using the reward model to compute the policy gradient.…”
Section: Gpt-4-llmmentioning
confidence: 99%
“…OPT-IML (175B) (Iyer et al, 2022) is a large language model trained by fine-tuning the OPT (175B) (Zhang et al, 2022a) model on the constructed Instruction Meta-Learning (IML) dataset, which consists of over 1500 NLP tasks from 8 publicly available benchmarks such as PromptSource (Bach et al, 2022), FLAN (Longpre et al, 2023), and Super-NaturalInstructions (Wang et al, 2022d). After fine-tuning, OPT-IML outperforms OPT across all benchmarks.…”
Section: Othersmentioning
confidence: 99%
“…It's worth noting that currently, open-source language models are larger and more diverse than vision models, making it easy to scale up our foundation models. For example, the largest OPT [98] has 175B parameters, while ViT-G [96] only has 1.8B.…”
Section: Progressive Pre-trainingmentioning
confidence: 99%
“…We use the UMT in both stages, significantly reducing the training sources and speeding up convergence. Thanks to readily-available image and language foundation models [62,59,53,98,16], our simple framework is easily scalable for video foundation models.…”
Section: Introductionmentioning
confidence: 99%
“…Given the ML-ESG task's limited sample size (around 1200) and imbalanced label distribution, training on the available data alone is insufficient to fully train on 35 labels. To overcome these challenges, we employed three renowned open-source generative models: Pythia (Biderman et al, 2023), CerebrasGPT (Dey et al, 2023), and OPT (Zhang et al, 2022). Due to limitations in computational resources, we utilized a 12B model for Pythia, while CerebrasGPT and OPT utilized 13B models.…”
Section: Approachesmentioning
confidence: 99%