2019
DOI: 10.48550/arxiv.1903.10676
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

SciBERT: A Pretrained Language Model for Scientific Text

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
341
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 244 publications
(347 citation statements)
references
References 0 publications
5
341
1
Order By: Relevance
“…We also fine-tune these models with text from our target datasets using the standard language model training objective to evaluate the effects of exposing task specific content. Additionally, we compare with SciBERT [40], a language model based on BERT but pre-trained on a large corpus of scientific text, to assess the impact of structure inclusion on different types of Transformer models. For comparisons with equivalent loss functions, we use SentBERT [13] that uses triples of anchor, semantically similar and dissimilar sentences to learn the embeddings.…”
Section: Evaluation Frameworkmentioning
confidence: 99%
“…We also fine-tune these models with text from our target datasets using the standard language model training objective to evaluate the effects of exposing task specific content. Additionally, we compare with SciBERT [40], a language model based on BERT but pre-trained on a large corpus of scientific text, to assess the impact of structure inclusion on different types of Transformer models. For comparisons with equivalent loss functions, we use SentBERT [13] that uses triples of anchor, semantically similar and dissimilar sentences to learn the embeddings.…”
Section: Evaluation Frameworkmentioning
confidence: 99%
“…For representing terms, we used a BERT-based model, called SciBERT [1] with 12 attention layers and the hidden layer of size 768 which was pretrained on 1.14M papers from Semantic Scholar 2 resulting in a corpus size of 3.17B tokens. SciBERT was then fine-tuned on the entire ACL corpus to represent semantic meanings of words in each time frame.…”
Section: B Text Representationmentioning
confidence: 99%
“…We propose a novel change-oriented summarization approach based on word semantic evolution to provide information on key changes that occurred in the data over time. As an underlying data source we use the Association for Computational Linguistics (ACL) dataset, and specifically, the Parcit Structured XML 1 version of the ACL Anthology Reference Corpus [2]. By analyzing this data we provide answers to questions about what concepts are highly changing and what are their directions of changes.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The sub-tokens can then spuriously match other terms, including sub-tokens from other split identifiers, but not the whole, unsplit term. Researchers in the biomedical domain recognized the same issue affecting medical terms and proposed domain-specific BERT adaptations [2,11,26].…”
Section: #Bugsmentioning
confidence: 99%