Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2022
DOI: 10.18653/v1/2022.naacl-main.397
|View full text |Cite
|
Sign up to set email alerts
|

CORWA: A Citation-Oriented Related Work Annotation Dataset

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(9 citation statements)
references
References 14 publications
0
9
0
Order By: Relevance
“…Similarly, PDF parsing/extraction techniques were applied in 65% (n=15) of studies, the remaining studies applied extraction to other document formats (e.g., journal articles available online in HTML format; see . While similar methods, which additionally take into account syntactic structure, including chunking and dependency parsing were less frequently applied (Angrosh et al, 2014;Li et al, 2022;Nayak et al, 2021;Pertsas & Constantopoulos, 2018). Tagging methods, including PoS tagging (assigning grammatical categories, e.g., noun, verb), followed by concept tagging (e.g., semantic annotation), or sequence tagging, where labels were assigned based on order of appearance, were used in 43% (n=15) of studies.…”
Section: Data Preprocessing and Feature Engineeringmentioning
confidence: 99%
See 4 more Smart Citations
“…Similarly, PDF parsing/extraction techniques were applied in 65% (n=15) of studies, the remaining studies applied extraction to other document formats (e.g., journal articles available online in HTML format; see . While similar methods, which additionally take into account syntactic structure, including chunking and dependency parsing were less frequently applied (Angrosh et al, 2014;Li et al, 2022;Nayak et al, 2021;Pertsas & Constantopoulos, 2018). Tagging methods, including PoS tagging (assigning grammatical categories, e.g., noun, verb), followed by concept tagging (e.g., semantic annotation), or sequence tagging, where labels were assigned based on order of appearance, were used in 43% (n=15) of studies.…”
Section: Data Preprocessing and Feature Engineeringmentioning
confidence: 99%
“…Bidirectional Encoder Representations from Transformers (BERT) and other BERT-based language models made up the majority of transformer-based approaches. Specifically BERT (Aumiller et al, 2020;Shen et al, 2022) and SciBERT (Goldfarb-Tarrant et al, 2020;Li et al, 2022) were the most utilized for tasks relevant to extracting data from research in social sciences. Others language models included BioBERT (Chen et al, 2020) and distilBERT (Goldfarb-Tarrant et al, 2020).…”
Section: Model Architectures and Componentsmentioning
confidence: 99%
See 3 more Smart Citations