2022
DOI: 10.48550/arxiv.2211.02849
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Coarse-to-fine Knowledge Graph Domain Adaptation based on Distantly-supervised Iterative Training

Abstract: Modern supervised learning neural network models require a large amount of manually labeled data, which makes the construction of domain-specific knowledge graphs timeconsuming and labor-intensive. In parallel, although there has been much research on named entity recognition and relation extraction based on distantly supervised learning, constructing a domain-specific knowledge graph from large collections of textual data without manual annotations is still an urgent problem to be solved. In response, we prop… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 29 publications
(34 reference statements)
0
2
0
Order By: Relevance
“…Datasets used to train language models can be substantial (1,29), often reaching hundreds of gigabytes, and they draw from various sources and domains (30)(31)(32)(33)(34). Consequently, even when trained on public data, these datasets can contain sensitive information, such as personally identifiable information (PII) including names, phone numbers, and addresses.…”
Section: The Emperor's New Clothes: Privacy Concernsmentioning
confidence: 99%
“…Datasets used to train language models can be substantial (1,29), often reaching hundreds of gigabytes, and they draw from various sources and domains (30)(31)(32)(33)(34). Consequently, even when trained on public data, these datasets can contain sensitive information, such as personally identifiable information (PII) including names, phone numbers, and addresses.…”
Section: The Emperor's New Clothes: Privacy Concernsmentioning
confidence: 99%
“…Transformer [1] was first proposed for translation tasks in NLP, which combines Multi-head Self Attention (MSA) with Feed-forward Networks (FFN) to offer a global perceptual field and multi-channel feature extraction capabilities. The subsequent development of the Transformer-based BERT [67] proved to be seminal in NLP, exhibiting exceptional performance across multiple language-related tasks [15,25,68]. Leveraging the great flexibility and scalability of the Transformer, researchers have started to train larger Transformer models, including GPT-1 [4], GPT-2 [5], GPT-3 [20], GPT-4 [22], T5 [3], PaLM [69], LLaMA [70] and others.…”
Section: Foundation Modelsmentioning
confidence: 99%