2020
DOI: 10.1186/s12859-020-03886-8
|View full text |Cite
|
Sign up to set email alerts
|

C-Norm: a neural approach to few-shot entity normalization

Abstract: Background Entity normalization is an important information extraction task which has gained renewed attention in the last decade, particularly in the biomedical and life science domains. In these domains, and more generally in all specialized domains, this task is still challenging for the latest machine learning-based approaches, which have difficulty handling highly multi-class and few-shot learning problems. To address this issue, we propose C-Norm, a new neural approach which synergistical… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 14 publications
(6 citation statements)
references
References 27 publications
0
6
0
Order By: Relevance
“…Named entity linking . Regarding the NEL evaluation, we assessed the performance of C-Norm [ 42 ] and BioSyn [ 45 ] for the prediction of trait and phenotype classes. We did not evaluate the prediction of species because the size of the NCBI taxonomy is beyond the capacity of the algorithms.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Named entity linking . Regarding the NEL evaluation, we assessed the performance of C-Norm [ 42 ] and BioSyn [ 45 ] for the prediction of trait and phenotype classes. We did not evaluate the prediction of species because the size of the NCBI taxonomy is beyond the capacity of the algorithms.…”
Section: Resultsmentioning
confidence: 99%
“…For the NEL task involving traits and phenotypes, we use the C-Norm and BioSyn algorithms. The C-Norm method [ 42 ] achieves state-of-the-art performance on the Bacteria Biotope dataset, which has good similarities to TaeC , i.e., deep ontology and complex entity terms. C-Norm represents terms in the texts using Word2vec embeddings [ 43 ], and it represents ontology classes using vectors that integrate hierarchical information from the ontology [ 44 ].…”
Section: Methodsmentioning
confidence: 99%
“…Several strategies were developed and investigated to exploit external lexical and semantic resources to improve machine learning models. These strategies include thematic masking [1] , named entity recognition by distant supervision [2] , and ontology-based normalization [3] . The biological roles of MOs depend mainly on their structure.…”
Section: Value Of the Datamentioning
confidence: 99%
“…INRA (Institut national de la recherche agronomique) and Cnrs (Centre national de la recherche scientifique) at University Paris Saclay proposed a two-step method to normalize multi-word terms with concepts from a domain-specific ontology. In this method, they used vector representations of terms computed with word embedding information and hierarchical information from ontology concepts [16]. Le and Mikolov presented word2vec and later introduced the doc2vec algorithm based on adjusted techniques for learning how to embed texts identical to word2vec, thus turning doc2vec into a branch of word2vec [17].…”
Section: Related Workmentioning
confidence: 99%