2009 Ninth IEEE International Conference on Data Mining 2009
DOI: 10.1109/icdm.2009.40
|View full text |Cite
|
Sign up to set email alerts
|

Semi-Supervised Sequence Labeling with Self-Learned Features

Abstract: Abstract-Typical information extraction (IE) systems can be seen as tasks assigning labels to words in a natural language sequence. The performance is restricted by the availability of labeled words. To tackle this issue, we propose a semisupervised approach to improve the sequence labeling procedure in IE through a class of algorithms with self-learned features (SLF). A supervised classifier can be trained with annotated text sequences and used to classify each word in a large set of unannotated sentences. By… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0
1

Year Published

2010
2010
2019
2019

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 17 publications
(10 citation statements)
references
References 27 publications
0
9
0
1
Order By: Relevance
“…Semi-supervised learning is useful for improving model performances when a target domain or language lacks of manual resources. Self-training is a commonly used strategy for various natural language processing (NLP) tasks, such as named entity recognition (NER) (Kozareva et al, 2005), part-of-speech (POS) tagging (Wang et al, 2007;Qi et al, 2009) and parsing (McClosky et al, 2006(McClosky et al, , 2008Sagae, 2010).…”
Section: Introductionmentioning
confidence: 99%
“…Semi-supervised learning is useful for improving model performances when a target domain or language lacks of manual resources. Self-training is a commonly used strategy for various natural language processing (NLP) tasks, such as named entity recognition (NER) (Kozareva et al, 2005), part-of-speech (POS) tagging (Wang et al, 2007;Qi et al, 2009) and parsing (McClosky et al, 2006(McClosky et al, , 2008Sagae, 2010).…”
Section: Introductionmentioning
confidence: 99%
“…Our proximity features are similar in spirit to selflabeled features previously studied for tagging problems [10]. Qi, et al proposed an iterative scheme, where feature vectors in each iteration are augmented with the predicted word-level class label distributions from the previous iteration in a semi-supervised manner.…”
Section: Related Workmentioning
confidence: 99%
“…Our current work does not feature this aspect, but we regard it as one of the key next steps to be tackled. Consequently, we mention the research performed by Schwartz et al 27 Teufel et al, 28 or Wu et al 29 that deal with using citation contexts for discerning a citation's function and analyzing how this influences or is influenced by the work it points to.…”
Section: Flux-cimmentioning
confidence: 99%