Proceedings of the First International Conference on Human Language Technology Research - HLT '01 2001
DOI: 10.3115/1072133.1072142
|View full text |Cite
|
Sign up to set email alerts
|

Automatic pattern acquisition for Japanese information extraction

Abstract: One of the central issues for information extraction is the cost of customization from one scenario to another. Research on the automated acquisition of patterns is important for portability and scalability. In this paper, we introduce Tree-Based Pattern representation where a pattern is denoted as a path in the dependency tree of a sentence. We outline the procedure to acquire Tree-Based Patterns in Japanese from un-annotated text. The system extracts the relevant sentences from the training data based on TF/… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
34
0
1

Year Published

2004
2004
2014
2014

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 25 publications
(35 citation statements)
references
References 1 publication
0
34
0
1
Order By: Relevance
“…To automatically decide whether a sentence is definitional, we could use a simple cutoff in which sentences that are ranked more highly are considered definitional. This is similar to the work by Sudo et al [19] who proposed unsupervised learning method for pattern discovery by utilizing TF-IDF weight to select a set of relevant documents and sentences, and then built patterns from them.…”
Section: Step 2: Unsupervised Labeling Using Prfmentioning
confidence: 64%
See 1 more Smart Citation
“…To automatically decide whether a sentence is definitional, we could use a simple cutoff in which sentences that are ranked more highly are considered definitional. This is similar to the work by Sudo et al [19] who proposed unsupervised learning method for pattern discovery by utilizing TF-IDF weight to select a set of relevant documents and sentences, and then built patterns from them.…”
Section: Step 2: Unsupervised Labeling Using Prfmentioning
confidence: 64%
“…Similarly, Yangarber et al [21] used a set of basic patterns as "seeds" and learn more scenario oriented extraction patterns automatically. Most relevant to our application of PRF, Sudo et al [19] put forward an unsupervised learning for pattern discovery. They utilized TF脳IDF to get a set of relevant documents and sentences and built patterns from them.…”
Section: Related Workmentioning
confidence: 99%
“…Several recent approaches to IE have used patterns based on a dependency analysis of the input text (Yangarber, 2003;Sudo et al, 2001;Sudo et al, 2003;Bunescu and Mooney, 2005;. These approaches have used a variety of pattern models (schemes for representing IE patterns based on particular parts of the dependency tree).…”
Section: Introductionmentioning
confidence: 99%
“…The predicate-argument (SVO) model allows subtrees containing only a verb and its direct subject and object as extraction pattern candidates (Yangarber, 2003). The chain model represents extraction patterns as a chain-shaped path from each target slot value to the root node of the dependency tree (Sudo et al, 2001). A couple of chain model patterns sharing the same verb are linked to each other and construct a linked-chain model pattern (Greenwood and Stevenson, 2006).…”
Section: Introductionmentioning
confidence: 99%
“…A number of pattern induction approaches have recently been researched based on the dependency analysis (Yangarber, 2003) (Sudo et al, 2001) (Greenwood and Stevenson, 2006) (Sudo et al, 2003). The natural language texts in training instances are parsed by dependency analyzer and converted into dependency trees.…”
Section: Introductionmentioning
confidence: 99%