2009
DOI: 10.1007/s10994-009-5106-x
|View full text |Cite
|
Sign up to set email alerts
|

Search-based structured prediction

Abstract: We present SEARN, an algorithm for integrating SEARch and lEARNing to solve complex structured prediction problems such as those that occur in natural language, speech, computational biology, and vision. SEARN is a meta-algorithm that transforms these complex problems into simple classification problems to which any binary classifier may be applied. Unlike current algorithms for structured learning that require decomposition of both the loss function and the feature functions over the predicted structure, SEAR… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
383
0

Year Published

2013
2013
2019
2019

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 349 publications
(385 citation statements)
references
References 33 publications
2
383
0
Order By: Relevance
“…The use of sequential predictions-feeding the output of predictors from a previous stage to the next-has been revisited in the literature from time to time. Methods such as [23,24] applied sequential prediction to natural language processing tasks. While [25] explored the use of context from neighboring pixel classifiers for computer vision tasks.…”
Section: Related Workmentioning
confidence: 99%
“…The use of sequential predictions-feeding the output of predictors from a previous stage to the next-has been revisited in the literature from time to time. Methods such as [23,24] applied sequential prediction to natural language processing tasks. While [25] explored the use of context from neighboring pixel classifiers for computer vision tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Such algorithms decompose a joint prediction task into a sequence of action predictions, such as predicting the label of the next word in sequence labeling or predicting a shift/reduce action in dependency parsing 3 ; these predictions are tied by features and/or internal state. Algorithms in this family have recently met with success in neural networks (Bengio et al, 2015;Wiseman and Rush, 2016), though date back to models typically based on linear policies (Collins and Roark, 2004;Daumé III and Marcu, 2005;Xu et al, 2007;Daumé III et al, 2009;Ross et al, 2010;Ross and Bagnell, 2014;Doppa et al, 2014;Chang et al, 2015).…”
Section: Learning With Bandit Feedbackmentioning
confidence: 99%
“…Imitation learning 1 algorithms such as SEARN (Daumé III et al, 2009) and DAGGER (Ross et al, 2011) have been applied successfully to a variety of structured prediction tasks due to their flexibility in incorporating features and their ability to learn with non-decomposable loss functions. Sample applications include biomedical event extraction (Vlachos and Craven, 2011), dynamic feature selection (He et al, 2013), and machine translation (Grissom II et al, 2014).…”
Section: Imitation Learningmentioning
confidence: 99%