2019
DOI: 10.1609/aaai.v33i01.33016991
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Embed Sentences Using Attentive Recursive Trees

Abstract: Sentence embedding is an effective feature representation for most deep learning-based NLP tasks. One prevailing line of methods is using recursive latent tree-structured networks to embed sentences with task-specific structures. However, existing models have no explicit mechanism to emphasize taskinformative words in the tree structure. To this end, we propose an Attentive Recursive Tree model (AR-Tree), where the words are dynamically located according to their importance in the task. Specifically, we constr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
1
1
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 21 publications
0
2
0
Order By: Relevance
“…Linguistic structure induction from text. Recent work has proposed several approaches for inducing latent syntactic structures, including constituency trees (Choi et al, 2018;Yogatama et al, 2017;Maillard and Clark, 2018;Havrylov et al, 2019;Kim et al, 2019;Drozdov et al, 2019) and dependency trees (Shi et al, 2019), from the distant supervision of downstream tasks. However, most of the methods are not able to produce linguistically sound structures, or even consistent ones with fixed data and hyperparameters but different random initializations (Williams et al, 2018).…”
Section: Related Workmentioning
confidence: 99%
“…Linguistic structure induction from text. Recent work has proposed several approaches for inducing latent syntactic structures, including constituency trees (Choi et al, 2018;Yogatama et al, 2017;Maillard and Clark, 2018;Havrylov et al, 2019;Kim et al, 2019;Drozdov et al, 2019) and dependency trees (Shi et al, 2019), from the distant supervision of downstream tasks. However, most of the methods are not able to produce linguistically sound structures, or even consistent ones with fixed data and hyperparameters but different random initializations (Williams et al, 2018).…”
Section: Related Workmentioning
confidence: 99%
“…Liu et al [16] proposed attentive tree-structured LSTM for VQA. To address the unbalanced distribution of weights, Shi et al [17] proposed an attentive recursive neural network for sentence embedding, which integrated task-specific attention mechanism into Tree-LSTM. Geng et al [18] utilized attentive Tree-LSTM and sequential respectively to extract semantic relation, and proved the effectiveness of the attention mechanism.…”
Section: Related Workmentioning
confidence: 99%