Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-70
|View full text |Cite
|
Sign up to set email alerts
|

Gated Convolutional Neural Network for Sentence Matching

Abstract: The recurrent neural networks (RNN) have shown promising results in sentence matching tasks, such as paraphrase identification (PI), natural language inference (NLI) and answer selection (AS). However, the recurrent architecture prevents parallel computation within a sequence and is highly time-consuming. To overcome this limitation, we propose a gated convolutional neural network (GCNN) for sentence matching tasks. In this model, the stacked convolutions encode hierarchical contextaware representations of a s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
91
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
3
2

Relationship

1
9

Authors

Journals

citations
Cited by 85 publications
(91 citation statements)
references
References 18 publications
0
91
0
Order By: Relevance
“…These neurons are responsible for covering small sub-areas of the visual area, named the receptive area. Then, the receptive areas are tiled to detect the overall visual area [49]. Hence, receptive areas are deemed as filters in the CNN deep learning model.…”
Section: A Convolutional Neural Networkmentioning
confidence: 99%
“…These neurons are responsible for covering small sub-areas of the visual area, named the receptive area. Then, the receptive areas are tiled to detect the overall visual area [49]. Hence, receptive areas are deemed as filters in the CNN deep learning model.…”
Section: A Convolutional Neural Networkmentioning
confidence: 99%
“…Ferreira et al (2018) feed sentence similarity measured with hand-crafted features to machine learning algorithms. Convolutional neural networks have been introduced by Yin and Schütze (2015) and Chen et al (2018), and further augmented with LSTMs (Kubal and Nimkar, 2018) and attention mechanisms (Fan et al, 2018).…”
Section: Leveraging Paraphrases In Nlpmentioning
confidence: 99%
“…In this section, we will employ the gating mechanisms for the xvector system in two ways. As depicted in Figure 1, we first use four GCNN [15] layers to replace the first four TDNN layers for extracting the frame-level representations. On the other hand, we propose the gated-attention statistics pooling as an alternative attention method for aggregating the frame-level vectors.…”
Section: Dnn With Gating Mechanismsmentioning
confidence: 99%