Proceedings of the 55th Annual Meeting of the Association For Computational Linguistics (Volume 1: Long Papers) 2017
DOI: 10.18653/v1/p17-1121
|View full text |Cite
|
Sign up to set email alerts
|

A Neural Local Coherence Model

Abstract: We propose a local coherence model based on a convolutional neural network that operates over the entity grid representation of a text. The model captures long range entity transitions along with entity-specific features without loosing generalization, thanks to the power of distributed representation. We present a pairwise ranking method to train the model in an end-to-end fashion on a task and learn task-specific high level features. Our evaluation on three different coherence assessment tasks demonstrates t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
89
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 65 publications
(89 citation statements)
references
References 15 publications
0
89
0
Order By: Relevance
“…Neural Grid (N&J) 3 . These are the neural versions of the entity grid models as proposed by (Tien Nguyen and Joty, 2017). They use convolutions over grammatical roles to model entity transitions in the distributed space.…”
Section: Models Comparedmentioning
confidence: 99%
“…Neural Grid (N&J) 3 . These are the neural versions of the entity grid models as proposed by (Tien Nguyen and Joty, 2017). They use convolutions over grammatical roles to model entity transitions in the distributed space.…”
Section: Models Comparedmentioning
confidence: 99%
“…To reduce model variance, we run the WSJ experiments 5 times with different random initializations and the GCDC ones 10 times (following Lai and Tetreault (2018)), and average the predicted scores of the ensembles for the final evaluation. For the WSJ data, we use the same train/dev splits as Tien Nguyen and Joty (2017), and for GCDC, we follow Lai and Tetreault (2018) and split the training data with a 9:1 ratio for tuning.…”
Section: Methodsmentioning
confidence: 99%
“…The first task, discrimination, is usually evaluated as accuracy of the model in ranking the original text higher than a permuted one (we use 20 permutations per document following previous work [11,14,30]). In order to better analyse our results, we add to this metric two widely used ranking metrics, i.d.…”
Section: Methodsmentioning
confidence: 99%
“…The typical tasks on which local coherence models are currently evaluated are: sentence ordering discrimination, where the system needs to rank original documents higher than randomly permuted ones, and insertion, introduced by [14], where the system has to rank the position of a sentence removed from a document. The state of the art for these tasks was recently achieved by [30], which uses the entity grid as input to a Convolutional Neural Network. Table 1: We report the count of Dialogue Act tags, the average number of tokens per turn, the average count of turns per dialogue and counts of our Training, test and Developments splits for our three datasets: SWBD, AMI and Oasis.…”
Section: The Dialogue Act Tags Are Directly Taken From the Swbd Damslmentioning
confidence: 99%