2020
DOI: 10.1609/aaai.v34i05.6284
|View full text |Cite
|
Sign up to set email alerts
|

Two-Level Transformer and Auxiliary Coherence Modeling for Improved Text Segmentation

Abstract: Breaking down the structure of long texts into semantically coherent segments makes the texts more readable and supports downstream applications like summarization and retrieval. Starting from an apparent link between text coherence and segmentation, we introduce a novel supervised model for text segmentation with simple but explicit coherence modeling. Our model – a neural architecture consisting of two hierarchically connected Transformer networks – is a multi-task learning model that couples the sentence-le… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

2
34
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 32 publications
(36 citation statements)
references
References 23 publications
(66 reference statements)
2
34
0
Order By: Relevance
“…The central ideas of each topic is summarized with one sentence, covering information from multiple utterances, i.e., t 1 for S 1 , t 2 for S 2 , and t 3 for S 3 . We also observe that utterances residing in the same topic (e.g., S 1 , S 2 and S 3 ) is inherently more coherent than those coming from different topics (e.g., the inter-topic snippet S 4 and S 5 ), which reveals the underlying relationships between topic and utterance coherence, also demonstrated by Glavaš and Somasundaran (2020).…”
Section: Introductionsupporting
confidence: 52%
“…The central ideas of each topic is summarized with one sentence, covering information from multiple utterances, i.e., t 1 for S 1 , t 2 for S 2 , and t 3 for S 3 . We also observe that utterances residing in the same topic (e.g., S 1 , S 2 and S 3 ) is inherently more coherent than those coming from different topics (e.g., the inter-topic snippet S 4 and S 5 ), which reveals the underlying relationships between topic and utterance coherence, also demonstrated by Glavaš and Somasundaran (2020).…”
Section: Introductionsupporting
confidence: 52%
“…The closest model to ours is proposed in (Glava and Somasundaran, 2020) 1 where transformers are used for both the levels of the architecture. They also developed a semantic coherence measure on distinguishing pairs of genuine and fake text snippets as an auxiliary loss alongside the segment classification loss.…”
Section: Supervised Segmentation Modelsmentioning
confidence: 99%
“…However, with the advancement of self-learning and transfer learning on deep neural networks, there are more recent supervised modeling approaches proposed that aim to predict labeled segment boundaries on smaller datasets. (Koshorek et al, 2018;Xing et al, 2020;Barrow et al, 2020;Glava and Somasundaran, 2020) To the best of our knowledge, the most straightforward remedy to the above problems is knowledge transfer and distillation from pre-trained models. The rich pre-trained knowledge enables the training of a more general segmentation model on a small labeled dataset.…”
Section: Introductionmentioning
confidence: 99%
“…Badjatiya et al [38] proposed an attention-based convolutional neural network bidirectional LSTM model that introduced the attention mechanism and learned the relative importance of each sentence in the text to achieve segmentation. Glavaš et al [39] proposed a multitask learning model that couples the sentence-level segmentation objective with the coherence objective that differentiates correct sequences of sentences from corrupt ones.…”
Section: Text Segmentationmentioning
confidence: 99%