2020
DOI: 10.1016/j.patcog.2020.107315
|View full text |Cite
|
Sign up to set email alerts
|

Textual data summarization using the Self-Organized Co-Clustering model

Abstract: Recently, different studies have demonstrated the use of co-clustering, a data mining technique which simultaneously produces row-clusters of observations and column-clusters of features. The present work introduces a novel co-clustering model to easily summarize textual data in a document-term format. In addition to highlighting homogeneous co-clusters as other existing algorithms do we also distinguish noisy co-clusters from significant co-clusters, which is particularly useful for sparse document-term matri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(7 citation statements)
references
References 24 publications
0
4
0
Order By: Relevance
“…Indeed, in LDA the notion of cluster of terms is replaced by topics, which is a kind of soft clustering of the terms (each term belonging to each topic with different probability). Nevertheless, some recent works in co-clustering for text mining have been proposed in order to ease the reading of the coclustering results, by designing explicitly which are the clusters of terms specific to each cluster of documents [Laclau and Nadif, 2016, Ailem et al, 2017, Selosse et al, 2020c (see Section 5.4 for more details).…”
Section: Some Typical Lbm Use Casesmentioning
confidence: 99%
“…Indeed, in LDA the notion of cluster of terms is replaced by topics, which is a kind of soft clustering of the terms (each term belonging to each topic with different probability). Nevertheless, some recent works in co-clustering for text mining have been proposed in order to ease the reading of the coclustering results, by designing explicitly which are the clusters of terms specific to each cluster of documents [Laclau and Nadif, 2016, Ailem et al, 2017, Selosse et al, 2020c (see Section 5.4 for more details).…”
Section: Some Typical Lbm Use Casesmentioning
confidence: 99%
“…In fact, clustering is a promising approach for topic modelling as well as for other NLP tasks. Even though Selosse et al (2020) focus on data summarization, they propose a unique co-clustering approach, which may be useful for topic alignment. Their method leads to the identification of "homogeneous co-clusters," which is also accomplished by a range of alternative algorithms, but the study also adds value by contrasting "noisy co-clusters" with "significant co-clusters, which is particularly useful for sparse document-term matrices."…”
Section: Topic Sentiment: Similarity Modelingmentioning
confidence: 99%
“…In fact, clustering is a promising approach for topic modelling as well as for other NLP tasks. Even though Selosse et al (2020) focus on data summarization, they propose a unique co-clustering approach, which may be useful for topic alignment. Their method leads to the identification of "homogeneous co-clusters", which is also accomplished by a range of alternative algorithms, but the study also adds value by contrasting "noisy co-clusters" with "significant coclusters, which is particularly useful for sparse document-term matrices".…”
Section: Topic Sentiment: Similarity Modelingmentioning
confidence: 99%