2008
DOI: 10.1007/978-1-84800-046-9
|View full text |Cite
|
Sign up to set email alerts
|

Survey of Text Mining II

Abstract: transmitted, in any form or by any means, with the prior permission in writing of the publishers, or in the case of reprographic reproduction in accordance with the terms of licences issued by the Copyright Licensing Agency. Enquiries concerning reproduction outside those terms should be sent to the publishers. The use of registered names, trademarks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant laws and regulations and there… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2012
2012
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 57 publications
(5 citation statements)
references
References 98 publications
0
5
0
Order By: Relevance
“…We have also tested the modified Lee-Seung method of Finesso and Spreij [15,51], but it is only a scaled version of the Lee-Seung method for KL divergence and gave nearly identical results which are omitted. All implementations are from Version 2.5 of Tensor Toolbox for MATLAB [4,3,2]; exact parameter settings are provided in Appendix E. We report the factor match score (FMS), a measure in [0, 1] of how close the computed solution is to the true solution. A value of 1 is ideal.…”
Section: Comparison Of Objective Functions For Sparse Count Datamentioning
confidence: 99%
See 1 more Smart Citation
“…We have also tested the modified Lee-Seung method of Finesso and Spreij [15,51], but it is only a scaled version of the Lee-Seung method for KL divergence and gave nearly identical results which are omitted. All implementations are from Version 2.5 of Tensor Toolbox for MATLAB [4,3,2]; exact parameter settings are provided in Appendix E. We report the factor match score (FMS), a measure in [0, 1] of how close the computed solution is to the true solution. A value of 1 is ideal.…”
Section: Comparison Of Objective Functions For Sparse Count Datamentioning
confidence: 99%
“…In this paper, we consider the problem of multilinear modeling of sparse count data. For instance, we may consider data that encodes the number of papers published by each author at each conference per year for a given time frame [13], the number of packets sent from one IP address to another using a specific port [47], or to/from and term counts on emails [2]. Our goal is to develop a descriptive model of such data, along with appropriate algorithms and theory.…”
mentioning
confidence: 99%
“…In contrast, other text mining methods such as topic modelling and clustering can analyse groups of words together instead of counting them separately (Berry and Castellanos, 2008). In this way, these techniques can consider the variations in each word's meaning across different contexts.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Although conventional clustering methods overcome some of the limitations of term-frequency-based techniques, topic modelling presents several advantages over them. Clustering is based on the idea of classifying documents into different groups based on a specific similarity measure (Berry and Castellanos, 2008). With clustering, every document (reviews, in our case) appears only in one group (such as topics or attributes).…”
Section: Literature Reviewmentioning
confidence: 99%
“…Therefore we have to carefully define the experimentation environment [12] and the dataset that we use in our tests. tion we can show in Figure 4.…”
Section: Development Of Evolutionary System For Document Clusteringmentioning
confidence: 99%