2018
DOI: 10.3390/data3040066
|View full text |Cite
|
Sign up to set email alerts
|

Similar Text Fragments Extraction for Identifying Common Wikipedia Communities

Abstract: Similar text fragments extraction from weakly formalized data is the task of natural language processing and intelligent data analysis and is used for solving the problem of automatic identification of connected knowledge fields. In order to search such common communities in Wikipedia, we propose to use as an additional stage a logical-algebraic model for similar collocations extraction. With Stanford Part-Of-Speech tagger and Stanford Universal Dependencies parser, we identify the grammatical characteristics … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 11 publications
0
1
0
Order By: Relevance
“…Characterizing Wikipedia readers. A substantial amount of prior work has focused on understanding user engagement with Wikipedia from the point of view of the editor community [2,41,43,68]. Studies on the behavior of Wikipedia readers have mostly considered interest in contents [31,49,63], content popularity [8,47,56], or event timing [37].…”
Section: Related Workmentioning
confidence: 99%
“…Characterizing Wikipedia readers. A substantial amount of prior work has focused on understanding user engagement with Wikipedia from the point of view of the editor community [2,41,43,68]. Studies on the behavior of Wikipedia readers have mostly considered interest in contents [31,49,63], content popularity [8,47,56], or event timing [37].…”
Section: Related Workmentioning
confidence: 99%