DOI: 10.3990/1.9789036541237
|View full text |Cite
|
Sign up to set email alerts
|

Deep web content monitoring

Abstract: The research reported in this thesis has been carried out under the auspices of SIKS,

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 50 publications
(155 reference statements)
0
2
0
Order By: Relevance
“…Another approach uses the TF-IDF as keyword or appropriate word extraction method, link classification method, and classification technique for its crawler. In some approaches, it is proposed to use the TF-IDF algorithm with various modifications based on the specific requirements of the crawler [4][5][6]. An approach discusses the use of TF-ICF (Term Frequency-Inverse Class Frequency), which calculates the popularity score for various categories instead of TF-IDF which calculates the popularity score of words.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Another approach uses the TF-IDF as keyword or appropriate word extraction method, link classification method, and classification technique for its crawler. In some approaches, it is proposed to use the TF-IDF algorithm with various modifications based on the specific requirements of the crawler [4][5][6]. An approach discusses the use of TF-ICF (Term Frequency-Inverse Class Frequency), which calculates the popularity score for various categories instead of TF-IDF which calculates the popularity score of words.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Second, the different query execution and document retrieval and processing schedules that we discuss in Section 3, and that we evaluate in Section 5, can lead to fundamentally different (e.g., in terms of quality and efficiency) focused crawling executions. Importantly, our sampling strategies are crucial for other important building blocks of deep-web crawling, in general, namely, automatic filling of search forms (Kantorski et al, 2015) and content monitoring (Mohammad Khelghati, 2016), since they require high-quality and efficient document samples from the collection to select which queries to issue and to decide when to update the content summary of the collection, respectively.…”
Section: Related Workmentioning
confidence: 99%