2020
DOI: 10.1016/j.knosys.2019.104987
|View full text |Cite
|
Sign up to set email alerts
|

Probabilistic data structures for big data analytics: A comprehensive review

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
3

Relationship

3
7

Authors

Journals

citations
Cited by 43 publications
(21 citation statements)
references
References 42 publications
0
21
0
Order By: Relevance
“…This is due to the large memory and high latency issues for processing queries using these traditional data structures. The probabilistic data structure, as a kind of data structure, is particularly advantageous for large data, because it reduces latency, and analytical procedures [27]. They are tremendously handy data systems for reducing the space and time trade-off, and to a great extent, equivalent to retrieval and storage for querying of data [28].…”
Section: Probabilistic Data Structurementioning
confidence: 99%
“…This is due to the large memory and high latency issues for processing queries using these traditional data structures. The probabilistic data structure, as a kind of data structure, is particularly advantageous for large data, because it reduces latency, and analytical procedures [27]. They are tremendously handy data systems for reducing the space and time trade-off, and to a great extent, equivalent to retrieval and storage for querying of data [28].…”
Section: Probabilistic Data Structurementioning
confidence: 99%
“…Hence, it is assumed that the probabilistic data structure, as a group of data structures, is used in the proposed model. This kind of data structure is extremely useful for big data, because it reduces latency and the analytical process [18]. The bloom filter (BF), cuckoo filter (CF) and quotient filter (QF) are three different types of space-efficient probabilistic data structures that are used to check whether an element is a member of a massive dataset or not.…”
Section: Type Of Node Detectionmentioning
confidence: 99%
“…A rather new technique to approach the problem of processing large datasets are Probabilistic Data Structures [41]. A specific example is a cardinality estimation algorithm called HyperLogLog [42].…”
Section: Privacy-aware Storagementioning
confidence: 99%