2015 3rd International Conference on Future Internet of Things and Cloud 2015
DOI: 10.1109/ficloud.2015.15
|View full text |Cite
|
Sign up to set email alerts
|

Statistical Modeling and Visualizing Open Big Data Using a Terrorism Case Study

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0
1

Year Published

2016
2016
2022
2022

Publication Types

Select...
6
3

Relationship

4
5

Authors

Journals

citations
Cited by 24 publications
(12 citation statements)
references
References 25 publications
0
11
0
1
Order By: Relevance
“…Actually, several researchers had already pointed out that a barrier to performing big data analytics was that most statistical software could not handle the large file sizes. 9 However, researchers have found ways around the big data five Vs—at least the volume, velocity, and variety attributes—by using cloud-based and distributed software such as Hadoop along with sampling techniques to reduce the five Vs. 8,48,49 Nonetheless, this is where another hidden healthcare big data problem lurks. There are several tacit issues that revolve around research design assumptions and statistical sampling assumptions.…”
Section: Literature Review and Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Actually, several researchers had already pointed out that a barrier to performing big data analytics was that most statistical software could not handle the large file sizes. 9 However, researchers have found ways around the big data five Vs—at least the volume, velocity, and variety attributes—by using cloud-based and distributed software such as Hadoop along with sampling techniques to reduce the five Vs. 8,48,49 Nonetheless, this is where another hidden healthcare big data problem lurks. There are several tacit issues that revolve around research design assumptions and statistical sampling assumptions.…”
Section: Literature Review and Discussionmentioning
confidence: 99%
“…8 Value in big data can be viewed as a constraint because it can be challenging to derive a benefit from analytics that is worth the investment time and cost to accommodate the other factors. Big data veracity can refer to ethics, accuracy, validity, or truthfulness 9 as well as social-cultural relevance. 10 In addition to the above characteristics, each discipline and industry has unique big data analytics issues.…”
Section: Introductionmentioning
confidence: 99%
“…The use of parametric statistical techniques requires rigorous designs that ensure the prerequisites of the data are satisfied including distribution, population-sample homogeneity, sample group size, data type, and other inferential thresholds including collinearity and variance tolerance (Strang, 2015d). Learning analytics software generally involves nonparametric distribution-free nonlinear techniques utilised in big data analytics (Chatti et al, 2012, p.10;Strang and Sun, 2015;Sun, Strang and Yearwood, 2014;Xing et al, 2015), which include cluster analysis, neural network analysis with Bayes probability theory, nonlinear math programming, correspondence analysis, and genetic nonlinear programming (Nersesian and Strang, 2013;Strang, 2012;Vajjhala, Strang and Sun, 2015;Xing et al, 2015). The strategy for this study is to accept learning analytics as a 'black box' big data summarisation tool by using its output for input into the unit of analysis during hypothesis testing.…”
Section: Empirical Learning Analytics Researchmentioning
confidence: 99%
“…It is impractical to define a universal threshold for big data volume (i.e., what constitutes a 'big dataset') because the time and type of data can influence its definition [23]. Currently, datasets that reside in the exabyte (EB) or ZB ranges are generally considered as big data [8,24], however challenges still exist for datasets in smaller size ranges. For example, Walmart collects 2.5 PB from over a million customers every hour [25].…”
Section: Big Datamentioning
confidence: 99%