2021
DOI: 10.1007/s12145-021-00601-w
|View full text |Cite
|
Sign up to set email alerts
|

A comprehensive social media data processing and analytics architecture by using big data platforms: a case study of twitter flood-risk messages

Abstract: The main objective of the article is to propose an advanced architecture and workflow based on Apache Hadoop and Apache Spark big data platforms. The primary purpose of the presented architecture is collecting, storing, processing, and analysing intensive data from social media streams. This paper presents how the proposed architecture and data workflow can be applied to analyse Tweets with a specific flood topic. The secondary objective, trying to describe the flood alert situation by using only Tweet message… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(3 citation statements)
references
References 40 publications
0
3
0
Order By: Relevance
“…In this framework, the data collection layer employs the Kafka tool to ingest data, transferring it to the ETL layer for necessary transformations and processing before storing it in the data warehouse. Similarly, (12) used another tool called Apache Flume for data collection, representing the data ingestion phase. Another innovative framework presented by (13) uses two tools, Apache Kafka and Apache Flume, as an ingestion data layer to address the challenges of data heterogeneity and velocity.…”
Section: Types and Classificationmentioning
confidence: 99%
See 1 more Smart Citation
“…In this framework, the data collection layer employs the Kafka tool to ingest data, transferring it to the ETL layer for necessary transformations and processing before storing it in the data warehouse. Similarly, (12) used another tool called Apache Flume for data collection, representing the data ingestion phase. Another innovative framework presented by (13) uses two tools, Apache Kafka and Apache Flume, as an ingestion data layer to address the challenges of data heterogeneity and velocity.…”
Section: Types and Classificationmentioning
confidence: 99%
“…With the appropriate data ingestion tool, one can instantly import, process and store data from various data sources. (12) A first category of tools comprises software that iteratively collects data through preconceived and industrialized tasks. Most of these tools are offered by the Apache Foundation, and can also be used to aggregate, convert and clean data before ingestion.…”
Section: Types and Classificationmentioning
confidence: 99%
“…Furthermore, organizations are considered to operate with ethical care, but on the other hand, they are collecting data without consent. It may create a conflict of interest and ethical violations (Podhoranyi, 2021). This study aims to fill the literature gap regarding the ethicality of social media analytical practices and how it harms consumer self-disclosure and marketing discomfort.…”
Section: Theoretical Framework and Hypothesismentioning
confidence: 99%