2023
DOI: 10.48550/arxiv.2302.04156
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Prompting for Multimodal Hateful Meme Classification

Abstract: Hateful meme classification is a challenging multimodal task that requires complex reasoning and contextual background knowledge. Ideally, we could leverage an explicit external knowledge base to supplement contextual and cultural information in hateful memes. However, there is no known explicit external knowledge base that could provide such hate speech contextual information. To address this gap, we propose PromptHate, a simple yet effective prompt-based model that prompts pre-trained language models (PLMs) … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 26 publications
(49 reference statements)
0
3
0
Order By: Relevance
“…Without any training or fine-tuning, a powerful model like CLIP can produce zero-shot predictions. In order to achieve that, we offer the model some text prompts [27][28]. These text labels or prompts are encoded by the CLIP classifier into a learned latent space, and their similarity to the image latent space is assessed.…”
Section: Prompt Engineeringmentioning
confidence: 99%
“…Without any training or fine-tuning, a powerful model like CLIP can produce zero-shot predictions. In order to achieve that, we offer the model some text prompts [27][28]. These text labels or prompts are encoded by the CLIP classifier into a learned latent space, and their similarity to the image latent space is assessed.…”
Section: Prompt Engineeringmentioning
confidence: 99%
“…Each supervised classification tasks are transformed into a language problems for training. • PromptHate [1]. An encoder model that effectively utilizes pre-trained language models by employing prompts and leveraging implicit knowledge for classifying hateful memes.…”
Section: Modelsmentioning
confidence: 99%
“…These datasets serve as the foundation for training and evaluating models, enabling researchers to create more accurate and robust solutions for classifying memes across different domains. Researchers have explored different approaches, including classic two-stream models that combine text and visual features for classifying hateful memes [4,26], as well as fine-tuning large-scale pre-trained multimodal models for multimodal classification tasks [1,13,16,21,23,29]. However, a challenge arises when models and datasets are scattered across separate GitHub repositories.…”
Section: Introductionmentioning
confidence: 99%