Proceedings of the 2016 ACM International Conference on the Theory of Information Retrieval 2016
DOI: 10.1145/2970398.2970414
|View full text |Cite
|
Sign up to set email alerts
|

Understanding the Message of Images with Knowledge Base Traversals

Abstract: The message of news articles is often supported by the pointed use of iconic images. These images together with their captions encourage emotional involvement of the reader. Current algorithms for understanding the semantics of news articles focus on its text, often ignoring the image. On the other side, works that target the semantics of images, mostly focus on recognizing and enumerating the objects that appear in the image. In this work, we explore the problem from another perspective: Can we devise algorit… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
2
1
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 36 publications
0
5
0
Order By: Relevance
“…TagMe is a service for identifying short phrases ( spots ) in a text that can be linked to a pertinent Wikipedia page 78 . TagMe is used for text contextualization and understanding applications in English, German, and Italian 79‐82 . The method augments a plain text by identifying “anchors,” that is, portions of the text that point to Wikipedia pages related to their meanings.…”
Section: Methodsmentioning
confidence: 99%
“…TagMe is a service for identifying short phrases ( spots ) in a text that can be linked to a pertinent Wikipedia page 78 . TagMe is used for text contextualization and understanding applications in English, German, and Italian 79‐82 . The method augments a plain text by identifying “anchors,” that is, portions of the text that point to Wikipedia pages related to their meanings.…”
Section: Methodsmentioning
confidence: 99%
“…To the best of our knowledge there is no public dataset providing gist annotations for image-caption pairs. 5 Consequently, we make use of our experiments of the dataset for understanding the message of images covering the topic of non-literal and literal image-caption pairs introduced in our previous work [83].…”
Section: Methodsmentioning
confidence: 99%
“…The main idea of the approach is to represent the images and their affiliated text by concepts from a knowledge base and conduct a query-specific knowledge graph generation, which in the case of [3] represents the gist and which in our case is used to create a mapping between query image and target class. In the knowledge base representation of Wikipedia we are using, article pages and categories are the concepts.…”
Section: Methodsmentioning
confidence: 99%
“…Ranking the Nodes. To conduct the ranking of concepts best representing the image-text pair, we use 16 different features, e.g., graph connectivity and content based measures, boolean and features based on the article texts of Wikipedia concepts (for further details, please refer to [3]). As there is no gold standard about the relevance of the so far collected concepts given (recall: a concept in the knowledge base has an equivalent node in the knowledge graph, but as we are not doing a graph ranking, we switch back the notation of concept), we rely on a pre-trained model using the data and gold standard of [3]to rank the concepts.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation