2019
DOI: 10.1007/978-3-030-11012-3_52
|View full text |Cite
|
Sign up to set email alerts
|

How to Read Paintings: Semantic Art Understanding with Multi-modal Retrieval

Abstract: Automatic art analysis has been mostly focused on classifying artworks into different artistic styles. However, understanding an artistic representation involves more complex processes, such as identifying the elements in the scene or recognizing author influences. We present SemArt, a multi-modal dataset for semantic art understanding. SemArt is a collection of fine-art painting images in which each image is associated to a number of attributes and a textual artistic comment, such as those that appear in art … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
101
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 77 publications
(102 citation statements)
references
References 22 publications
1
101
0
Order By: Relevance
“…The two proposed context-aware embeddings are evaluated on the SemArt dataset [17] in four different art classification tasks and in two cross-modal retrieval tasks. We show that, although none of the proposed models show a superior performance with respect to the other one in all of the evaluated tasks, context-aware embeddings consistently outperform methods based on visual embeddings only.…”
Section: Task-specific Embeddingsmentioning
confidence: 99%
See 2 more Smart Citations
“…The two proposed context-aware embeddings are evaluated on the SemArt dataset [17] in four different art classification tasks and in two cross-modal retrieval tasks. We show that, although none of the proposed models show a superior performance with respect to the other one in all of the evaluated tasks, context-aware embeddings consistently outperform methods based on visual embeddings only.…”
Section: Task-specific Embeddingsmentioning
confidence: 99%
“…For example, [21] proposed to detect authors by analysing their brushwork using wavelet decompositions, [24,41] combined color, edge, or texture features for author, style, and school classification and [5,32] used SIFT features [28] to classify paintings into different attributes. In the last years, deep visual features extracted from CNNs have been repeatedly shown to be very effective in many computer vision tasks, including automatic art analysis [2,17,23,29,30,37,44,45]. At first, deep features were extracted from pre-trained networks and used off-the-shelf for automatic art classification [2,23,37].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Detection of people in artworks is implemented in [10] and weakly supervised detection of objects in paintings is investigated in [11]. Multimodal retrieval accross visual and textual representations of paintings is discussed by [12]. The demand for artrelated visual recognition algorithms is best demonstrated by the challenge of [13].…”
Section: Prior Workmentioning
confidence: 99%
“…Recent research on the semantic understanding of paintings in art, however, shows that it is possible to apply transfer learning of vision to artworks while training new models. For example, Garcia and Vogiatzis [16] considered the Text2Art challenge where they successfully retrieved the related artwork from the set of test images 45.5% (Top-10) of times. They also showed a technique to store visual and textual information of the same artwork in the same semantic space, thereby making the task of retrieval much easier.…”
Section: Related Workmentioning
confidence: 99%