2017 IEEE International Conference on Computer Vision (ICCV) 2017
DOI: 10.1109/iccv.2017.71
|View full text |Cite
|
Sign up to set email alerts
|

Towards Context-Aware Interaction Recognition for Visual Relationship Detection

Abstract: The first two authors contributed equally to this work. This work was in part supported by an ARC Future Fellowship to C.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
121
0

Year Published

2018
2018
2019
2019

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 155 publications
(121 citation statements)
references
References 39 publications
(110 reference statements)
0
121
0
Order By: Relevance
“…A visual relationship is discovered (starting from the proposals coming from an object detector) with a graph traversal algorithm in a reinforcement learning setting. Context-AwareVRD [35] and WeaklySup [23] encode the features of pairs of bounding boxes similarly to Equation (9). However, in Context-AwareVRD the learning is performed with a neural network, whereas in WeaklySup the learning is based on a weakly-supervised discriminative clustering I.e., the supervision on a given relationship is not at triples level but on an image level.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…A visual relationship is discovered (starting from the proposals coming from an object detector) with a graph traversal algorithm in a reinforcement learning setting. Context-AwareVRD [35] and WeaklySup [23] encode the features of pairs of bounding boxes similarly to Equation (9). However, in Context-AwareVRD the learning is performed with a neural network, whereas in WeaklySup the learning is based on a weakly-supervised discriminative clustering I.e., the supervision on a given relationship is not at triples level but on an image level.…”
Section: Methodsmentioning
confidence: 99%
“…In [2] the background knowledge is statistical information (learnt with statistical link prediction methods [21]) about the training set triples. Contextual information between objects is used also in [23], [35] with different learning methods. In [31] the background knowledge (from the training set and Wikipedia) is a probability distribution of a relationship given the subject/object.…”
Section: Related Workmentioning
confidence: 99%
“…VRD and Visual Genome have spurred the development of new approaches for visual relationship detection. Suc-cessful methods typically build on top of an object detection module, and reason jointly over language and visual features [16,37,19,5,40,35]. Multiple independent directions have been proven fruitful, including learning features that are agnostic to object categories [34], facilitating the interaction between object features and predicate features [34,5,16], overcoming the scarcity of labeled data through weakly supervised learning [27,38], and detecting the relationships among multiple objects jointly as scene graphs [33,18,32,17].…”
Section: Related Workmentioning
confidence: 99%
“…We also report the performance of combining their predictions by a weighted average. We evaluate five state-of-the-art models: Vip-CNN [16], Peyre et al [27], PPR-FCN [40], DRNet [5] and VTransE [37]. They were created for visual relationship detection but can be adapted to our task straightforwardly: First, object detectors are replaced by ground truth objects.…”
Section: Baselines For Spatial Relation Recognitionmentioning
confidence: 99%
“…Considering the fact that visual features provide limited knowledge for distinguishing the predicates, many works focus on introducing different modal features into predicate recognition stage. For example, [40,41,45,47] prove that language prior and location information denoting categories and location of object pairs are effective to improve the performances of visual relationship recognition. However, comparing to language prior, relative location information, as strong inferring to relationships, are not fully exploited in those works.…”
Section: Related Workmentioning
confidence: 99%