2019 IEEE International Conference on Multimedia and Expo (ICME) 2019
DOI: 10.1109/icme.2019.00094
|View full text |Cite
|
Sign up to set email alerts
|

PANet: A Context Based Predicate Association Network for Scene Graph Generation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(5 citation statements)
references
References 18 publications
0
5
0
Order By: Relevance
“…It is composed of two general parts: the local propagation within triplet items [26] and the global propagation among all the elements. The global items, according to its specific prior layout structure, can be further divided into the following forms: fully-connected graph [59], [77], chain [73], [74] and tree [78], [79]. contextual information.…”
Section: Message Passing Mechanismmentioning
confidence: 99%
See 1 more Smart Citation
“…It is composed of two general parts: the local propagation within triplet items [26] and the global propagation among all the elements. The global items, according to its specific prior layout structure, can be further divided into the following forms: fully-connected graph [59], [77], chain [73], [74] and tree [78], [79]. contextual information.…”
Section: Message Passing Mechanismmentioning
confidence: 99%
“…Many other message passing methods based on RNN have been developed. Chen et al [74] used an RNN module to capture instance-level context, including objects co-occurrence, spatial location dependency and label relation. Dai et al [76] used a Bi-directional RNN and Shin et al [73] used Bi-directional LSTM as a replacement.…”
Section: Message Passing Mechanismmentioning
confidence: 99%
“…Currently, SGG can be divided into two classes: 1) with facts alone and 2) introducing prior information. Besides, these SGG methods pay more attention to the methods with facts alone, including CRF-based (conditional random field) SGG [117,120] , VTransE-based (visual translation embedding) SGG [121,122] , RNN/LSTM-based SGG [123,124] , Faster RCNN-based SGG [125,126] , graph neural network (GNN) [127,128] , etc. Furthermore, SGG adds different types of prior information, such as language priors [129] , knowledge priors [130,131] , visual contextual information [132] , visual cue [133] , etc.…”
Section: Visual Representation Learning: Stateof-the-artmentioning
confidence: 99%
“…SPP performs three maximum pooling operations of different sizes on the input, and concats the output results to splicing and splicing the output results, so that the output depth of the network is the same as the input depth. The third part is the feature fusion network Neck, which consists of Feature Pyramid Networks (FPN) [24] and Path Aggregation Networks (PAN) [25] structures. The FPN structure combines the category information of high-level large targets To the lower layer, the PAN structure transfers the location information of the low-level large target and the category and location features of the small target upwards.…”
Section: The Basic Principle Of Yolov5 Algorithmmentioning
confidence: 99%