2022
DOI: 10.1007/s11263-022-01702-9
|View full text |Cite
|
Sign up to set email alerts
|

AOE-Net: Entities Interactions Modeling with Adaptive Attention Mechanism for Temporal Action Proposals Generation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(12 citation statements)
references
References 51 publications
0
12
0
Order By: Relevance
“…On NMS, AOE-Net obtains the best on AR@100 and the second best on AR@200 and AR@500 with very close gap with the SOTA, 57.49 vs. 57.74 and 62.40 vs. 62.74, respectively. Notably, the performance on TAPG in both datasets of our AOE-Net are a very competitive with AEI-B [52] and followed closely by ABN [6], both of which also incorporate local actors and global environment. This experiment strongly supports our observation and motivation on using the human perception principle to analyze human actions in untrimmed videos.…”
Section: Methodsmentioning
confidence: 85%
See 1 more Smart Citation
“…On NMS, AOE-Net obtains the best on AR@100 and the second best on AR@200 and AR@500 with very close gap with the SOTA, 57.49 vs. 57.74 and 62.40 vs. 62.74, respectively. Notably, the performance on TAPG in both datasets of our AOE-Net are a very competitive with AEI-B [52] and followed closely by ABN [6], both of which also incorporate local actors and global environment. This experiment strongly supports our observation and motivation on using the human perception principle to analyze human actions in untrimmed videos.…”
Section: Methodsmentioning
confidence: 85%
“…In our AOE-Net, BMM module is adopted from previous works i.e. BSN [1], BMN [3], ABN [7], AEN [6], AEI [52] because of its standard and simple design. BMM takes the output V-L features sequence F = {f i } T i=1 from PMR module as its input.…”
Section: Boundary-matching Module (Bmm)mentioning
confidence: 99%
“…Top k similarity scores are chosen as language-based frame embedding feature F l i . (iii) -language feature extraction: In this step, we employ Adaptive Attention Mechanism (AAM) [3] to select the most relevant representative language features:…”
Section: Proposed Methodsmentioning
confidence: 99%
“…Strong supervision gives precise boundary labels and category labels for training. There are two detailed pipelines: the top-down framework [65,63,12,7,35,73,67,98,70,75] pre-defines extensive anchors, adopts fixed-length sliding windows to produce initial proposals, then regresses to refine boundaries; the bottom-up framework [92,36,34,68,90,1] learns frame-wise boundary detectors for the boundary frames, then groups extreme frames or estimates action lengths for proposal generation. In addition, several works [10,39,78] used various fusion strategies to complement these frameworks.…”
Section: Related Workmentioning
confidence: 99%