2022
DOI: 10.1007/978-3-031-19830-4_30
|View full text |Cite
|
Sign up to set email alerts
|

Mining Relations Among Cross-Frame Affinities for Video Semantic Segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 18 publications
(2 citation statements)
references
References 52 publications
0
2
0
Order By: Relevance
“…Given they can't provide the instance mask for each building, the prediction results also need to be post-processed. 3) video semantic segmentation methods: CFFM [13] and MRCFA [14] are adopted since they can process the multi-temporal image series which is similar to the video frames. To ensure fair comparisons, we employ the same post-processing algorithm in conjunction with the compared methods.…”
Section: Compared With the State-of-the-artmentioning
confidence: 99%
“…Given they can't provide the instance mask for each building, the prediction results also need to be post-processed. 3) video semantic segmentation methods: CFFM [13] and MRCFA [14] are adopted since they can process the multi-temporal image series which is similar to the video frames. To ensure fair comparisons, we employ the same post-processing algorithm in conjunction with the compared methods.…”
Section: Compared With the State-of-the-artmentioning
confidence: 99%
“…The work of [53] has devised a spatio‐temporal continuity module to capture the spatial and temporal relations inherent in videos to improve the segmentation performance. In [61], better temporal information aggregation in videos has been achieved by mining relations among cross‐frame affinities instead of developing new techniques to calculate the cross‐frame affinities such as optical flow and attention. To reduce the computation cost, some works apply a strong segmentation network only at predefined key frames, and propagate to non‐key frames using optical flows to reuse the better features in preceding frames [39, 43–45].…”
Section: Related Workmentioning
confidence: 99%