2020
DOI: 10.1007/s41095-020-0156-x
|View full text |Cite
|
Sign up to set email alerts
|

WaterNet: An adaptive matching pipeline for segmenting water with volatile appearance

Abstract: We develop a novel network to segment water with significant appearance variation in videos. Unlike existing state-of-the-art video segmentation approaches that use a pre-trained feature recognition network and several previous frames to guide segmentation, we accommodate the object's appearance variation by considering features observed from the current frame. When dealing with segmentation of objects such as water, whose appearance is non-uniform and changing dynamically, our pipeline can produce more reliab… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
18
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 14 publications
(18 citation statements)
references
References 19 publications
0
18
0
Order By: Relevance
“…Since the webcams provided images with different resolutions, the images were resized with padding to a size of 512 × 512 × 3 making them compatible with the models. For the segmentation model training, two additional datasets were provided by Kaggle Water Segmentation Dataset (see Liang et al., 2020) as well as Flood Segmentation Dataset (Pally et al., 2022) was used. Liang et al.…”
Section: Case Studies and Datamentioning
confidence: 99%
“…Since the webcams provided images with different resolutions, the images were resized with padding to a size of 512 × 512 × 3 making them compatible with the models. For the segmentation model training, two additional datasets were provided by Kaggle Water Segmentation Dataset (see Liang et al., 2020) as well as Flood Segmentation Dataset (Pally et al., 2022) was used. Liang et al.…”
Section: Case Studies and Datamentioning
confidence: 99%
“…(Lin, Qi, and Jia 2019;Voigtlaender et al 2019;Wang et al 2019;Yang, Wei, and Yang 2020) further leverage both the first and the previous frames. Several recent methods (Hu, Huang, and Schwing 2018;Liang et al 2020a;Duke et al 2021) turn to use several latest frames to further improve the local temporal guidance. Moreover, STM-based networks (Oh et al 2019;Seong, Hyun, and Kim 2020;Lu et al 2020;Liang et al 2020b,c;Cheng, Tai, and Tang 2021b;Wang et al 2021;Xie et al 2021;Hu et al 2021;Seong et al 2021) boost the performance with memory networks that memorize information from past frames for further reuse, which relieve the error propagation to some extent.…”
Section: Related Workmentioning
confidence: 99%
“…For propagation-based models [3,23,42], the guidance of segmentation masks from past frames are introduced during the process of mask decoding. For matching-based models [5,10,17,17,26,27,30,35,49,55,63,65,67], an embedding space is learnt for target objects. Recently, STM-based networks [6, 16, 28, 32, 39, 46, 47, 52? ] achieve impressive results with memory networks that memorize and read information from past frames.…”
Section: Related Workmentioning
confidence: 99%