We present an approach for jointly matching and segmenting object instances of the same category within a collection of images. In contrast to existing algorithms that tackle the tasks of semantic matching and object co-segmentation in isolation, our method exploits the complementary nature of the two tasks. The key insights of our method are two-fold. First, the estimated dense correspondence field from semantic matching provides supervision for object co-segmentation by enforcing consistency between the predicted masks from a pair of images. Second, the predicted object masks from object co-segmentation in turn allow us to reduce the adverse effects due to background clutters for improving semantic matching. Our model is end-to-end trainable and does not require supervision from manually annotated correspondences and object masks. We validate the efficacy of our approach on four benchmark datasets: TSS, Internet, PF-PASCAL, and PF-WILLOW, and show that our algorithm performs favorably against the state-of-the-art methods on both semantic matching and object co-segmentation tasks.Index Terms-Semantic matching, object co-segmentation, weakly-supervised learning.
!Our work. In this paper, we propose to jointly tackle both semantic matching and object co-segmentation with a twostream network in an end-to-end trainable fashion. Our key insights are two-fold. First, to suppress the effect of background clutters, the predicted object masks by object co-segmentation allow the model to focus on matching the segmented foreground regions while excluding background matching. Second, the estimated dense correspondence fields by semantic matching provide supervision for enforcing the model to generate geometrically consistent object masks across images. Therefore, we exploit the interdependency between the two network outputs, i.e., the estimated dense correspondence fields and the predicted foreground object masks, by introducing the cross-network consistency loss. Incorporating this loss improves both tasks since it encourages arXiv:1906.05857v1 [cs.CV]