2014 IEEE Conference on Computer Vision and Pattern Recognition 2014
DOI: 10.1109/cvpr.2014.172
|View full text |Cite
|
Sign up to set email alerts
|

A Compositional Model for Low-Dimensional Image Set Representation

Abstract: Learning a low-dimensional representation of images is useful for various applications in graphics and computer vision. Existing solutions either require manually specified landmarks for corresponding points in the images, or are restricted to specific objects or shape deformations. This paper alleviates these limitations by imposing a specific model for generating images; the nested composition of color, shape, and appearance. We show that each component can be approximated by a low-dimensional subspace when … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
17
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 21 publications
(17 citation statements)
references
References 25 publications
0
17
0
Order By: Relevance
“…affine). RASL [31], Collection Flow [21] and Mobahi et al [29] first estimate a low-rank subspace of the image collection, and then perform joint alignment among images projected onto the subspace. FlowWeb [40] builds a fully-connected graph for the image collection with images as nodes and pairwise flow fields as edges, and establishes globally-consistent dense correspondences by maximizing the cycle consistency among all edges.…”
Section: Related Workmentioning
confidence: 99%
“…affine). RASL [31], Collection Flow [21] and Mobahi et al [29] first estimate a low-rank subspace of the image collection, and then perform joint alignment among images projected onto the subspace. FlowWeb [40] builds a fully-connected graph for the image collection with images as nodes and pairwise flow fields as edges, and establishes globally-consistent dense correspondences by maximizing the cycle consistency among all edges.…”
Section: Related Workmentioning
confidence: 99%
“…In addition, Learned-Miller [21] generalises the dense correspondences between image pairs to an arbitrary number of images by continuously warping each image via a parametric transformation. RSA [37], Collection Flow [18] and Mobahi et al [29] project a collection of images into a lower dimensional subspace and perform a joint alignment among the projected images. AnchorNet [34] learns semantically meaningful parts across categories, although is trained with image labels.…”
Section: Related Workmentioning
confidence: 99%
“…This parameterization allows us to describe a wide variety of motion (or warp) fields. Generic "smooth" warp fields can be described by using a truncated Discrete Cosine Transform (DCT) basis as M. However, more compressed motion bases can also be used [42], [43]. In this work, results are shown using an affine transformation parametrized with a fourdimensional θ that captures rotation, shear, and scaling.…”
Section: Dynamic Modelmentioning
confidence: 99%