2016
DOI: 10.1109/tpami.2015.2511748
|View full text |Cite
|
Sign up to set email alerts
|

Dissimilarity-Based Sparse Subset Selection

Abstract: Finding an informative subset of a large collection of data points or models is at the center of many problems in computer vision, recommender systems, bio/health informatics as well as image and natural language processing. Given pairwise dissimilarities between the elements of a 'source set' and a 'target set,' we consider the problem of finding a subset of the source set, called representatives or exemplars, that can efficiently describe the target set. We formulate the problem as a row-sparsity regularized… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
105
0
4

Year Published

2016
2016
2020
2020

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 152 publications
(109 citation statements)
references
References 58 publications
0
105
0
4
Order By: Relevance
“…In this paper, we model dictionary learning as a subsect selection problem which seeks a small set of representatives to summarize and describe the whole dataset X , thereby removing outliers that are not true representatives and improving the runtime computational efficiency. A straightforward way to conduct sparse subsect selection is to regularize the coefficient matrix with an ℓ 2,1 or ℓ ∞, 1 norm [6]. Meanwhile, since the number of constraints for shape control is limited, cell shapes actually lie on a low-dimensional manifold [16].…”
Section: Methodsmentioning
confidence: 99%
“…In this paper, we model dictionary learning as a subsect selection problem which seeks a small set of representatives to summarize and describe the whole dataset X , thereby removing outliers that are not true representatives and improving the runtime computational efficiency. A straightforward way to conduct sparse subsect selection is to regularize the coefficient matrix with an ℓ 2,1 or ℓ ∞, 1 norm [6]. Meanwhile, since the number of constraints for shape control is limited, cell shapes actually lie on a low-dimensional manifold [16].…”
Section: Methodsmentioning
confidence: 99%
“…For each image in Scene and Pascal, we take 90 and 150 descriptors as the input of each aggregation approach, respectively. Our ProLFA model in (9) is then employed to train the aggregation function on the training set, thus obtaining the global representations of testing set via (22). Finally, to evaluate the discrimination of aggregated representations, we choose the 1-Nearest Neighbor (1-NN) classifier with Euclidean distance since it is parameter free and the results will be easily reproducible.…”
Section: Evaluation By Image Searchmentioning
confidence: 99%
“…Additionally, to improve the interpretability of codebook, we consider the process of finding codebook as the task of Prototype Selection (PS), which aims at finding exemplar samples from a feature collection. PS has been actively discussed in other fields [22], such as video summarization and product recommendation, since it holds several advantages over data storage, compression, synthesis and cleansing. Besides helping to reduce the computational time and memory of algorithms, due to working on several prototypical samples, PS has further improved performances of numerous applications.…”
Section: Introductionmentioning
confidence: 99%
“…Other efforts have been devoted to finding the representatives by using pairwise similarities between data points (Frey & Dueck, 2007;Elhamifar et al, 2016). For example, Elhamifar et al (2016) proposed a dissimilarity-based sparse subset selection (DS3) method to minimize the difference between source data and target data. However, some methods, e.g.…”
Section: Related Workmentioning
confidence: 99%