2018
DOI: 10.1109/tip.2018.2842152
|View full text |Cite
|
Sign up to set email alerts
|

Scalable Online Convolutional Sparse Coding

Abstract: Convolutional sparse coding (CSC) improves sparse coding by learning a shift-invariant dictionary from the data. However, most existing CSC algorithms operate in the batch mode and are computationally expensive. In this paper, we alleviate this problem by online learning. The key is a reformulation of the CSC objective so that convolution can be handled easily in the frequency domain, and much smaller history matrices are needed. To solve the resultant optimization problem, we use the alternating direction met… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
44
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 42 publications
(44 citation statements)
references
References 29 publications
0
44
0
Order By: Relevance
“…Another aspect of this advantage is our ability to run in an online manner, even for a single input image. This stands in sharp contrast to other recent online methods [30,31] which allow for online training but only in the case of streaming images. Other approaches took a step further and proposed partitioning the image into smaller sub-images [32], but this is still far from [32] and the SBDL algorithm [1].…”
Section: Relation To Other Methodsmentioning
confidence: 67%
“…Another aspect of this advantage is our ability to run in an online manner, even for a single input image. This stands in sharp contrast to other recent online methods [30,31] which allow for online training but only in the case of streaming images. Other approaches took a step further and proposed partitioning the image into smaller sub-images [32], but this is still far from [32] and the SBDL algorithm [1].…”
Section: Relation To Other Methodsmentioning
confidence: 67%
“…Let η This can be solved by closed form solution. Using the reordering trick used in [21], [37], we first put all d k 's in the columns ofD = [d 1 , . .…”
Section: A Filter Updatementioning
confidence: 99%
“…Moreover, inferring z i from x is an iterative process that may require a large number of iterations, thus, less suitable for real-time tasks. Work on speeding up the ADMM method for CD learning is done in [13]. In [14], a consensus-based optimization framework has been proposed that makes CSC tractable on large-scale datasets.…”
Section: Convolutional Sparse Codingmentioning
confidence: 99%