More about HKUST
Scalable Convolutional Sparse Coding
PhD Thesis Proposal Defence
Title: "Scalable Convolutional Sparse Coding"
by
Miss Yaqing WANG
Abstract:
Convolutional sparse coding (CSC) improves sparse coding by learning a
shift-invariant dictionary from the data. It has been successfully extracting
local patterns from various data types, such as trajectories, images, audios,
videos, multi-spectral and light field images and biomedical data. However,
most existing CSC algorithms operate in the batch mode and are computationally
expensive. This lack of scalability restricts the use of CSC on large-scale
data.
First, we propose a scalable online CSC algorithm for data sets of large
quantity. The algorithm, which will be called Online Convolutional Sparse
Coding (OCSC). The key is a reformulation of the CSC objective so that
convolution can be handled easily in the frequency domain, and much smaller
space is needed. To solve the resultant optimization problem, we use the
alternating direction method of multipliers (ADMM), and its subproblems have
efficient closed-form solutions. Theoretical analysis shows that the learned
dictionary converges to a stationary point of the optimization problem. Results
on large image data sets such as ImageNet, Flower and CIFAR-10 show that the
proposed algorithm outperforms state-of-the-art batch and online CSC methods.
It is more scalable, has faster convergence and better reconstruction
performance.
Then, we enable CSC to learn with a large set of filters. Instead of convolving
with a dictionary shared by all samples, we propose the use of a
sample-dependent dictionary in which each filter is a linear combination of a
small set of base filters learned from data. This added flexibility allows a
large number of sample-dependent patterns to be captured, which is especially
useful in the handling of large or high-dimensional data sets. Computationally,
the resultant model can be efficiently learned by online learning.
Specifically, the base filter can still be updated by OCSC, while the codes and
combination weights can be learned by accelerated nonconvex proximal
algorithms. Extensive experimental results on multiple kinds of data sets such
as images, multispectral images, videos, and light field images show that the
proposed methods outperform existing CSC algorithms with significantly reduced
time and space complexities.
Finally, we propose some further improvement. For example, extending CSC to
deal with unknown noise, designing multiscale filters, introducing stochastic
version, and extending the proposed sample-dependent filters to convolutional
neural networks and graph convolutional networks.
Date: Monday, 12 November 2018
Time: 4:30pm - 6:30pm
Venue: Room 5501
(lifts 25/26)
Committee Members: Prof. Lionel Ni (Supervisor)
Prof. James Kwok (Supervisor)
Dr. Qiong Luo (Chairperson)
Prof. Nevin Zhang
**** ALL are Welcome ****