More about HKUST
A Survey of Fast Convex Optimization Methods in Machine Learning
PhD Qualifying Examination
Title: "A Survey of Fast Convex Optimization Methods in Machine Learning"
by
Mr. Wenliang Zhong
Abstract:
With the development of Internat and storage technics, large datasets
become more and more popular in machine learning research. How to e
ciently solve convex optimization problem with these datasets is an
important topic, which attracts many researchers' interest. Traditional
Gradient methods, though highly scalable and easy to implement, are known
to converge slowly. Some more sophisticated algorithms, like Newton
method, can converges fast w.r.t number of iteration. How- ever, it is
impractical to compute or save Hessian matrix even for one iteration when
the data is of millions dimensions. To overcome these obstacles, several
fast convex optimization methods have been pro- posed recently. This paper
gives an a general introduction to these algorithms and a review of the
literature. Specially, both deterministic and stochastic, normal and
accelerated gradient decent methods are presented. Another fast
optimization style, called coordinate decent, is also included. These
algorithm frameworks cover a wide range of convex optimization problems in
machine learning, e.g. SVM, logistic regression, LASSO, elastic net
regression, convex multi-tasks learning, etc. Moreover, brie y comparison,
convergence rate analysis, applica- tion examples and some empirical
evidence are also provided.
Date: Friday, 7 January 2011
Time: 10:00am - 12:00noon
Venue: Room 3501
lifts 25/26
Committee Members: Dr. James Kwok (Supervisor)
Prof. Dit-Yan Yeung (Chairperson)
Dr. Raymond Wong
Prof. Nevin Zhang
**** ALL are Welcome ****