More about HKUST
Effective Optimization Algorithms for Multi-task Learning with Conflicting Tasks
PhD Thesis Proposal Defence
Title: "Effective Optimization Algorithms for Multi-task Learning with
Conflicting Tasks"
by
Mr. Hansi YANG
Abstract:
Real-world applications of machine learning often involve multiple tasks
that may conflict with each other, resulting in longer training time and
often unsatisfying performance. In this thesis proposal, we consider
developing optimization algorithms specifically designed to manage task
conflicts and enhance learning outcomes. We introduce a cohesive framework
comprising three innovative strategies aimed at resolving high gradient
variance, mitigating inaccurate labels, and balancing multi-task learning
conflicts.
First, we focus on variance reduction techniques for few-shot learning,
where limited samples lead to significant gradient variance across different
tasks. Our momentum-based variance reduction methods are designed to deliver
precise gradient estimates, resulting in faster convergence and improved
generalization in few-shot scenarios. These improvements are substantiated
through rigorous theoretical analysis and empirical validation.
Next, we explore bi-level learning with cubic regularization to tackle the
challenges posed by label noise. Label noise can cause overfitting and
impair model performance. Our approach leverages bi-level learning to
provide flexible control over the learning process, using cubic
regularization to address complex curvatures in bi-level optimization
problems. This method stabilizes training dynamics and enhances robustness
against mislabeled data.
Lastly, we propose a gradient balancing framework for multi-task learning,
where conflicts between tasks and samples are common. Our method dynamically
adjusts the sample order during optimization to ensure fair representation
of all tasks, facilitating effective learning across diverse tasks even when
individual datasets are limited.
Collectively, these strategies address fundamental optimization challenges
in machine learning with conflicting tasks. By reducing variance in few-shot
learning, enhancing robustness to inaccurate labels, and ensuring balanced
multi-task learning, this work lays a robust foundation for developing
effective optimization algorithms. Our aim is to provide both theoretical
insights and practical solutions that advance the state-of-the-art in
machine learning amidst conflicting task scenarios.
Date: Wednesday, 18 June 2025
Time: 9:30am - 11:30am
Venue: Room 3494
Lifts 25/26
Committee Members: Prof. James Kwok (Supervisor)
Prof. Raymond Wong (Chairperson)
Dr. Dan Xu