More about HKUST
A Survey on Advancements and Challenges of Model-based Reinforcement Learning
PhD Qualifying Examination
Title: "A Survey on Advancements and Challenges of Model-based Reinforcement
Learning"
by
Mr. Hon Hing CHAK
Abstract:
Model-based reinforcement learning is a sub-field in reinforcement learning. It
studies the problem of optimal control, a sequential decision-making problem
that maximizes an objective value based on the decisions acted on an
environment. This sub-field describes a class of heuristics that learns a model
(called a world model) of the environment, and a second model (called an agent)
that produces decisions with high objective values. The agent utilizes the
world model during its learning process.
In the literature, the control problem, or namely the reinforcement learning
problem, is typically described under a mathematical framework, called the
Markov Decision Process (MDP). MDP specifies how the environment and the
objective value change in response to decisions. Many problems with an
interactive nature can be framed as MDPs, such as games, robotic control, stock
trading, etc.. Typical reinforcement learning algorithms without environment
modeling, i.e., model-free reinforcement learning, can only be effective for
some problems in practice. If interacting with the environment is costly or
unsafe, such as in self-driving or medical applications, model-free algorithms
would not be suitable. This is because model-free algorithms rely on large
amounts of interactive data to learn agents with high performances.
Recent studies have explored new methodologies in model-based reinforcement
learning. Utilizing the recent advancement of deep learning, current
model-based methods could create world models that generate high-quality
simulated interactions in some environments, which are crucial for training
high-performing agents. For example, modern model-based methods use large
language models to implement the world models, which have achieved great
results in the past years. In this survey, we first review different types of
reinforcement learning problems and their challenges to modern reinforcement
learning techniques. We then study some recent approaches in the model-based
reinforcement learning field and their results in common baseline reinforcement
learning problems in the field. Finally, we conclude this survey by
highlighting weaknesses that are still present in current model-based
methodologies and how we might address them in future research.
Date: Friday, 28 June 2024
Time: 2:00pm - 4:00pm
Venue: Room 4472
Lifts 25/26
Committee Members: Prof. Raymond Wong (Supervisor)
Prof. Dit-Yan Yeung (Chairperson)
Prof. Gary Chan
Prof. Shing-Chi Cheung