More about HKUST
A Survey on Explainable AI: From Interpreting Models to Driving Model Improvements
PhD Qualifying Examination
Title: "A Survey on Explainable AI: From Interpreting Models to Driving Model
Improvements"
by
Miss Luyu QIU
Abstract:
The rapid advancement of Artificial Intelligence (AI) has had transformative
effects across a wide array of fields. However, the lack of transparency and
interpretability in AI models has created significant obstacles, particularly
in applications where trust, accountability, and regulatory compliance are
essential. This has led to the emergence of Explainable AI (XAI), a field
dedicated to enhancing the transparency of AI systems and making their
decision-making processes understandable to human users. This survey provides
a comprehensive overview of XAI, detailing its background, importance, and
categorizing key approaches to XAI methods. We analyze classic methods and
discuss their strengths and limitations, highlighting the need for solutions
that balance interpretability with predictive accuracy.
To demonstrate the depth of my research in the field of XAI, this survey will
discuss one of my published papers, which addresses the challenge of
Out-of-Distribution (OoD) data in perturbation-based XAI algorithms.
Perturbation-based methods are promising for interpreting black-box models
due to their simplicity and effectiveness, yet they often suffer from
unreliable explanations when the perturbed data diverges from the original
dataset distribution, an issue that has received limited attention.We address
this OoD problem by introducing a novel module that measures the consistency
between perturbed samples and the original dataset distribution. Our approach
penalizes the influence of unreliable OoD data, enhancing the robustness of
explanations by integrating inlier scores with target model predictions. This
method significantly improves the reliability of popular perturbation-based
XAI algorithms, such as RISE, OCCLUSION, and LIME, and notably mitigates the
performance degradation issue observed in RISE. Extensive experiments confirm
the effectiveness of our solution, yielding superior results on computational
and cognitive metrics and introducing a novel faithfulness indicator that
demonstrates robustness against the OoD problem.
Moreover, this survey highlights my ongoing research on explaining the
behavior of large language models (LLMs). Specifically, we delve into
understanding why LLMs underperform in multiplication tasks. By leveraging
explainable AI (XAI) methods, we uncover the underlying causes of this issue
and propose improvements to the Transformer model, achieving enhanced
performance.
Finally, the survey concludes with a summary of key insights and future
research directions, providing a comprehensive overview of the field and
potential pathways for advancement.
Date: Wednesday, 18 December 2024
Time: 4:00pm - 6:00pm
Venue: Room 2128A
Lift 19
Committee Members: Prof. Lei Chen (Supervisor)
Prof. Qiong Luo (Chairperson)
Dr. Xiaomin Ouyang
Prof. Qian Zhang