More about HKUST
SCALABLE COMMONSENSE KNOWLEDGE ACQUISITION AND KNOWLEDGE FUSION
The Hong Kong University of Science and Technology
Department of Computer Science and Engineering
PhD Thesis Defence
Title: "SCALABLE COMMONSENSE KNOWLEDGE ACQUISITION AND KNOWLEDGE FUSION"
By
Mr. Changlong YU
Abstract:
Understanding human languages requires the ability to reason about rich
commonsense knowledge concerning everyday concepts and events. Recent advances
have been made in leveraging linguistically pattern-based methods for automatic
commonsense knowledge acquisition. Compared with crowdsourced annotations,
these methods can significantly reduce human labeling efforts but still have
several limitations when scaling up in different scenarios.
In this thesis, we investigate ways to improve pattern-based knowledge
extraction at scale. First, we confirm the inherent low-recall issues of
pattern-based methods for hypernymy prediction tasks and propose a
complementary framework that utilizes contextualized representations to
supplement semantic information. Experimental results demonstrate the
superiority of this approach for term pairs that are not covered by patterns.
Next, we argue that patterns cannot be easily generalized across different
languages, and creating high-quality annotation benchmarks is time-consuming,
especially for low-resource languages. We explore different cross-lingual and
multilingual training paradigms and find that meta-learning can effectively
transfer knowledge from high-resource languages to low-resource ones.
Furthermore, extending general patterns to specific domains like e-commerce is
infeasible. E-commerce commonsense regarding user shopping intentions is not
explicitly stated in the products' metadata but can be mined from vast amounts
of user interaction behaviors. We propose a novel framework to distill
intention knowledge by explaining co-purchase behaviors with the help of large
language models and human-inthe- loop annotations. Intrinsic and extrinsic
evaluations demonstrate the effectiveness of our proposed framework.
After harvesting large-scale structured commonsense knowledge, how to better
incorporate it for downstream tasks becomes crucial. Considering the high-order
information stored in the knowledge graph, we propose injecting complex
commonsense knowledge obtained from random walk paths into pretrained language
models like BERT. We design advanced masking strategies and new training
objectives for effective knowledge fusion. Lastly, we revisit the evaluations
of knowledge fusion on natural language understanding tasks and find that even
fusing wrong or random knowledge can achieve comparable or better performance,
which calls for fair and faithful evaluations in the future.
Date: Thursday, 31 August 2023
Time: 10:00am - 12:00noon
Venue: Room 3494
Lifts 25/26
Chairman: Prof. Pak Wo LEUNG (PHYS)
Committee Members: Prof. Wilfred NG (Supervisor)
Prof. Yangqiu SONG
Prof. Dan XU
Prof. Jingdi ZHANG (PHYS)
Prof. Meng JIANG (University of Notre Dame)
**** ALL are Welcome ****