More about HKUST
Commonsense Knowledge Base Population and Reasoning for Inferential Knowledge
The Hong Kong University of Science and Technology
Department of Computer Science and Engineering
PhD Thesis Defence
Title: "Commonsense Knowledge Base Population and Reasoning for Inferential
Knowledge"
By
Mr. Tianqing FANG
Abstract:
Commonsense knowledge includes facts about the everyday world that ordinary
people are expected to know. It plays a crucial role in natural language
processing (NLP) systems, enabling them to make presumptions about common
situations encountered by humans. However, acquiring and incorporating
commonsense knowledge into NLP systems poses challenges, as such knowledge is
typically implicit and not readily available in standard corpora.
To tackle the data scarcity issue, a standard way to study commonsense is to
construct commonsense knowledge bases (CSKBs). Previous attempts have focused
on (1) human annotation, which is expensive and has limited scalability; (2)
information extraction, which suffers from relatively poor quality and
reporting bias; or (3) text generation from Large Language Models (LLMs), which
suffers from selection bias and limited novelty of generated knowledge.
Moreover, the power of LLMs to elicit commonsense knowledge also requires
fine-tuning on large-scale corpora and human- annotated commonsense data in the
first place.
We propose an alternative commonsense knowledge acquisition framework, called
Commonsense Knowledge Base Population (CKBP), which automatically populates
complex commonsense knowledge from more affordable linguistic knowledge
resources. We establish a benchmark for CKBP based on event-event discourse
relations extracted through semantic and discourse parsing of large corpora,
and we manually annotate 60K populated triples for verification.
To carry out the population process, we introduce a Graph Neural Network (GNN)-
based model that leverages the rich contextual information in the knowledge
graph as additional supervision signals. Since CKBP is a semi-supervised
learning problem with a large amount of unlabeled data (discourse knowledge
from large corpora), we also propose a pseudo-labeling-based model that
achieves excellent performance. We evaluate the effectiveness of the populated
knowledge on downstream commonsense reasoning tasks and observe that it
enhances generative commonsense inference and commonsense question answering by
providing more diverse knowledge.
Furthermore, with the knowledge at hand, we explore commonsense reasoning based
on commonsense knowledge from two perspectives. First, we directly utilize the
populated knowledge for downstream commonsense question answering by converting
it into question-answering (QA) form with templates, serving as supervision
data for training QA models and generative commonsense inference models.
Second, we perform reasoning on complex logical queries derived from
commonsense knowledge graphs. We sample conjunctive logical queries from the
knowledge graphs and verbalize them using LLMs to generate narratives for both
training and evaluating models for complex reasoning. Experimental results
demonstrate that while LLMs exhibit proficiency in handling one-hop commonsense
knowledge, performing complex reasoning involving multiple hops and
intersections on commonsense knowledge graphs remains challenging. Models
trained on complex logical queries show improvement in terms of general
narrative understanding and complex commonsense reasoning ability.
Date: Thursday, 8 August 2024
Time: 4:00pm - 6:00pm
Venue: Room 5510
Lifts 25/26
Chairman: Prof. Irene Man Chi LO (CIVL)
Committee Members: Dr. Yangqiu SONG (Supervisor)
Dr. Junxian HE
Dr. Brian MAK
Dr. Jing WANG (ISOM)
Dr. Wen HUA (PolyU)