More about HKUST
Privacy-Preserved Learning on Graph-Structured Data
PhD Thesis Proposal Defence Title: "Privacy-Preserved Learning on Graph-Structured Data" by Mr. Qi HU Abstract: Graph-structured data and graph learning techniques, particularly graph neural networks (GNN), have revolutionized numerous applications in recent years. These advancements enable sophisticated knowledge extraction from complex graph data, unlocking unprecedented opportunities for discovery and decision-making. However, the growing accessibility of graph data and the increasing complexity of graph learning models have raised significant privacy concerns that remain inadequately addressed despite ongoing research efforts. This thesis investigates privacy preservation in graph learning models, focusing on developing effective and robust approaches to mitigate these concerns. The first section examines privacy preservation in traditional graph learning models, specifically addressing graph embedding tasks. While graph embeddings, low-dimensional representations of graph-structured data, facilitate various downstream applications such as node classification and link prediction, they are vulnerable to attribute inference attacks. Existing privacy-preserving methods often rely on adversarial learning techniques and assume full access to sensitive attributes during training. However, this assumption is problematic given the diverse privacy preferences, and the adversarial learning approach often suffers from training instability. To address these limitations, we introduce a novel approach that incorporates the independent distribution penalty as a regularization term. The second section explores privacy preservation in distributed learning environments. Although federated learning protects raw data by design, we show that learned embeddings remain vulnerable to privacy leakage attacks. Moreover, traditional privacy preservation approaches often fail to account for varying privacy preferences among participants, resulting in suboptimal trade-offs between utility and privacy. To address this, we propose a flexible framework that adaptively accommodates diverse privacy requirements while minimizing utility loss. The final section addresses privacy challenges in neural graph databases (NGDBs), which integrate traditional graph databases with neural networks and can be integrated with large language models. Although NGDBs offer powerful capabilities for handling incomplete data through neural embedding storage and complex query answering (CQA), their ability to reveal hidden relationships introduces additional privacy vulnerabilities. Specifically, malicious actors may exploit carefully crafted queries to extract sensitive information. To mitigate these risks, we develop a privacy-preserved NGDB framework that mitigates these risks by increasing the complexity of inferring sensitive information through query combinations. Date: Friday, 24 January 2025 Time: 10:00am - 12:00noon Venue: Room 3494 Lifts 25/26 Committee Members: Dr. Yangqiu Song (Supervisor) Dr. Wei Wang (Chairperson) Dr. Dongdong She