More about HKUST
A Survey on Knowledge-driven Autonomous Driving
PhD Thesis Proposal Defense Title: "A Survey on Knowledge-driven Autonomous Driving" by Mr. Haoran LI Abstract: The rise of transformer models has significantly advanced machine learning models. Large language models (LLMs), trained on massive data and supported by extensive computational resources, unify the conventional Natural Language Processing (NLP) paradigm and can effectively handle a variety of downstream tasks by integrating these tasks into generative workflows. For real-world impacts, LLMs have revolutionized accessibility and usability for researchers, developers and users. Consequently, powerful LLMs have empowered new possibilities across various fields, including agents, smart assistants, chatbots, and more. However, the widespread availability and accessibility of these models also introduce potential risks, including malicious use and privacy concerns. The free-form generation pipelines that make LLMs valuable can also be misused to compromise privacy or for harmful purposes. Even though heavy efforts have already been made to enhance LLMs' trustworthiness to target LLMs' safety and privacy issues, new attacks are frequently proposed to bypass existing defense mechanisms and comprise LLMs for malicious uses. Therefore, a consistent and adversarial game exists between malicious attackers and defenders regarding LLMs' trustworthiness, leaving significant challenges undiscovered. To fully study LLMs' trustworthiness issues, we identify new attacks in terms of information leakage, improve defense mechanisms to address various attacks and empirically evaluate the attacks' outcomes with and without defenses. For identified attacks, we target information leakage issues on vector databases to study embeddings' information leakage. Beyond information leakage on embeddings, we also demonstrate how to compromise LLMs with jailbreaking prompts. After discussing attacks, we present new defenses to prevent information leakage on embeddings. Lastly, we implement a benchmark to evaluate attacks' performance with and without defenses empirically. Extensive experiments are conducted to support the effectiveness of our discovered attacks and defenses. Our evaluation benchmark's results reveal the gap between attacks' and defenses' assumptions. Date: Tuesday, 16 July 2024 Time: 10:00am - 12:00noon Venue: Room 5510 Lifts 25/26 Committee Members: Dr. Yangqiu Song (Supervisor) Dr. Binhang Yuan (Chairperson) Dr. Junxian He Dr. Dongdong She