More about HKUST
Side Channel Analysis for AI Infrastructures
The Hong Kong University of Science and Technology Department of Computer Science and Engineering PhD Thesis Defence Title: "Side Channel Analysis for AI Infrastructures" By Mr. Yuanyuan YUAN Abstract: Side channel analysis (SCA) investigates the unintended secret leakage in a system's non-functional characteristics such as execution time or memory access patterns. Given the growing adoption of AI systems in security-critical and privacy-preserving applications, this thesis comprehensively studies side channel leakages in infrastructures that underpin the entire life cycle of modern AI systems, including data processing libraries, trusted execution environments (TEEs), runtime interpreters, executables on edge devices, etc. It also proposes highly practical solutions for SCA. On the offensive side, this thesis demonstrates end-to-end attacks that recover user's inputs (i.e., the user's privacy) and the underlying neural networks (i.e., the intellectual property) of AI systems from various side channels. We first consider a non-privileged co-process as the attacker and exploit cache side channels. Modern AI systems adopt data processing libraries (e.g., Libjpeg, FFmpeg) to handle various formats of user inputs like images and audios. Our work for the first time recovers such complex inputs from these libraries' cache side channels. Then, we consider untrusted hosts as adversaries. While TEEs are widely employed to shield AI systems from malicious host platforms, our works exploit the ciphertext side channels in TEEs and reconstruct both inputs and neural networks from TEE-shielded AI systems, breaking the security belief of TEEs. We also propose the first unified recovery scheme for complex input data like images, audios, text, videos, etc., and the first practical reconstruction technique for model weights of real-world neural networks. On the defensive side, this thesis localizes hundreds of new leakage sources in the exploited systems. Prior works often adopt program analysis techniques to model leakage patterns, whose localization is leakage-specific and suffers from the scalability issue. We recast the information leakage as a cooperative game among all leakage sources and reduce the cost from exponential to nearly constant. Our work presents a generic localization pipeline and supports analyzing production-size programs and AI infrastructures for the first time. We systematically examine the leakage in AI runtime interpreters including TensorFlow and PyTorch, and study how their different computation paradigms affect the leakage. Our analysis also reveals that optimizations in AI compilers (e.g., TVM, Glow) enlarge the leakage in compiled neural network executables. Date: Friday, 30 August 2024 Time: 9:00am - 11:00am Venue: Room 3494 Lifts 25/26 Chairman: Dr. Briana CHANG (FINA) Committee Members: Dr. Shuai WANG (Supervisor) Dr. Wei WANG Prof. Nevin ZHANG Dr. Zili MENG (ECE) Dr. Kexin PEI (UChicago)