More about HKUST
A Survey on Efficient Transformers: from Training to Inference
PhD Qualifying Examination Title: "A Survey on Efficient Transformers: from Training to Inference" by Mr. Shih-yang LIU Abstract: Transformer-based models have shown remarkable improvements over earlier techniques and ushered in significant progress in many tasks, especially in the domains of vision and language. However, as their size and complexity continue to increase, latency and model size have become a noticeable issues in both training and deployment. To tackle these challenges, researchers have begun investigating methods to reduce training overhead—often referred to as Parameter Efficient Fine-Tuning (PEFT)—which adjust only a subset of the model parameters, as well as compression strategies that help shrink models for quicker inference. However, many existing studies regard these two strategies as independent, missing a chance to leverage the advantages of both line works to achieve simultaneous gains in both training and inference efficiency. In this context, this summary provides a comprehensive review of both PEFT and model compression, showcasing various examples and examining how their integration can enhance the efficiency of transformer architectures, particularly in the realm of Large Language Models (LLMs). Notably, this work is among the first broad surveys in the efficient transformers domain that encompasses both PEFT and model compression, potentially inspiring readers to further investigate the synergy between these research directions. Date: Tuesday, 1 April 2025 Time: 2:30pm - 4:30pm Venue: Room 3494 Lifts 25/26 Committee Members: Prof. Tim Cheng (Supervisor) Dr. Yangqiu Song (Chairperson) Dr. Dan Xu