More about HKUST
Towards Trustworthy Visual Generative Models: Reliable and Controllable Generation of Diffusion Models
The Hong Kong University of Science and Technology
Department of Computer Science and Engineering
MPhil Thesis Defence
Title: "Towards Trustworthy Visual Generative Models: Reliable and Controllable
Generation of Diffusion Models"
By
Mr. Sen LI
Abstract:
Visual generative models, especially diffusion models, have demonstrated
incredible performance on high-quality visual generation, attracting more and
more attention in both academia and industry. Representative models or tools
such as DALLE-3 and MidJourney have been widely used in daily life to
facilitate the creation of artworks or pictures. However, these powerful tools
also bring potential risks since they can be maliciously used to generate and
disseminate unsafe content such as pornographic and violent pictures, which may
cause severe results. In this thesis, we discuss how to make visual generative
models more reliable and controllable from different aspects. In particular, we
focus on diffusion models as they are the most widely used visual generative
models.
Firstly, we uncover the potential risks existing in diffusion models, showing
that they can be easily inserted with (malicious) invisible backdoors during
training which can result in unreliable and harmful behaviors. To this end, we
propose a novel bi-level optimization framework to formulate the training
process, which can be instantiated by proposed different algorithms for
unconditional and conditional diffusion models, respectively. Extensive
experiments show that backdoors can be effectively inserted without affecting
the benign performance of models, making the backdoors more stealthy and
robust. Also, we empirically find that current various defense methods cannot
mitigate the proposed invisible backdoors, enhancing the usability in practice.
Moreover, the proposed invisible backdoors can be directly applied to model
watermarking for model ownership verification in black-box setting, further
enhancing the significance of the proposed framework.
Then, we focus on the controllable generation of text-to-image diffusion
models. We introduce MuLan, a Multimodal-LLM agent, to progressively generate
objects given a text prompt. MuLan firstly decompose the prompt to several
sub-prompts, and each sub-prompt focuses on only one object. Each object is
generated conditioned on previously generated objects. With a VLM
(Vision-Language Model) checker, MuLan can timely monitor the process and
adaptively correct possible mistakes after each generation stage. MuLan can
greatly boost the generation performance in terms of object attributes and
spatial relationships in text prompts. Evaluated by GPT-4V and human, extensive
experiments show the superior performance of MuLan. In addition, we show that
MuLan can enable human-agent interaction during generation, further enhancing
the flexibility and effectiveness of the generation process.
Date: Monday, 19 August 2024
Time: 12:00noon - 2:00pm
Venue: Room 5501
Lifts 25/26
Chairman: Dr. Dan XU
Committee Members: Dr. Shuai WANG (Supervisor)
Dr. Long CHEN