ProSpect: Prompt Spectrum for Attribute-Aware Personalization of Diffusion Models
Zhang, Yuxin1,2; Dong, Weiming1,2; Tang, Fan3; Huang, Nisha1,2; Huang, Haibin4; Ma, Chongyang4; Lee, Tong-Yee5; Deussen, Oliver6; Xu, Changsheng1,2
发表期刊ACM TRANSACTIONS ON GRAPHICS
ISSN0730-0301
2023-12-01
卷号42期号:6页码:14
通讯作者Dong, Weiming(weiming.dong@ia.ac.cn)
摘要Personalizing generative models offers a way to guide image generation with user-provided references. Current personalization methods can invert an object or concept into the textual conditioning space and compose new natural sentences for text-to-image diffusion models. However, representing and editing specific visual attributes such as material, style, and layout remains a challenge, leading to a lack of disentanglement and editability. To address this problem, we propose a novel approach that leverages the step-by-step generation process of diffusion models, which generate images from low to high frequency information, providing a new perspective on representing, generating, and editing images. We develop the Prompt Spectrum Space P*, an expanded textual conditioning space, and a new image representation method called ProSpect. ProSpect represents an image as a collection of inverted textual token embeddings encoded from per-stage prompts, where each prompt corresponds to a specific generation stage (i.e., a group of consecutive steps) of the diffusion model. Experimental results demonstrate that P* and ProSpect offer better disentanglement and controllability compared to existing methods. We apply ProSpect in various personalized attribute-aware image generation applications, such as image-guided or text-driven manipulations of materials, style, and layout, achieving previously unattainable results from a single image input without fine-tuning the diffusion models. Our source code is available at https: //github.com/zyxElsa/ProSpect.
关键词Image generation Diffusion models Attribute-aware editing Model personalization
DOI10.1145/3618342
收录类别SCI
语种英语
资助项目National Key R&D Program of China[2020AAA0106200] ; National Natural Science Foundation of China[61832016] ; National Natural Science Foundation of China[62102162] ; National Natural Science Foundation of China[U20B2070] ; Beijing Natural Science Foundation[L221013] ; National Science and Technology Council[111-2221-E-006-112-MY3] ; Deutsche Forschungsgemeinschaft (DFG)[413891298]
项目资助者National Key R&D Program of China ; National Natural Science Foundation of China ; Beijing Natural Science Foundation ; National Science and Technology Council ; Deutsche Forschungsgemeinschaft (DFG)
WOS研究方向Computer Science
WOS类目Computer Science, Software Engineering
WOS记录号WOS:001139790400072
出版者ASSOC COMPUTING MACHINERY
引用统计
被引频次:1[WOS]   [WOS记录]     [WOS相关记录]
文献类型期刊论文
条目标识符http://ir.ia.ac.cn/handle/173211/55395
专题多模态人工智能系统全国重点实验室
通讯作者Dong, Weiming
作者单位1.Chinese Acad Sci, Inst Automat, MAIS, Beijing, Peoples R China
2.UCAS, Sch Artificial Intelligence, Beijing, Peoples R China
3.Chinese Acad Sci, Inst Comp Technol, Beijing, Peoples R China
4.Kuaishou Technol, Beijing, Peoples R China
5.Natl Cheng Kung Univ, Tainan, Taiwan
6.Univ Konstanz, Constance, Germany
第一作者单位中国科学院自动化研究所
通讯作者单位中国科学院自动化研究所
推荐引用方式
GB/T 7714
Zhang, Yuxin,Dong, Weiming,Tang, Fan,et al. ProSpect: Prompt Spectrum for Attribute-Aware Personalization of Diffusion Models[J]. ACM TRANSACTIONS ON GRAPHICS,2023,42(6):14.
APA Zhang, Yuxin.,Dong, Weiming.,Tang, Fan.,Huang, Nisha.,Huang, Haibin.,...&Xu, Changsheng.(2023).ProSpect: Prompt Spectrum for Attribute-Aware Personalization of Diffusion Models.ACM TRANSACTIONS ON GRAPHICS,42(6),14.
MLA Zhang, Yuxin,et al."ProSpect: Prompt Spectrum for Attribute-Aware Personalization of Diffusion Models".ACM TRANSACTIONS ON GRAPHICS 42.6(2023):14.
条目包含的文件
条目无相关文件。
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[Zhang, Yuxin]的文章
[Dong, Weiming]的文章
[Tang, Fan]的文章
百度学术
百度学术中相似的文章
[Zhang, Yuxin]的文章
[Dong, Weiming]的文章
[Tang, Fan]的文章
必应学术
必应学术中相似的文章
[Zhang, Yuxin]的文章
[Dong, Weiming]的文章
[Tang, Fan]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。