CLC number:
On-line Access: 2024-12-20
Received: 2024-05-14
Revision Accepted: 2024-10-25
Crosschecked: 0000-00-00
Cited: 0
Clicked: 49
Shiyuan YANG, Zheng GU, Wenyue HAO, Yi WANG1, Huaiyu CAI, Xiaodong CHEN. Few-shot exemplar-driven inpainting with parameter-efficient diffusion fine-tuning[J]. Frontiers of Information Technology & Electronic Engineering, 1998, -1(-1): .
@article{title="Few-shot exemplar-driven inpainting with parameter-efficient diffusion fine-tuning",
author="Shiyuan YANG, Zheng GU, Wenyue HAO, Yi WANG1, Huaiyu CAI, Xiaodong CHEN",
journal="Frontiers of Information Technology & Electronic Engineering",
volume="-1",
number="-1",
pages="",
year="1998",
publisher="Zhejiang University Press & Springer",
doi="10.1631/FITEE.2400395"
}
%0 Journal Article
%T Few-shot exemplar-driven inpainting with parameter-efficient diffusion fine-tuning
%A Shiyuan YANG
%A Zheng GU
%A Wenyue HAO
%A Yi WANG1
%A Huaiyu CAI
%A Xiaodong CHEN
%J Journal of Zhejiang University SCIENCE C
%V -1
%N -1
%P
%@ 2095-9184
%D 1998
%I Zhejiang University Press & Springer
%DOI 10.1631/FITEE.2400395
TY - JOUR
T1 - Few-shot exemplar-driven inpainting with parameter-efficient diffusion fine-tuning
A1 - Shiyuan YANG
A1 - Zheng GU
A1 - Wenyue HAO
A1 - Yi WANG1
A1 - Huaiyu CAI
A1 - Xiaodong CHEN
J0 - Journal of Zhejiang University Science C
VL - -1
IS - -1
SP -
EP -
%@ 2095-9184
Y1 - 1998
PB - Zhejiang University Press & Springer
ER -
DOI - 10.1631/FITEE.2400395
Abstract: Text-to-image diffusion models have demonstrated impressive capabilities in image generation that have been effectively applied to image inpainting. While text prompt provides an intuitive guidance for conditional inpainting, users often seek the ability to inpaint a specific object with customized appearance by providing an exemplar image. Unfortunately, existing methods struggle to achieve high-fidelity in exemplar-driven inpainting. To address this, we utilized a plug-and-play low-rank adaptation(LoRA) module based on a pretrained text-driven inpainting model. The LoRA module is dedicated to learn the exemplar-specific concepts through few-shot fine-tuning, bringing improved fitting capability to customized exemplar images, without intensive training on large-scale datasets. Additionally, we introduced GPT-4v prompting and prior noise initialization techniques to further facilitate the fidelity in inpainting results. In brief, the denoising diffusion process first starts with the noise derived from a composite exemplar-background image, and is subsequently guided by an expressive prompt generated from the exemplar using the GPT-4v model. Extensive experiments demonstrate that our method achieves state-of-the-art performance, both qualitatively and quantitatively, offering users an exemplar-driven inpainting tool with enhanced customization capability.
Open peer comments: Debate/Discuss/Question/Opinion
<1>