GRIG: Few-Shot Generative Residual Image Inpainting
Wanglong Lu
Xianta Jiang
Xiaogang Jin
Yongliang Yang
Minglun Gong
Tao Wang
Kaijie Shi
Hanli Zhao*
[Paper]
[GitHub]
Overall pipeline of our few-shot generative residual image inpainting framework (GRIG).

Abstract

Image inpainting is the task of filling in missing or masked region of an image with semantically meaningful contents. Recent methods have shown significant improvement in dealing with large-scale missing regions. However, these methods usually require large training datasets to achieve satisfactory results and there has been limited research into training these models on a small number of samples. To address this, we present a novel few-shot generative residual image inpainting method that produces high-quality inpainting results. The core idea is to propose an iterative residual reasoning method that incorporates Convolutional Neural Networks (CNNs) for feature extraction and Transformers for global reasoning within generative adversarial networks, along with image-level and patch-level discriminators. We also propose a novel forgery-patch adversarial training strategy to create faithful textures and detailed appearances. Extensive evaluations show that our method outperforms previous methods on the few-shot image inpainting task, both quantitatively and qualitatively.


Demo vedio


BiliBIli channel


 [GitHub]


Paper and Supplementary Material

Wanglong Lu and Xianta Jiang and Xiaogang Jin and Yong-Liang Yang and Minglun Gong and Tao Wang and Kaijie Shi and Hanli Zhao.
GRIG: Few-Shot Generative Residual Image Inpainting (hosted on ArXiv)


[Bibtex]


Acknowledgements

This template was originally made by Phillip Isola and Richard Zhang for a colorful ECCV project; Thanks for their awesome work!