loading page

MR-InpaintNet: Toward Deep Multi-Resolution Learning for Progressive Image Inpainting
  • +3
  • Huan Zhang ,
  • Zhao Zhang ,
  • Haijun Zhang ,
  • Yi Yang ,
  • Shuicheng Yan ,
  • Meng Wang
Huan Zhang
Author Profile
Zhao Zhang
Hefei University of Technology, Hefei University of Technology, Hefei University of Technology

Corresponding Author:[email protected]

Author Profile
Haijun Zhang
Author Profile
Shuicheng Yan
Author Profile
Meng Wang
Author Profile

Abstract

Deep learning based image inpainting methods have improved the performance greatly due to powerful representation ability of deep learning. However, current deep inpainting methods still tend to produce unreasonable structure and blurry texture, implying that image inpainting is still a challenging topic due to the ill-posed property of the task. To address these issues, we propose a novel deep multi-resolution learning-based progressive image inpainting method, termed MR-InpaintNet, which takes the damaged images of different resolutions as input and then fuses the multi-resolution features for repairing the damaged images. The idea is motivated by the fact that images of different resolutions can provide different levels of feature information. Specifically, the low-resolution image provides strong semantic information and the high-resolution image offers detailed texture information. The middle-resolution image can be used to reduce the gap between low-resolution and high-resolution images, which can further refine the inpainting result. To fuse and improve the multi-resolution features, a novel multi-resolution feature learning (MRFL) process is designed, which is consisted of a multi-resolution feature fusion (MRFF) module, an adaptive feature enhancement (AFE) module and a memory enhanced mechanism (MEM) module for information preservation. Then, the refined multi-resolution features contain both rich semantic information and detailed texture information from multiple resolutions. We further handle the refined multiresolution features by the decoder to obtain the recovered image. Extensive experiments on the Paris Street View, Places2 and CelebA-HQ datasets demonstrate that our proposed MRInpaintNet can effectively recover the textures and structures, and performs favorably against state-of-the-art methods.