• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

Wang, Chen (Wang, Chen.) | Wang, Jin (Wang, Jin.) | Zhu, Qing (Zhu, Qing.) | Yin, Baocai (Yin, Baocai.)

Indexed by:

EI

Abstract:

Image inpainting is a challenging task in image processing and widely applied in many areas such as photo editing. Traditional patch-based methods are not effective to deal with complex or non-repetitive structures. Recently, deep learning-based approaches have shown promising results for image inpainting. However, they usually generate contents with artificial boundaries, distorted structures or blurry textures. To handle this problem, we propose a novel image inpainting method based on wavelet transform attention model (WTAM). The wavelet transform decomposes features into multi-frequency subbands for extracting and transmitting deep information, and the attention mechanism enhances the ability of wavelet transform to capture significant detailed information in each level's subband images. Extensive experimental results on multiple datasets (Paris StreetView, CelebA and CelebAMask-HQ) demonstrate that our method can not only synthesize sharp image structures but also generate fine-detailed textures in missing regions, significantly outperforming the state-of-the-art methods. © 2020 IEEE

Keyword:

Wavelet transforms Computer vision Image compression Image enhancement Textures Deep learning

Author Community:

  • [ 1 ] [Wang, Chen]Faculty of Information Technology, Beijing University of Technology, Beijing, China
  • [ 2 ] [Wang, Jin]Faculty of Information Technology, Beijing University of Technology, Beijing, China
  • [ 3 ] [Zhu, Qing]Faculty of Information Technology, Beijing University of Technology, Beijing, China
  • [ 4 ] [Yin, Baocai]Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology, Dalian, China

Reprint Author's Address:

Email:

Show more details

Related Keywords:

Related Article:

Source :

ISSN: 0271-4310

Year: 2020

Volume: 2020-October

Language: English

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count:

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 6

Affiliated Colleges:

Online/Total:718/10672818
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.