• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

Wei, Fuhao (Wei, Fuhao.) | Wang, Shaofan (Wang, Shaofan.) | Sun, Yanfeng (Sun, Yanfeng.) | Yin, Baocai (Yin, Baocai.)

Indexed by:

EI Scopus SCIE

Abstract:

Developing real-time cloud segmentation technology is urgent for many remote sensing based applications such as weather forecasting. Existing deep learning based cloud segmentation methods involve two shortcomings. (a): They tend to produce discontinuous boundaries and fail to capture less salient feature, which corresponds to thin cloud pixels; (b): they are unrobust towards different scenarios. Those issues are circumvented by integrating U-Net and the swin transformer together, with an efficiently designed dual attention mechanism based skip connection. Typically, a swin transformer based encoder-decoder network, by incorporating a dual attentional skip connection with Swin-UNet (DASUNet) is proposed. DASUNet captures the global relationship of image patches based on its window attention mechanism, which fits the real-time requirement. Moreover, DASUNet characterizes the less salient features by equipping with token dual attention modules among the skip connection, which compensates the ignorance of less salient features incurred from traditional attention mechanism during the stacking of transformer layers. Experiments on ground-based images (SWINySeg) and remote sensing images (HRC-WHU, 38-Cloud) show that, DASUNet achieves the state-of-the-art or competitive results for cloud segmentation (six top-1 positions of six metrics among 11 methods on SWINySeg, two top-1 positions of five metrics among 10 methods on HRC-WHU, two top-1 positions of four metrics among 12 methods with ParaNum <= 30M$\le 30{\rm M}$ on 38-Cloud), with 100FPS implementation speed averagely for each 224x224$224\times 224$ image. We propose a swin transformer based encoder-decoder network, by incorporating a dual attentional skip connection with Swin-UNet (DASUNet) for cloud segmentation. DASUNet characterizes the less salient features by equipping with token dual attention modules among the skip connection. The token dual attention module consists of token similarity attention and token importance attention, which compensates the ignorance of less salient features incurred from traditional attention mechanism during the stacking of transformer layers. image

Keyword:

feature selection image segmentation image processing

Author Community:

  • [ 1 ] [Wei, Fuhao]Beijing Univ Technol, Sch Informat Sci & Technol, Beijing Key Lab Multimedia & Intelligent Software, Beijing, Peoples R China
  • [ 2 ] [Wang, Shaofan]Beijing Univ Technol, Sch Informat Sci & Technol, Beijing Key Lab Multimedia & Intelligent Software, Beijing, Peoples R China
  • [ 3 ] [Sun, Yanfeng]Beijing Univ Technol, Sch Informat Sci & Technol, Beijing Key Lab Multimedia & Intelligent Software, Beijing, Peoples R China
  • [ 4 ] [Yin, Baocai]Beijing Univ Technol, Sch Informat Sci & Technol, Beijing Key Lab Multimedia & Intelligent Software, Beijing, Peoples R China

Reprint Author's Address:

  • [Wang, Shaofan]Beijing Univ Technol, Sch Informat Sci & Technol, Beijing Key Lab Multimedia & Intelligent Software, Beijing, Peoples R China;;

Show more details

Related Keywords:

Source :

IET IMAGE PROCESSING

ISSN: 1751-9659

Year: 2024

Issue: 12

Volume: 18

Page: 3460-3479

2 . 3 0 0

JCR@2022

Cited Count:

WoS CC Cited Count: 2

SCOPUS Cited Count: 2

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 4

Affiliated Colleges:

Online/Total:1569/10568374
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.