Indexed by:
Abstract:
End-to-end learned image compression exploits the expressive power of nonlinear transform modules to de-correlate the spatial redundancies of image contents. Due to its long-range attention scheme, transformer-based transforms can explore more global features for better reconstruction. However, transformer modules bring in indispensable computational costs, and the coarse utilization of transformer in learned image compression cannot meet the coding efficiency. In this paper, we propose a novel graph-structured swin-transformer for learned image compression, shown in Figure 1. We assume that the global receptive field of attention map should be sparse not dense, while the local neighboring correlations must be strong. © 2024 IEEE.
Keyword:
Reprint Author's Address:
Email:
Source :
ISSN: 1068-0314
Year: 2024
Page: 592-
Language: English
Cited Count:
WoS CC Cited Count: 36
SCOPUS Cited Count:
ESI Highly Cited Papers on the List: 0 Unfold All
WanFang Cited Count:
Chinese Cited Count:
30 Days PV: 11
Affiliated Colleges: