Indexed by:
Abstract:
Few-shot semantic segmentation aims to segment new objects in the image with limited annotations. Typically, in metric-based few-shot learning, the expression of categories is obtained by averaging global support object information. However, a single prototype cannot accurately describe a category. Meanwhile, simple foreground averaging operations also ignore the dependencies between objects and their surroundings. In this paper, we propose a novel Transformer-based Prototype Search Network (TPSN) for few-shot segmentation. We use the transformer encoder to integrate information between different image regions and then use the decoder to express a category in terms of multiple prototypes. The multi-prototype approach can effectively alleviate the feature fluctuation caused by limited annotation data. Moreover, we use adaptive prototype search during multi-prototype extraction instead of the ordinary averaging operation compared with the previous few-shot prototype framework. This helps the network integrate the different image regions’ information and fuse object features with their dependent background information, obtaining more reasonable prototype expressions. In addition, to encourage the category's prototypes to focus on different parts and maintain consistency in high-level semantics, we use the diversity and consistency loss to constrain the multi-prototype training. Experiments show that our algorithm achieves state-of-the-art performance in few-shot segmentation on two datasets: PASCAL-5i and COCO-20i. © 2022 Elsevier Ltd
Keyword:
Reprint Author's Address:
Email:
Source :
Computers and Electrical Engineering
ISSN: 0045-7906
Year: 2022
Volume: 103
4 . 3
JCR@2022
4 . 3 0 0
JCR@2022
ESI Discipline: COMPUTER SCIENCE;
ESI HC Threshold:46
JCR Journal Grade:2
CAS Journal Grade:3
Cited Count:
WoS CC Cited Count: 0
SCOPUS Cited Count: 13
ESI Highly Cited Papers on the List: 0 Unfold All
WanFang Cited Count:
Chinese Cited Count:
30 Days PV: 14
Affiliated Colleges: