Mix-layers semantic extraction and multi-scale aggregation transformer for semantic segmentation
Abstract Recently, a number of vision transformer models for semantic segmentation have been proposed, with the majority of these achieving impressive results. However, they lack the ability to exploit the intrinsic position and channel features of the image and are less capable of multi-scale featu...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Springer
2024-11-01
|
Series: | Complex & Intelligent Systems |
Subjects: | |
Online Access: | https://doi.org/10.1007/s40747-024-01650-6 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832571197643030528 |
---|---|
author | Tianping Li Xiaolong Yang Zhenyi Zhang Zhaotong Cui Zhou Maoxia |
author_facet | Tianping Li Xiaolong Yang Zhenyi Zhang Zhaotong Cui Zhou Maoxia |
author_sort | Tianping Li |
collection | DOAJ |
description | Abstract Recently, a number of vision transformer models for semantic segmentation have been proposed, with the majority of these achieving impressive results. However, they lack the ability to exploit the intrinsic position and channel features of the image and are less capable of multi-scale feature fusion. This paper presents a semantic segmentation method that successfully combines attention and multiscale representation, thereby enhancing performance and efficiency. This represents a significant advancement in the field. Multi-layers semantic extraction and multi-scale aggregation transformer decoder (MEMAFormer) is proposed, which consists of two components: mix-layers dual channel semantic extraction module (MDCE) and semantic aggregation pyramid pooling module (SAPPM). The MDCE incorporates a multi-layers cross attention module (MCAM) and an efficient channel attention module (ECAM). In MCAM, horizontal connections between encoder and decoder stages are employed as feature queries for the attention module. The hierarchical feature maps derived from different encoder and decoder stages are integrated into key and value. To address long-term dependencies, ECAM selectively emphasizes interdependent channel feature maps by integrating relevant features across all channels. The adaptability of the feature maps is reduced by pyramid pooling, which reduces the amount of computation without compromising performance. SAPPM is comprised of several distinct pooled kernels that extract context with a deeper flow of information, forming a multi-scale feature by integrating various feature sizes. The MEMAFormer-B0 model demonstrates superior performance compared to SegFormer-B0, exhibiting gains of 4.8%, 4.0% and 3.5% on the ADE20K, Cityscapes and COCO-stuff datasets, respectively. |
format | Article |
id | doaj-art-cbd44f27c47d4139b42105b1d087189c |
institution | Kabale University |
issn | 2199-4536 2198-6053 |
language | English |
publishDate | 2024-11-01 |
publisher | Springer |
record_format | Article |
series | Complex & Intelligent Systems |
spelling | doaj-art-cbd44f27c47d4139b42105b1d087189c2025-02-02T12:49:03ZengSpringerComplex & Intelligent Systems2199-45362198-60532024-11-0111111510.1007/s40747-024-01650-6Mix-layers semantic extraction and multi-scale aggregation transformer for semantic segmentationTianping Li0Xiaolong Yang1Zhenyi Zhang2Zhaotong Cui3Zhou Maoxia4School of Physics and Electronics, Shandong Normal UniversitySchool of Physics and Electronics, Shandong Normal UniversitySchool of Physics and Electronics, Shandong Normal UniversitySchool of Physics and Electronics, Shandong Normal UniversitySchool of Physics and Electronics, Shandong Normal UniversityAbstract Recently, a number of vision transformer models for semantic segmentation have been proposed, with the majority of these achieving impressive results. However, they lack the ability to exploit the intrinsic position and channel features of the image and are less capable of multi-scale feature fusion. This paper presents a semantic segmentation method that successfully combines attention and multiscale representation, thereby enhancing performance and efficiency. This represents a significant advancement in the field. Multi-layers semantic extraction and multi-scale aggregation transformer decoder (MEMAFormer) is proposed, which consists of two components: mix-layers dual channel semantic extraction module (MDCE) and semantic aggregation pyramid pooling module (SAPPM). The MDCE incorporates a multi-layers cross attention module (MCAM) and an efficient channel attention module (ECAM). In MCAM, horizontal connections between encoder and decoder stages are employed as feature queries for the attention module. The hierarchical feature maps derived from different encoder and decoder stages are integrated into key and value. To address long-term dependencies, ECAM selectively emphasizes interdependent channel feature maps by integrating relevant features across all channels. The adaptability of the feature maps is reduced by pyramid pooling, which reduces the amount of computation without compromising performance. SAPPM is comprised of several distinct pooled kernels that extract context with a deeper flow of information, forming a multi-scale feature by integrating various feature sizes. The MEMAFormer-B0 model demonstrates superior performance compared to SegFormer-B0, exhibiting gains of 4.8%, 4.0% and 3.5% on the ADE20K, Cityscapes and COCO-stuff datasets, respectively.https://doi.org/10.1007/s40747-024-01650-6Semantic segmentationMEMAFormerMDCESAPPM |
spellingShingle | Tianping Li Xiaolong Yang Zhenyi Zhang Zhaotong Cui Zhou Maoxia Mix-layers semantic extraction and multi-scale aggregation transformer for semantic segmentation Complex & Intelligent Systems Semantic segmentation MEMAFormer MDCE SAPPM |
title | Mix-layers semantic extraction and multi-scale aggregation transformer for semantic segmentation |
title_full | Mix-layers semantic extraction and multi-scale aggregation transformer for semantic segmentation |
title_fullStr | Mix-layers semantic extraction and multi-scale aggregation transformer for semantic segmentation |
title_full_unstemmed | Mix-layers semantic extraction and multi-scale aggregation transformer for semantic segmentation |
title_short | Mix-layers semantic extraction and multi-scale aggregation transformer for semantic segmentation |
title_sort | mix layers semantic extraction and multi scale aggregation transformer for semantic segmentation |
topic | Semantic segmentation MEMAFormer MDCE SAPPM |
url | https://doi.org/10.1007/s40747-024-01650-6 |
work_keys_str_mv | AT tianpingli mixlayerssemanticextractionandmultiscaleaggregationtransformerforsemanticsegmentation AT xiaolongyang mixlayerssemanticextractionandmultiscaleaggregationtransformerforsemanticsegmentation AT zhenyizhang mixlayerssemanticextractionandmultiscaleaggregationtransformerforsemanticsegmentation AT zhaotongcui mixlayerssemanticextractionandmultiscaleaggregationtransformerforsemanticsegmentation AT zhoumaoxia mixlayerssemanticextractionandmultiscaleaggregationtransformerforsemanticsegmentation |