Dual Embedding Transformer Network for Hyperspectral Unmixing
Hyperspectral unmixing is an essential task for achieving accurate perception of hyperspectral remote sensing information, aiming to overcome the limitation of spatial resolution and interpret the distribution of land features. To achieve the spatial and spectral feature representation of hyperspect...
Saved in:
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2025-01-01
|
Series: | IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10818529/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832592951456301056 |
---|---|
author | Huadong Yang Chengbi Zhang |
author_facet | Huadong Yang Chengbi Zhang |
author_sort | Huadong Yang |
collection | DOAJ |
description | Hyperspectral unmixing is an essential task for achieving accurate perception of hyperspectral remote sensing information, aiming to overcome the limitation of spatial resolution and interpret the distribution of land features. To achieve the spatial and spectral feature representation of hyperspectral images, we propose a dual embedding transformer network (DET-Net) based on an encoder-decoder architecture, which utilizes two transformer modules, including three-view spatial attention (TVA) module with 2-D embedding and multiscale spectral band group feature fusion (BGF) module with 3-D embedding to accomplish the task of hyperspectral unmixing. In TVA module, based on 2-D embedding, we introduce a three-view attention mechanism to extract more comprehensive spatial features. In BGF module, the transformer embedding is extended to band group spatial-spectral 3-D cubed embedding and establishes a series of spectral band groups. A cross-feature fusion mechanism is adopted to achieve multiscale spatial-spectral feature decoupling. With the collaboration of these two embeddings, DET-Net effectively captures complex spatial and spectral dependencies to decouple the tridimensional unmixing feature representation. Experimental results on synthetic and real datasets demonstrates the generalization performance of the proposed method, and the ablation experiments confirm the effectiveness of the TVA and BGF modules. |
format | Article |
id | doaj-art-b6cd6ec9c8b64cd982435aab9a26857d |
institution | Kabale University |
issn | 1939-1404 2151-1535 |
language | English |
publishDate | 2025-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing |
spelling | doaj-art-b6cd6ec9c8b64cd982435aab9a26857d2025-01-21T00:00:18ZengIEEEIEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing1939-14042151-15352025-01-01183514352910.1109/JSTARS.2024.352374710818529Dual Embedding Transformer Network for Hyperspectral UnmixingHuadong Yang0https://orcid.org/0000-0002-2657-063XChengbi Zhang1School of Information Science and Engineering, Shenyang Ligong University, Shenyang, ChinaSchool of Information Science and Engineering, Shenyang Ligong University, Shenyang, ChinaHyperspectral unmixing is an essential task for achieving accurate perception of hyperspectral remote sensing information, aiming to overcome the limitation of spatial resolution and interpret the distribution of land features. To achieve the spatial and spectral feature representation of hyperspectral images, we propose a dual embedding transformer network (DET-Net) based on an encoder-decoder architecture, which utilizes two transformer modules, including three-view spatial attention (TVA) module with 2-D embedding and multiscale spectral band group feature fusion (BGF) module with 3-D embedding to accomplish the task of hyperspectral unmixing. In TVA module, based on 2-D embedding, we introduce a three-view attention mechanism to extract more comprehensive spatial features. In BGF module, the transformer embedding is extended to band group spatial-spectral 3-D cubed embedding and establishes a series of spectral band groups. A cross-feature fusion mechanism is adopted to achieve multiscale spatial-spectral feature decoupling. With the collaboration of these two embeddings, DET-Net effectively captures complex spatial and spectral dependencies to decouple the tridimensional unmixing feature representation. Experimental results on synthetic and real datasets demonstrates the generalization performance of the proposed method, and the ablation experiments confirm the effectiveness of the TVA and BGF modules.https://ieeexplore.ieee.org/document/10818529/Abundance mapautoencoder (AE)deep learningendmember extractionhyperspectral image (HSI)hyperspectral unmixing (HU) |
spellingShingle | Huadong Yang Chengbi Zhang Dual Embedding Transformer Network for Hyperspectral Unmixing IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing Abundance map autoencoder (AE) deep learning endmember extraction hyperspectral image (HSI) hyperspectral unmixing (HU) |
title | Dual Embedding Transformer Network for Hyperspectral Unmixing |
title_full | Dual Embedding Transformer Network for Hyperspectral Unmixing |
title_fullStr | Dual Embedding Transformer Network for Hyperspectral Unmixing |
title_full_unstemmed | Dual Embedding Transformer Network for Hyperspectral Unmixing |
title_short | Dual Embedding Transformer Network for Hyperspectral Unmixing |
title_sort | dual embedding transformer network for hyperspectral unmixing |
topic | Abundance map autoencoder (AE) deep learning endmember extraction hyperspectral image (HSI) hyperspectral unmixing (HU) |
url | https://ieeexplore.ieee.org/document/10818529/ |
work_keys_str_mv | AT huadongyang dualembeddingtransformernetworkforhyperspectralunmixing AT chengbizhang dualembeddingtransformernetworkforhyperspectralunmixing |