Swin-transformer for weak feature matching

Abstract Feature matching in computer vision is crucial but challenging in weakly textured scenes due to the lack of pattern repetition. We introduce the SwinMatcher feature matching method, aimed at addressing the issues of low matching quantity and poor matching precision in weakly textured scenes...

Full description

Saved in:
Bibliographic Details
Main Authors: Yuan Guo, Wenpeng Li, Ping Zhai
Format: Article
Language:English
Published: Nature Portfolio 2025-01-01
Series:Scientific Reports
Subjects:
Online Access:https://doi.org/10.1038/s41598-025-87309-9
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832585917638901760
author Yuan Guo
Wenpeng Li
Ping Zhai
author_facet Yuan Guo
Wenpeng Li
Ping Zhai
author_sort Yuan Guo
collection DOAJ
description Abstract Feature matching in computer vision is crucial but challenging in weakly textured scenes due to the lack of pattern repetition. We introduce the SwinMatcher feature matching method, aimed at addressing the issues of low matching quantity and poor matching precision in weakly textured scenes. Given the inherently significant local characteristics of image features, we employ a local self-attention mechanism to learn from weakly textured areas, maximally preserving the features of weak textures. To address the issue of incorrect matches in scenes with repetitive patterns, we use a cross-attention and positional encoding mechanism to learn the correct matches of repetitive patterns in two scenes, achieving higher matching precision. We also introduce a matching optimization algorithm that calculates the spatial expected coordinates of local two-dimensional heat maps of correspondences to obtain the final sub-pixel level matches. Experiments indicate that, under identical training conditions, the SwinMatcher outperforms other standard methods in pose estimation, homography estimation, and visual localization. It exhibits strong robustness and superior matching in weakly textured areas, offering a new research direction for feature matching in weakly textured images.
format Article
id doaj-art-67c301ae47fd4861b13c438b0d42f77f
institution Kabale University
issn 2045-2322
language English
publishDate 2025-01-01
publisher Nature Portfolio
record_format Article
series Scientific Reports
spelling doaj-art-67c301ae47fd4861b13c438b0d42f77f2025-01-26T12:23:37ZengNature PortfolioScientific Reports2045-23222025-01-0115111410.1038/s41598-025-87309-9Swin-transformer for weak feature matchingYuan Guo0Wenpeng Li1Ping Zhai2Department of Computer Science and Technology, Heilongjiang UniversitySchool of Computer and Control Engineering, Qiqihar UniversityDepartment of Computer Science and Technology, Qilu University of TechnologyAbstract Feature matching in computer vision is crucial but challenging in weakly textured scenes due to the lack of pattern repetition. We introduce the SwinMatcher feature matching method, aimed at addressing the issues of low matching quantity and poor matching precision in weakly textured scenes. Given the inherently significant local characteristics of image features, we employ a local self-attention mechanism to learn from weakly textured areas, maximally preserving the features of weak textures. To address the issue of incorrect matches in scenes with repetitive patterns, we use a cross-attention and positional encoding mechanism to learn the correct matches of repetitive patterns in two scenes, achieving higher matching precision. We also introduce a matching optimization algorithm that calculates the spatial expected coordinates of local two-dimensional heat maps of correspondences to obtain the final sub-pixel level matches. Experiments indicate that, under identical training conditions, the SwinMatcher outperforms other standard methods in pose estimation, homography estimation, and visual localization. It exhibits strong robustness and superior matching in weakly textured areas, offering a new research direction for feature matching in weakly textured images.https://doi.org/10.1038/s41598-025-87309-9Feature matchingDeep learningWeak textureTransformer
spellingShingle Yuan Guo
Wenpeng Li
Ping Zhai
Swin-transformer for weak feature matching
Scientific Reports
Feature matching
Deep learning
Weak texture
Transformer
title Swin-transformer for weak feature matching
title_full Swin-transformer for weak feature matching
title_fullStr Swin-transformer for weak feature matching
title_full_unstemmed Swin-transformer for weak feature matching
title_short Swin-transformer for weak feature matching
title_sort swin transformer for weak feature matching
topic Feature matching
Deep learning
Weak texture
Transformer
url https://doi.org/10.1038/s41598-025-87309-9
work_keys_str_mv AT yuanguo swintransformerforweakfeaturematching
AT wenpengli swintransformerforweakfeaturematching
AT pingzhai swintransformerforweakfeaturematching