Multi-Scale Channel Distillation Network for Image Compressive Sensing
Recently, convolutional neural networks (CNNs) have demonstrated striking success in computer vision tasks. Methods based on CNNs for image compressive sensing (CS) have also gained prominence. However, existing methods tend to increase the depth of the network in feature space for better reconstruc...
Saved in:
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2025-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10835084/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Recently, convolutional neural networks (CNNs) have demonstrated striking success in computer vision tasks. Methods based on CNNs for image compressive sensing (CS) have also gained prominence. However, existing methods tend to increase the depth of the network in feature space for better reconstruction quality, neglecting the hierarchical representation of intermediate features in pixel space. In order to coordinate the feature space and pixel space to complete the deep reconstruction of images, and further improve the reconstruction performance of current CS methods, we propose a multi-scale channel distillation network (MSCDN). This network first obtains images of multiple scales using a scale-space image decomposition method at the sampling stage, followed by sampling these decomposed images through a convolutional operation. In this way, multi-scale information in the compressed domain is aggregated. During the reconstruction phase, a low-frequency information recovery network generates a preliminary image, whereas a high-frequency feature aggregation network refines the image further. Specifically, we design a dual-branch deep reconstruction architecture with channel distillation residual block (CDRB) as the core component. One branch extracts features gradually by cascading multiple CDRB modules, thereby supplementing the initial reconstructed image with a large amount of high-frequency content in feature space. The other branch takes the initial reconstructed image as input and sequentially fuses the intermediate feature outputs by CDRBs to increase the local details of the image in pixel space. Combining outputs from both branches, we achieve an optimal reconstructed image. Extensive experimental results on four benchmark datasets demonstrate that MSCDN surpasses state-of-the-art CS methods not only in reconstruction accuracy but also in perceptual visual quality. |
---|---|
ISSN: | 2169-3536 |