Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning

While Convolutional Neural Networks (CNNs) excel at learning complex latent-space representations, their over-parameterization can lead to overfitting and reduced performance, particularly with limited data. This, alongside their high computational and memory demands, limits the applicability of CNN...

Full description

Saved in:
Bibliographic Details
Main Authors: Manish Sharma, Jamison Heard, Eli Saber, Panagiotis Markopoulos
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10851278/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832576772300865536
author Manish Sharma
Jamison Heard
Eli Saber
Panagiotis Markopoulos
author_facet Manish Sharma
Jamison Heard
Eli Saber
Panagiotis Markopoulos
author_sort Manish Sharma
collection DOAJ
description While Convolutional Neural Networks (CNNs) excel at learning complex latent-space representations, their over-parameterization can lead to overfitting and reduced performance, particularly with limited data. This, alongside their high computational and memory demands, limits the applicability of CNNs for edge deployment and applications where computational resources are constrained. Low-rank matrix approximation has emerged as a promising approach to reduce CNN parameters, but existing methods often require pre-determined ranks or involve complex post-training adjustments, leading to challenges in rank selection, performance loss, and limited practicality in resource-constrained environments. This underscores the need for an adaptive compression method that integrates into the training process, dynamically adjusting model complexity based on data and task requirements. To address this, we propose an efficient training method for CNN compression via dynamic parameter rank pruning. Our approach integrates efficient matrix factorization and novel regularization techniques, forming a robust framework for dynamic rank pruning and model compression. By using Singular Value Decomposition (SVD) to model low-rank convolutional filters and dense weight matrices, and training the SVD factors with back-propagation in an end-to-end manner, we achieve model compression. We evaluate our method on modern CNNs, including ResNet-18, ResNet-20, and ResNet-32, using datasets like CIFAR-10, CIFAR-100, and ImageNet (2012). Our experiments demonstrate that the proposed method can reduce model parameters by up to 50% and improve classification accuracy by up to 2% over baseline models, making CNNs more feasible for practical applications.
format Article
id doaj-art-94387e2af56549a189b7e49471dfad53
institution Kabale University
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-94387e2af56549a189b7e49471dfad532025-01-31T00:01:24ZengIEEEIEEE Access2169-35362025-01-0113184411845610.1109/ACCESS.2025.353341910851278Convolutional Neural Network Compression via Dynamic Parameter Rank PruningManish Sharma0https://orcid.org/0000-0002-1867-9379Jamison Heard1https://orcid.org/0000-0001-6860-0844Eli Saber2https://orcid.org/0009-0002-8593-4015Panagiotis Markopoulos3https://orcid.org/0000-0001-9686-779XChester F. Carlson Center for Imaging Science, Rochester Institute of Technology, Rochester, NY, USADepartment of Electrical and Microelectronic Engineering, Rochester Institute of Technology, Rochester, NY, USAChester F. Carlson Center for Imaging Science, Rochester Institute of Technology, Rochester, NY, USADepartment of Electrical and Computer Engineering, The University of Texas at San Antonio, San Antonio, TX, USAWhile Convolutional Neural Networks (CNNs) excel at learning complex latent-space representations, their over-parameterization can lead to overfitting and reduced performance, particularly with limited data. This, alongside their high computational and memory demands, limits the applicability of CNNs for edge deployment and applications where computational resources are constrained. Low-rank matrix approximation has emerged as a promising approach to reduce CNN parameters, but existing methods often require pre-determined ranks or involve complex post-training adjustments, leading to challenges in rank selection, performance loss, and limited practicality in resource-constrained environments. This underscores the need for an adaptive compression method that integrates into the training process, dynamically adjusting model complexity based on data and task requirements. To address this, we propose an efficient training method for CNN compression via dynamic parameter rank pruning. Our approach integrates efficient matrix factorization and novel regularization techniques, forming a robust framework for dynamic rank pruning and model compression. By using Singular Value Decomposition (SVD) to model low-rank convolutional filters and dense weight matrices, and training the SVD factors with back-propagation in an end-to-end manner, we achieve model compression. We evaluate our method on modern CNNs, including ResNet-18, ResNet-20, and ResNet-32, using datasets like CIFAR-10, CIFAR-100, and ImageNet (2012). Our experiments demonstrate that the proposed method can reduce model parameters by up to 50% and improve classification accuracy by up to 2% over baseline models, making CNNs more feasible for practical applications.https://ieeexplore.ieee.org/document/10851278/Convolutional neural networkdynamic rank selectionimage classificationlow-rank factorizationmodel compressionmodel pruning
spellingShingle Manish Sharma
Jamison Heard
Eli Saber
Panagiotis Markopoulos
Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
IEEE Access
Convolutional neural network
dynamic rank selection
image classification
low-rank factorization
model compression
model pruning
title Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
title_full Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
title_fullStr Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
title_full_unstemmed Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
title_short Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
title_sort convolutional neural network compression via dynamic parameter rank pruning
topic Convolutional neural network
dynamic rank selection
image classification
low-rank factorization
model compression
model pruning
url https://ieeexplore.ieee.org/document/10851278/
work_keys_str_mv AT manishsharma convolutionalneuralnetworkcompressionviadynamicparameterrankpruning
AT jamisonheard convolutionalneuralnetworkcompressionviadynamicparameterrankpruning
AT elisaber convolutionalneuralnetworkcompressionviadynamicparameterrankpruning
AT panagiotismarkopoulos convolutionalneuralnetworkcompressionviadynamicparameterrankpruning