Back to Search Start Over

Filter pruning for convolutional neural networks in semantic image segmentation.

Authors :
López-González, Clara I.
Gascó, Esther
Barrientos-Espillco, Fredy
Besada-Portas, Eva
Pajares, Gonzalo
Source :
Neural Networks. Jan2024, Vol. 169, p713-732. 20p.
Publication Year :
2024

Abstract

The remarkable performance of Convolutional Neural Networks (CNNs) has increased their use in real-time systems and devices with limited resources. Hence, compacting these networks while preserving accuracy has become necessary, leading to multiple compression methods. However, the majority require intensive iterative procedures and do not delve into the influence of the used data. To overcome these issues, this paper presents several contributions, framed in the context of e x plainable Artificial Intelligence (xAI): (a) two filter pruning methods for CNNs, which remove the less significant convolutional kernels; (b) a fine-tuning strategy to recover generalization; (c) a layer pruning approach for U-Net; and (d) an explanation of the relationship between performance and the used data. Filter and feature maps information are used in the pruning process: Principal Component Analysis (PCA) is combined with a next-convolution influence-metric, while the latter and the mean standard deviation are used in an importance score distribution-based method. The developed strategies are generic, and therefore applicable to different models. Experiments demonstrating their effectiveness are conducted over distinct CNNs and datasets, focusing mainly on semantic segmentation (using U-Net, DeepLabv3+, SegNet, and VGG-16 as highly representative models). Pruned U-Net on agricultural benchmarks achieves 98.7% parameters and 97.5% FLOPs drop, with a 0.35% gain in accuracy. DeepLabv3+ and SegNet on CamVid reach 46.5% and 72.4% parameters reduction and a 51.9% and 83.6% FLOPs drop respectively, with almost no decrease in accuracy. VGG-16 on CIFAR-10 obtains up to 86.5% parameter and 82.2% FLOPs decrease with a 0.78% accuracy gain. • New filter pruning and fine-tuning strategies for segmentation and classification. • Novel layer pruning method for U-Net, a relevant semantic segmentation model. • Performing a detailed pruning survey on encoder–decoder and classification models. • Significant reduction in parameters and FLOPs maintaining networks efficiency. • Analyzing the dependence of the pruned models performance on the used dataset. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
08936080
Volume :
169
Database :
Academic Search Index
Journal :
Neural Networks
Publication Type :
Academic Journal
Accession number :
174322346
Full Text :
https://doi.org/10.1016/j.neunet.2023.11.010