*Result*: Dynamic network compression via probabilistic channel pruning.

Title:
Dynamic network compression via probabilistic channel pruning.
Authors:
Lee K; Dept. of CSE, Konkuk University, Seoul, 05029, Republic of Korea. Electronic address: kwan7595@konkuk.ac.kr., Lee HW; Dept. of CSE, Konkuk University, Seoul, 05029, Republic of Korea. Electronic address: leehw@konkuk.ac.kr.
Source:
Neural networks : the official journal of the International Neural Network Society [Neural Netw] 2026 Jan; Vol. 193, pp. 108080. Date of Electronic Publication: 2025 Sep 04.
Publication Type:
Journal Article
Language:
English
Journal Info:
Publisher: Pergamon Press Country of Publication: United States NLM ID: 8805018 Publication Model: Print-Electronic Cited Medium: Internet ISSN: 1879-2782 (Electronic) Linking ISSN: 08936080 NLM ISO Abbreviation: Neural Netw Subsets: MEDLINE
Imprint Name(s):
Original Publication: New York : Pergamon Press, [c1988-
Contributed Indexing:
Keywords: Connectivity; Model compression; Neural network pruning; Probabilistic channel pruning
Entry Date(s):
Date Created: 20250909 Date Completed: 20251217 Latest Revision: 20251217
Update Code:
20260130
DOI:
10.1016/j.neunet.2025.108080
PMID:
40925121
Database:
MEDLINE

*Further Information*

*Neural network compression problems have been extensively studied to overcome the limitations of compute-intensive deep learning models. Most of the state-of-the-art solutions in this context are based on network pruning that identify and remove unimportant weights, filters or channels. However, existing methods often lack actual speedup or require complex pruning criteria and additional training (fine-tuning) overhead. To address these limitations, we develop probability-based connectivity module that determines the connection of each channel to the next layer. Our connectivity module enables to dynamically activate and deactivate channel connections during training, and hence, does not necessitate fine-tuning of the pruned model. We show that the convolution decomposition, which decomposes convolution with connectivity module and depth-wise convolution can effectively induce sparsity, resulting in 52.76 %, 46.05 % reduction of parameter counts, with even boosting accuracy (+0.19 %, + 0.3 %) compared to baseline architectures in ResNet-56, VGG-19 Models. We also introduce resource-aware regularization that exploits the probabilistic activation of connectivity module in order to control the level of compression. We show that our method achieves comparable level of compression and accuracy to the state-of-the-art pruning methods.
(Copyright © 2025 Elsevier Ltd. All rights reserved.)*

*Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.*