Ahmadi, A., Mahboobi Esfanjani, R.. (1403). Structure Learning for Deep Neural Networks with Competitive Synaptic Pruning. فناوری آموزش, 13(1), 189-196. doi: 10.22061/jecei.2024.11017.758
A. Ahmadi; R. Mahboobi Esfanjani. "Structure Learning for Deep Neural Networks with Competitive Synaptic Pruning". فناوری آموزش, 13, 1, 1403, 189-196. doi: 10.22061/jecei.2024.11017.758
Ahmadi, A., Mahboobi Esfanjani, R.. (1403). 'Structure Learning for Deep Neural Networks with Competitive Synaptic Pruning', فناوری آموزش, 13(1), pp. 189-196. doi: 10.22061/jecei.2024.11017.758
Ahmadi, A., Mahboobi Esfanjani, R.. Structure Learning for Deep Neural Networks with Competitive Synaptic Pruning. فناوری آموزش, 1403; 13(1): 189-196. doi: 10.22061/jecei.2024.11017.758
Department of Electrical Engineering, Sahand University of Technology, Tabriz, Iran.
تاریخ دریافت: 05 مرداد 1403،
تاریخ بازنگری: 25 مهر 1403،
تاریخ پذیرش: 10 آبان 1403
چکیده
Background and Objectives: A predefined structure is usually employed for deep neural networks, which results in over- or underfitting, heavy processing load, and storage overhead. Training along with pruning can decrease redundancy in deep neural networks; however, it may lead to a decrease in accuracy. Methods: In this note, we provide a novel approach for structure optimization of deep neural networks based on competition of connections merged with brain-inspired synaptic pruning. The efficiency of each network connection is continuously assessed in the proposed scheme based on the global gradient magnitude criterion, which also considers positive scores for strong and more effective connections and negative scores for weak connections. But a connection with a weak score is not removed quickly; instead, it is eliminated when its net score reaches a predetermined threshold. Moreover, the pruning rate is obtained distinctly for each layer of the network. Results: Applying the suggested algorithm to a neural network model of a distillation column in a noisy environment demonstrates its effectiveness and applicability. Conclusion: The proposed method, which is inspired by connection competition and synaptic pruning in the human brain, enhances learning speed, preserves accuracy, and reduces costs due to its smaller network size. It also handles noisy data more efficiently by continuously assessing network connections