Next Article in Journal
Advanced Servo Control and AI Integration in 3-DoF Platforms for Enhanced Simulation Interactivity
Previous Article in Journal
Enhanced and Combined Representations in Extended Reality through Creative Industries
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Matching the Ideal Pruning Method with Knowledge Distillation for Optimal Compression

Department of Computer Vision, Institute of Cognitive Science, Osnabrück University, 49074 Osnabrück, Germany
*
Author to whom correspondence should be addressed.
Appl. Syst. Innov. 2024, 7(4), 56; https://doi.org/10.3390/asi7040056
Submission received: 13 November 2023 / Revised: 14 February 2024 / Accepted: 20 June 2024 / Published: 29 June 2024

Abstract

In recent years, model compression techniques have gained significant attention as a means to reduce the computational and memory requirements of deep neural networks. Knowledge distillation and pruning are two prominent approaches in this domain, each offering unique advantages in achieving model efficiency. This paper investigates the combined effects of knowledge distillation and two pruning strategies, weight pruning and channel pruning, on enhancing compression efficiency and model performance. The study introduces a metric called “Performance Efficiency” to evaluate the impact of these pruning strategies on model compression and performance. Our research is conducted on the popular datasets CIFAR-10 and CIFAR-100. We compared diverse model architectures, including ResNet, DenseNet, EfficientNet, and MobileNet. The results emphasize the efficacy of both weight and channel pruning in achieving model compression. However, a significant distinction emerges, with weight pruning showing superior performance across all four architecture types. We realized that the weight pruning method better adapts to knowledge distillation than channel pruning. Pruned models show a significant reduction in parameters without a significant reduction in accuracy.
Keywords: knowledge distillation; network efficiency; parameter reduction; unstructured pruning; structured pruning knowledge distillation; network efficiency; parameter reduction; unstructured pruning; structured pruning

Share and Cite

MDPI and ACS Style

Malihi, L.; Heidemann, G. Matching the Ideal Pruning Method with Knowledge Distillation for Optimal Compression. Appl. Syst. Innov. 2024, 7, 56. https://doi.org/10.3390/asi7040056

AMA Style

Malihi L, Heidemann G. Matching the Ideal Pruning Method with Knowledge Distillation for Optimal Compression. Applied System Innovation. 2024; 7(4):56. https://doi.org/10.3390/asi7040056

Chicago/Turabian Style

Malihi, Leila, and Gunther Heidemann. 2024. "Matching the Ideal Pruning Method with Knowledge Distillation for Optimal Compression" Applied System Innovation 7, no. 4: 56. https://doi.org/10.3390/asi7040056

Article Metrics

Back to TopTop