0% Complete
Home
/
13th International Conference on Computer and Knowledge Engineering
Enhancing Lighter Neural Network Performance with Layer-wise Knowledge Distillation and Selective Pixel Attention
Authors :
Siavash Zaravashan
1
Sajjad Torabi
2
Hesam Zaravashan
3
1- Iran University of Science and Technology
2- Part AI Research Center
3- Iran University of Science and Technology
Keywords :
Knowledge distillation،Image classification،Deep learning،Model compression،Computer vision
Abstract :
Deep neural networks have made a revolution in different areas of Artificial Intelligence such as image classification and object detection. However, most of these models are heavy and time-consuming at the inference stage. If we desire to use real-time applications, it's necessary to lighten them up. However, this could negatively affect the performance of models. One suggested approach for enhancing the capabilities of smaller models involves a technique called knowledge distillation. In this method, a less advanced model, referred to as the student model, learns from and emulates the more advanced teacher model to enhance its performance. This method has been proven as an effective technique to compress the model and increase its accuracy. Previous methods mostly focus on proposing feature transformation and loss functions between the same level’s features to improve the effectiveness. In this paper, we investigate the effect of connections between different layers of teacher and student networks and reveal their great importance. We have employed three different loss functions—attention-driven distillation, hierarchical context loss, and non-local loss—to enhance the classifier's overall performance. Extensive experiments on the CIFAR-100 dataset show that the proposed method has a great performance improvement. For the benefit of the research community, we will make the code for this study available on GitHub.
Papers List
List of archived papers
Span-prediction of Unknown Values for Long-sequence Dialogue State Tracking
Marzieh Naghdi Dorabati - Reza Ramezani - Mohammad Ali Nematbakhsh
Multi-Layered Defense Against Modern Phishing: A Dual-Sandbox and CDR Approach
Mahdi Seyfipoor - Mohammad Mahdi Eskandari
Implementation of a Low-Overhead 2-Bit Parity-Preserving Reversible Vedic Multiplier for Quantum Architectures
Shekoofeh Moghimi - Negin Mashayekhi - Mohammad Reza Reshadinezhad
A Synergistic Hybrid Architecture with Residual Attention and Mixture-of-Experts for Robust Hour-Ahead Forex Forecasting
Alireza Abbaszadeh - Seyyed Abed Hosseini - Mohammad Reza Akbarzadeh Totonchi
T-Rank: Graph Data Analytics for Urban Traffic Modeling
Alireza Safarpour - Iman Gholampour - Amirhossain Aghazadeh Fard - Seyed Mohammad Karbasi
BioBERT-based SNP-traits Associations Extraction from Biomedical Literature
Mohammad Dehghani - Behrouz Bokharaeian - Zahra Yazdanparast
Depression Diagnosis Using Optimization of Nonlinear EEG Features Based on Parametric Learning Tactics
Ali Asadi Zeidabadi - Melika Changizi - Mahdi Zolfagharzadeh Kermani - Sara Bargi Barkouk
Machine Learning-Driven Prediction of Anti-Alzheimer Drug Efficacy Using PubChem Molecular Fingerprints
Mohammad Javad Sadeghi - Mohammad Javad Nemati - AliAsghar Zare - Mohammadreza Shams
A Comprehensive Approach to SMS Spam Filtering Integrating Embedded and Statistical Features
Shaghayegh Hosseinpour - Mohammad Reza Keyvanpour
Averting Mode Collapse for Generative Zero-Shot Learning
Shayan Ramazi - Setare Shabani
more
Samin Hamayesh - Version 42.7.0