0% Complete
Home
/
13th International Conference on Computer and Knowledge Engineering
Enhancing Lighter Neural Network Performance with Layer-wise Knowledge Distillation and Selective Pixel Attention
Authors :
Siavash Zaravashan
1
Sajjad Torabi
2
Hesam Zaravashan
3
1- Iran University of Science and Technology
2- Part AI Research Center
3- Iran University of Science and Technology
Keywords :
Knowledge distillation،Image classification،Deep learning،Model compression،Computer vision
Abstract :
Deep neural networks have made a revolution in different areas of Artificial Intelligence such as image classification and object detection. However, most of these models are heavy and time-consuming at the inference stage. If we desire to use real-time applications, it's necessary to lighten them up. However, this could negatively affect the performance of models. One suggested approach for enhancing the capabilities of smaller models involves a technique called knowledge distillation. In this method, a less advanced model, referred to as the student model, learns from and emulates the more advanced teacher model to enhance its performance. This method has been proven as an effective technique to compress the model and increase its accuracy. Previous methods mostly focus on proposing feature transformation and loss functions between the same level’s features to improve the effectiveness. In this paper, we investigate the effect of connections between different layers of teacher and student networks and reveal their great importance. We have employed three different loss functions—attention-driven distillation, hierarchical context loss, and non-local loss—to enhance the classifier's overall performance. Extensive experiments on the CIFAR-100 dataset show that the proposed method has a great performance improvement. For the benefit of the research community, we will make the code for this study available on GitHub.
Papers List
List of archived papers
Distilling Knowledge from CNN-Transformer Models for Enhanced Human Action Recognition
Hamid Ahmadabadi - Omid Nejati Manzari - Ahmad Ayatollahi
Optimizing the controller placement problem in SDN with uncertain parameters with robust optimization
Mohammad Kazemi - AhmadReza Montazerolghaem
AL-YOLO: Accurate and Lightweight Vehicle and Pedestrian Detector in Foggy Weather
Behdad Sadeghian Pour - Hamidreza Mohammadi Jozani - Shahriar Baradaran Shokouhi
Intracranial Hemorrhage Classification using CBAM Attention Module and Convolutional Neural Networks
Parnian Rahimi - Marjan Naderan - Amir Jamshidnezhad - Shahram Rafie
To Transfer or Not To Transfer (TNT): Action Recognition in Still Image Using Transfer Learning
Ali Soltani Nezhad - Hojat Asgarian Dehkordi - Seyed Sajad Ashrafi - Shahriar Baradaran Shokouhi
DFIG-WECS Renewable Integration to the Grid and Stability Improvement through Optimal Damping Controller Design
Theophilus Ebuka Odoh - Aliyu Sabo - Hossien Shahinzadeh - Noor Izzri Abdul Wahab - Farshad Ebrahimi
Developing Convolutional Neural Networks using a Novel Lamarckian Co-Evolutionary Algorithm
Zaniar Sharifi - Khabat Soltanian - Ali Amiri
Automatic Generation of XACML Code using Model-Driven Approach
Athareh Fatemian - Bahman Zamani - Marzieh Masoumi - Mehran Kamranpour - Behrouz Tork Ladani - Shekoufeh Kolahdouz Rahimi
An Ensemble CNN for Brain Age Estimation based on Hippocampal Region Applicable to Alzheimer's Diagnosis
Zahra Qodrati - Seyedeh Masoumeh Taji - Habibollah Danyali - Kamran Kazemi
A Genetic-based Fusion Approach of Persian and Universal Phonetic results for Spoken Language Identification
Ashkan Moradi - Yasser Shekofteh - Saeed Zarei
more
Samin Hamayesh - Version 41.7.6