0% Complete
Home
/
14th International Conference on Computer and Knowledge Engineering
Distilled BERT Model In Natural Language Processing
Authors :
Yazdan Zandiye Vakili
1
Avisa Fallah
2
Hedieh Sajedi
3
1- School of Mathematics, Statistics and Computer Science, University of Tehran, Tehran, Iran
2- School of Mathematics, Statistics and Computer Science, University of Tehran, Tehran, Iran
3- School of Mathematics, Statistics and Computer Science, University of Tehran, Tehran, Iran
Keywords :
NLP،Machine Learning،Distillation،BERT،Transformers
Abstract :
This paper reviews the evolution of Natural Language Processing (NLP) models, focusing on the distillation techniques used to create efficient and compact versions of large models. Traditional NLP models laid the foundation but had limitations in scalability and contextual understanding. Transformer models like BERT revolutionized NLP but required significant computational resources. This review examines TinyBERT, DistilBERT, MobileBERT, and MiniLM, which balance size and performance through knowledge distillation. These distilled models maintain high performance while being suitable for deployment on resource-constrained devices, making advanced NLP capabilities accessible in real-world applications.
Papers List
List of archived papers
Bipartite link prediction improvement using the effective utilization of edge betweenness centrality
Sadegh Sulaimany Sulaimany - Yasin Amini
Adaptive Channel Estimation for MIMO-OFDM Systems in Impulsive Noise Environments
Mojtaba Hajiabadi
Multi-Fusion Ensemble CNN for Drug–Target Binding Affinity Prediction Using Transformer-Based Molecular and Protein Representations
Betsabeh Tanoori
Predicting cascading failure with machine learning methods in the interdependent networks
Mohamad Hossein Maghsoodi - Mohamad Khansari
ExaAEC: A New Multi-label Emotion Classification Corpus in Arabic Tweets
Saeed Sarbazi-Azad - Ahmad Akbari - Mohsen Khazeni
Intelligent Resource Collision Management for Cellular Vehicular Systems Using Software-Defined Networking
Mohammad Kazemiesfeh - Neda Moghim - Ahmadreza Montazerolghaem
Plant Disease Detection Using Dynamic Knowledge Distillation and Attention Mechanism
Mohammad Ghasemi Arian - Mohammad Hossein Yaghmaee Moghaddam
Improving Soft Error Reliability of FPGA-based Deep Neural Networks with Reduced Approximate TMR
Anahita Hosseinkhani - Behnam Ghavami
Driving Violation Detection Using Vehicle Data and Environmental Conditions
Masood Ghasemi - Mahmood Fathy - Mohammad Shahverdy
Automating Theory of Mind Assessment with a LLaMA-3-Powered Chatbot: Enhancing Faux Pas Detection in Autism
Avisa Fallah - Ali Keramati - Mohammad Ali Nazari - Fatemeh Sadat Mirfazeli
more
Samin Hamayesh - Version 42.7.0