0% Complete
Home
/
14th International Conference on Computer and Knowledge Engineering
Distilled BERT Model In Natural Language Processing
Authors :
Yazdan Zandiye Vakili
1
Avisa Fallah
2
Hedieh Sajedi
3
1- School of Mathematics, Statistics and Computer Science, University of Tehran, Tehran, Iran
2- School of Mathematics, Statistics and Computer Science, University of Tehran, Tehran, Iran
3- School of Mathematics, Statistics and Computer Science, University of Tehran, Tehran, Iran
Keywords :
NLP،Machine Learning،Distillation،BERT،Transformers
Abstract :
This paper reviews the evolution of Natural Language Processing (NLP) models, focusing on the distillation techniques used to create efficient and compact versions of large models. Traditional NLP models laid the foundation but had limitations in scalability and contextual understanding. Transformer models like BERT revolutionized NLP but required significant computational resources. This review examines TinyBERT, DistilBERT, MobileBERT, and MiniLM, which balance size and performance through knowledge distillation. These distilled models maintain high performance while being suitable for deployment on resource-constrained devices, making advanced NLP capabilities accessible in real-world applications.
Papers List
List of archived papers
Using Deep Learning for Classification of Lung Cancer on CT Images in Ardabil Province
Mohammad Ali Javadzadeh Barzaki - Jafar Abdollahi - Mohammad Negaresh - Maryam Salimi - Hadi Zolfeghari - Mohsen Mohammadi - Asma Salmani - Rona Jannati - Firouz Amani
Supervised Contrastive Learning for Short Text Classification in Natural Language Processing
Mitra Esmaeili - Hamed Vahdat nejad
A Survey on Semi-Automated and Automated Approaches for Video Annotation
Samin Zare - Mehran Yazdi
Multi-Digit Handwritten Recognition: A CNN-LSTM Hybrid Approach with Wavelet Transforms
Amin Kazempour - Jafar Tanha
Recommending Popular Locations Based on Collected Trajectories
Mohammad Rabbani bidgoli - Saber Ziaei
Extracting Major Topics of COVID-19 Related Tweets
Faezeh Azizi - Hamed Vahdat-Nejad - Hamideh Hajiabadi - Mohammad Hossein Khosravi
Farsi Text in Scene: A new dataset
Ali Salmasi - Ehsanollah Kabir
A Systematic Embedded Software Design Flow for Robotic Applications
Navid Mahdian - Seyed-Hosein Attarzadeh-Niaki - Armin Salimi-Badr
Non-Negative Matrix Factorization improves Residual Neural Networks
Hojjat Moayed
Pruning and Mixed Precision Techniques for Accelerating Neural Network
Mahsa Zahedi - Mohammad Sediq Abazari Bozhgani - Abdorreza Savadi
more
Samin Hamayesh - Version 42.2.1