ICCK Transactions on Machine Intelligence
ISSN: 3068-7403 (Online)
Email: [email protected]

Submit Manuscript
Edit a Special Issue

TY - JOUR AU - Malik, Ishfaq Ahmad AU - Habib, Gousia PY - 2025 DA - 2025/11/13 TI - AI Enabled Resource-Constrained Computing Architectures for IoT Devices JO - ICCK Transactions on Machine Intelligence T2 - ICCK Transactions on Machine Intelligence JF - ICCK Transactions on Machine Intelligence VL - 1 IS - 3 SP - 138 EP - 147 DO - 10.62762/TMI.2025.225921 UR - https://www.icck.org/article/abs/TMI.2025.225921 KW - knowledge distillation KW - high-performance computing KW - transformers KW - DistilBERT KW - self-attention mechanism AB - Deep learning is a great success primarily because it encodes large amounts of data and manipulates billions of model parameters. Despite this, it is challenging to deploy these cumbersome deep models on devices with limited resources, such as mobile phones and embedded devices, due to the high computational complexity and the amount of storage required. Various techniques are available to compress and accelerate models for this purpose. Knowledge distillation is a novel technique for model compression and acceleration, which involves learning a small student model from a large teacher model. Then, that student network is fine-tuned on any downstream task to be applicable for resource-constrained applications. This paper explores various state-of-the-art model compression techniques, including knowledge distillation, for compressing large deep neural networks to make them deployable on resource-constrained devices. SN - 3068-7403 PB - Institute of Central Computation and Knowledge LA - English ER -
@article{Malik2025AI,
author = {Ishfaq Ahmad Malik and Gousia Habib},
title = {AI Enabled Resource-Constrained Computing Architectures for IoT Devices},
journal = {ICCK Transactions on Machine Intelligence},
year = {2025},
volume = {1},
number = {3},
pages = {138-147},
doi = {10.62762/TMI.2025.225921},
url = {https://www.icck.org/article/abs/TMI.2025.225921},
abstract = {Deep learning is a great success primarily because it encodes large amounts of data and manipulates billions of model parameters. Despite this, it is challenging to deploy these cumbersome deep models on devices with limited resources, such as mobile phones and embedded devices, due to the high computational complexity and the amount of storage required. Various techniques are available to compress and accelerate models for this purpose. Knowledge distillation is a novel technique for model compression and acceleration, which involves learning a small student model from a large teacher model. Then, that student network is fine-tuned on any downstream task to be applicable for resource-constrained applications. This paper explores various state-of-the-art model compression techniques, including knowledge distillation, for compressing large deep neural networks to make them deployable on resource-constrained devices.},
keywords = {knowledge distillation, high-performance computing, transformers, DistilBERT, self-attention mechanism},
issn = {3068-7403},
publisher = {Institute of Central Computation and Knowledge}
}
Portico
All published articles are preserved here permanently:
https://www.portico.org/publishers/icck/