Neural Architecture Search For Efficient Model Deployment On Mobile Devices

Authors

  • Hamza Javed Assistant Professor, Department of Mathematics, International Islamic University Islamabad (IIUI), Islamabad, Pakistan Author

DOI:

https://doi.org/10.71465/mrcis182

Keywords:

Neural Architecture Search, Mobile AI, Edge Computing, Lightweight Models

Abstract

Neural Architecture Search (NAS) has transformed the automation of deep learning model design, enabling the discovery of optimized architectures tailored for specific hardware constraints. With the rapid growth of mobile and edge-AI applications, deploying efficient deep learning models on resource-constrained devices has become a pressing challenge. NAS provides a systematic approach to generating lightweight, low-latency, and energy-efficient neural networks suitable for smartphones, IoT devices, drones, and wearables. This article examines modern NAS techniques—such as evolutionary search, reinforcement-learning-based search, differentiable NAS (DARTS), and hardware-aware NAS—alongside the challenges of memory limitations, power consumption, and inference speed on mobile devices. Two graphs illustrate latency improvements and accuracy–efficiency trade-offs across NAS-generated models. The article concludes with future research directions, including adaptive NAS, federated NAS, and device-specific optimization frameworks.

References

Ahmad, N. R. (2025). AI-enabled public governance in developing states: Service delivery gains, accountability risks, and a practical risk-based regulatory model. https://doi.org/10.52152/wja5db40

Irk, E. (2025). From subsidies to statutory markets: Leadership, institutional entrepreneurship, and welfare governance reform. https://doi.org/10.52152/s59sjh53

Downloads

Published

2025-12-31