Journal of Intelligent Systems and Internet of Things

Journal DOI

https://doi.org/10.54216/JISIoT

Submit Your Paper

2690-6791ISSN (Online) 2769-786XISSN (Print)

Volume 18 , Issue 1 , PP: 114-125, 2026 | Cite this article as | XML | Html | PDF | Full Length Article

Optimizing Neural Network Architectures with TensorFlow and Keras for Scalable Deep Learning

Muna Al-Saadi 1 , Bushra Al-Saadi 2 , Dheyauldeen Ahmed Farhan 3 , Oday Ali Hassen 4 *

  • 1 University of Information Technology and Communications (UoITC), Baghdad, Iraq - (muna.alsaadi75@gmail.com)
  • 2 University of Information Technology and Communications (UoITC), Baghdad, Iraq - (bysalsaadi@gmail.com)
  • 3 Department of Computer Science, University of Al Maarif, Al-Anbar, 31001, Iraq - (dheyauldeen.farhan@uoa.edu.iq)
  • 4 Ministry of Education, Wasit Education Directorate, Iraq; Computer Department, College of Education for Pure Sciences, Wasit University, 52001 Al-Kut, Wasit, Iraq - (odayali@uowasit.edu.iq)
  • Doi: https://doi.org/10.54216/JISIoT.180108

    Received: March 12, 2025 Revised: May 21, 2025 Accepted: July 04, 2025
    Abstract

    Deep studying architectures face fundamental demanding situations in balancing overall performance optimization, computational scalability, and operational interpretability. Current strategies show off an essential fragmentation: neural architecture search (NAS) techniques perform independently of interpretability requirements, while scalability answers remain detached from structure optimization pipelines. This disconnect hinders the improvement of a unified workflow from architecture layout to interpretable deployment. We endorse DeepOptiFrame, a TensorFlow/Keras-primarily based Python framework that combines three middle capabilities: (1) superior optimization algorithms (BOHB, Hyperband) with useful resource-restrained multi-objective search, (2) distributed training acceleration across GPU/GPU clusters via Horovod integration and blended-precision strategies, and (3) GPU-increased interpretability gear (SHAP, LIME) incorporated without delay into the education pipeline. Our framework demonstrates large experimental improvements: a 15-20% accuracy growth at the CIFAR-a hundred and ImageNet benchmarks compared to today's baselines, a 65% education speedup whilst scaled to eight GPUs with close to-linear performance, and a 30% development in interpretability reliability, as measured via the Mean Confidence Decrease metric. This implementation additionally reduces reminiscence intake via forty% throughout gradient checkpoints even as keeping numerical balance. These advances establish a new paradigm for coherent deep learning development, simultaneously improving overall performance, scalability, and transparency inside unified workflow surroundings.

    Keywords :

    Neural Architecture Search , Explainable AI , Distributed Deep Learning , Model Optimization , Interpretability Metrics

    References

    References

    [1]       D. Ketseas, "Stochastic Response of an Airfoil and Its Effects on Lco’s Behavior Under Stall Flutter Regime," Int. J. Math., Stat. Comput. Sci., vol. 2, pp. 168–172, 2024. doi: 10.59543/ijmscs.v2i.8663.

    [2]       Y. Kuvayskova and A. Nemykin, "Neural Network Architecture Search Algorithm for Technical Object State Prediction," in 2025 Int. Russian Smart Industry Conf. (SmartIndustryCon), IEEE, Mar. 2025, pp. 675–680.

    [3]       H. Lan, "Device Placement Optimization with Deep Reinforcement Learning," University of Toronto, Toronto, ON, Canada, 2023. Accessed: Jun. 05, 2025.

    [4]       C. Min, G. Liao, G. Wen, Y. Li, and X. Guo, "Ensemble Interpretation: A Unified Method for Interpretable Machine Learning," arXiv: 2312.06255, 2023.

    [5]       G. De Bernardi, S. Narteni, E. Cambiaso, and M. Mongelli, "Rule-Based Out-of-Distribution Detection," IEEE Trans. Artif. Intell., vol. 5, no. 6, pp. 2627–2637, Jun. 2024.

    [6]       Z. Lu, R. Cheng, Y. Jin, K. C. Tan, and K. Deb, "Neural architecture search as multiobjective optimization benchmarks: Problem formulation and performance assessment," IEEE Trans. Evol. Comput., vol. 28, no. 2, pp. 323–337, 2023.

    [7]       K. Zhou, X. Huang, Q. Song, R. Chen, and X. Hu, "Auto-GNN: Neural architecture search of graph neural networks," Front. Big Data, vol. 5, p. 1029307, 2022.

    [8]       S. Xue et al., "IDARTS: Interactive Differentiable Architecture Search," in Proc. IEEE Int. Conf. Computer Vision, 2021, pp. 1143–1152.

    [9]       P. Ren et al., "A comprehensive survey of neural architecture search: Challenges and solutions," ACM Comput. Surv., vol. 54, no. 4, pp. 1–34, 2021.

    [10]    S. Xiao, B. Zhao, and D. Liu, "Semi-supervised accuracy predictor-based multi-objective neural architecture search," Neurocomputing, vol. 609, p. 128472, Dec. 2024.

    [11]    A. Géron, Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow, O’Reilly Media, Inc., 2019. Accessed: Jun. 05, 2025. [Online]. Available: https://www.oreilly.com/library/view/hands-on-machine-learning/9781492032632/

    [12]    A. R. Menon, U. Menon, and K. Ahirwar, "Ravnest: Decentralized Asynchronous Training on Heterogeneous Devices," arXiv: 2401.01728, 2024.

    [13]    D. Narayanan, A. Phanishayee, K. Shi, X. Chen, and M. Zaharia, "Memory-efficient pipeline-parallel DNN training," in Proc. Int. Conf. Machine Learning, PMLR, Jul. 2021, pp. 7937-7947.

    [14]    A. Hassen et al., "Realistic Smile Expression Recognition Approach Using Ensemble Classifier with Enhanced Bagging," Comput. Mater. Continua, vol. 70, no. 2, pp. 123-138, 2022.

    [15]    X. Wan, B. Ru, P. M. Esparanca, and F. M. Carlucci, "Approximate Neural Architecture Search via Operation Distribution Learning," in Proc. 2022 IEEE/CVF Winter Conf. Applications of Computer Vision (WACV), 2022, pp. 3545–3554.

    [16]    W. Xu, "Efficient distributed image recognition algorithm of deep learning framework TensorFlow," J. Phys. Conf. Ser., vol. 2066, no. 1, p. 12070, 2021.

    [17]    R. Hesse, S. Schaub-Meyer, and S. Roth, "Fast Axiomatic Attribution for Neural Networks," in NIPS’21: Proc. 35th Int. Conf. Neural Information Processing Systems, 2021, pp. 19513–19524. Accessed: Jun. 05, 2025. [Online]. Available: https://dl.acm.org/doi/10.5555/3540261.3541754

    [18]    I. Cik, A. D. Rasamoelina, M. Mach, and P. Sincak, "Explaining Deep Neural Network using Layer-wise Relevance Propagation and Integrated Gradients," in SAMI 2021 - IEEE 19th World Symp. Applied Machine Intelligence and Informatics, IEEE, 2021, pp. 381–386.

    [19]    J. Pfau, A. T. Young, J. Wei, M. L. Wei, and M. J. Keiser, "Robust Semantic Interpretability: Revisiting Concept Activation Vectors," arXiv: 2104.02768, 2021.

    [20]    A. Agiollo, G. Ciatto, and A. Omicini, "Shallow2Deep: Restraining Neural Networks Opacity through Neural Architecture Search," in Lecture Notes in Computer Science, vol. 12688, Cham: Springer, 2021, pp. 63–82.

    [21]    S. R. Islam, W. Eberle, S. K. Ghafoor, and M. Ahmed, "Explainable Artificial Intelligence Approaches: A Survey," arXiv: 2101.09429, 2021.

    [22]    C. Rudin, C. Chen, Z. Chen, H. Huang, L. Semenova, and C. Zhong, "Interpretable machine learning: Fundamental principles and 10 grand challenges," Stat. Surv., vol. 16, pp. 1–85, 2022.

    [23]    N. Klyuchnikov et al., "NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing," IEEE Access, vol. 10, pp. 45736–45747, 2022.

    [24]    Microsoft, "Neural Network Intelligence: An open-source AutoML toolkit," in Proc. 26th ACM SIGKDD Int. Conf. Knowledge Discovery Data Mining, 2025. Accessed: Jun. 05, 2025.

    [25]    P. A. Schirmer and I. Mporas, "Non-Intrusive Load Monitoring: A Review," IEEE Trans. Smart Grid, vol. 14, no. 1, pp. 769–784, Jan. 2023.

    [26]    L. P. Swaminatha Rao and S. Jaganathan, "Hyperparameter Optimization Using Budget-Constrained BOHB for Traffic Forecasting," in Lecture Notes in Networks and Systems, Singapore: Springer, 2024, pp. 225–240.

    [27]    V. Geraeinejad, S. Sinaei, M. Modarressi, and M. Daneshtalab, "RoCo-NAS: Robust and Compact Neural Architecture Search," in Proc. Int. Joint Conf. Neural Networks, IEEE, 2021, pp. 1–8.

    [28]    M. Dorrich, M. Fan, and A. M. Kist, "Impact of Mixed Precision Techniques on Training and Inference Efficiency of Deep Neural Networks," IEEE Access, vol. 11, pp. 57627–57634, 2023.

    [29]    A. Wollek et al., "German CheXpert Chest X-ray Radiology Report Labeler," RoFo Fortschritte auf dem Gebiet der Rontgenstrahlen und der Bildgebenden Verfahren, vol. 196, no. 09, pp. 956–965, 2023.

    [30]    A. Audevart, K. Banachewicz, and L. Massaron, Machine Learning Using TensorFlow Cookbook: Create Powerful Machine Learning Algorithms with TensorFlow, Packt Publishing Ltd, 2021.

    [31]    S. R. Dubey, S. K. Singh, and B. B. Chaudhuri, "Activation functions in deep learning: A comprehensive survey and benchmark," Neurocomputing, vol. 503, pp. 92–108, 2022.

    [32]    M. Shafiq and Z. Gu, "Deep Residual Learning for Image Recognition: A Survey," Appl. Sci., vol. 12, no. 18, p. 8972, 2022.

    [33]    M. Tan and Q. Le, "EfficientNetV2: Smaller Models and Faster Training," in Proc. 38th Int. Conf. Machine Learning, M. Meila and T. Zhang, Eds., vol. 139, PMLR, Jun. 2021, pp. 10096–10106.

    [34]    S. Singhal et al., "Sentiment Analysis on Amazon Reviews of Mobile Phones using Machine Learning," Technology, vol. 15, p. 19.

    [35]    A. A. Ismail, H. Corrada Bravo, and S. Feizi, "Improving deep learning interpretability by saliency guided training," in Adv. Neural Inf. Process. Syst., vol. 34, pp. 26726-26739, 2021.

    [36]    Q. Jin et al., "F8Net: Fixed-Point 8-Bit Only Multiplication for Network Quantization," arXiv: 2202.05239, 2022.

    [37]    A. Paleyes, R. G. Urma, and N. D. Lawrence, "Challenges in Deploying Machine Learning: A Survey of Case Studies," ACM Comput. Surv., vol. 55, no. 6, pp. 1–29, 2022.

    [38]    Z. Liu et al., "Neural Architecture Search on Efficient Transformers and beyond," arXiv: 2207.13955, Jul. 2022.

    Cite This Article As :
    Al-Saadi, Muna. , Al-Saadi, Bushra. , Ahmed, Dheyauldeen. , Ali, Oday. Optimizing Neural Network Architectures with TensorFlow and Keras for Scalable Deep Learning. Journal of Intelligent Systems and Internet of Things, vol. , no. , 2026, pp. 114-125. DOI: https://doi.org/10.54216/JISIoT.180108
    Al-Saadi, M. Al-Saadi, B. Ahmed, D. Ali, O. (2026). Optimizing Neural Network Architectures with TensorFlow and Keras for Scalable Deep Learning. Journal of Intelligent Systems and Internet of Things, (), 114-125. DOI: https://doi.org/10.54216/JISIoT.180108
    Al-Saadi, Muna. Al-Saadi, Bushra. Ahmed, Dheyauldeen. Ali, Oday. Optimizing Neural Network Architectures with TensorFlow and Keras for Scalable Deep Learning. Journal of Intelligent Systems and Internet of Things , no. (2026): 114-125. DOI: https://doi.org/10.54216/JISIoT.180108
    Al-Saadi, M. , Al-Saadi, B. , Ahmed, D. , Ali, O. (2026) . Optimizing Neural Network Architectures with TensorFlow and Keras for Scalable Deep Learning. Journal of Intelligent Systems and Internet of Things , () , 114-125 . DOI: https://doi.org/10.54216/JISIoT.180108
    Al-Saadi M. , Al-Saadi B. , Ahmed D. , Ali O. [2026]. Optimizing Neural Network Architectures with TensorFlow and Keras for Scalable Deep Learning. Journal of Intelligent Systems and Internet of Things. (): 114-125. DOI: https://doi.org/10.54216/JISIoT.180108
    Al-Saadi, M. Al-Saadi, B. Ahmed, D. Ali, O. "Optimizing Neural Network Architectures with TensorFlow and Keras for Scalable Deep Learning," Journal of Intelligent Systems and Internet of Things, vol. , no. , pp. 114-125, 2026. DOI: https://doi.org/10.54216/JISIoT.180108