Optimizing Training Time in Deep Learning Models Using Distributed Computing Techniques
DOI:
https://doi.org/10.55145/ajest.2026.05.01.009Keywords:
Distributed Deep Learning, Convolutional Neural Networks, Data Parallelism, Model Scalability, Computational EfficiencyAbstract
Training deep learning models is often a time-consuming process, especially when hardware resources are tight. This bottleneck slows down experimentation and increases development costs. The primary goal of this study was to investigate whether distributed computing could cut down on training time while keeping model accuracy intact. To test this, we trained a Convolutional Neural Network (CNN) on the CIFAR-10 dataset using two different configurations: a standard single-device setup and a distributed synchronous setup powered by TensorFlow’s MirroredStrategy. To ensure a fair comparison, both trials used the exact same architecture, hyperparameters, and preprocessing. The results were clear: distributed training reduced the total time by about 19.5%, with validation accuracy remaining nearly identical to the single-device approach. These findings suggest that distributed training is a practical way to accelerate deep learning workflows without sacrificing performance, serving as an efficient solution even in environments with limited hardware.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 Nasiba M. Abdulkarim, Lozan M. Abdulrahman, Taha A. Ababakar, Omar M. Ahmed

This work is licensed under a Creative Commons Attribution 4.0 International License.