Skip to content
2000
Volume 18, Issue 10
  • ISSN: 2352-0965
  • E-ISSN: 2352-0973

Abstract

Introduction

Convolutional Neural Networks (CNNs) and deep learning algorithms have significantly advanced image processing and classification. This study compares three CNN architectures VGG-16, ResNet-50, and ResNet-18, and evaluates their performance on the CIFAR-100 and MNIST datasets. Training time is prioritized as a critical metric, along with test accuracy and training loss, under varying hardware configurations (NVIDIA-GPU and Intel-CPU).

Methods

Experiments were conducted on an Ubuntu 22.04 system using the PyTorch framework. The hardware configurations included an NVIDIA GeForce GTX 1660 Super GPU and an Intel Core i5-10400 CPU. The CIFAR-100 dataset, containing 60,000 color images across 100 classes, and the MNIST dataset, comprising 70,000 grayscale images, were used for benchmarking.

Results

The results highlight the superior efficiency of GPUs, with training times reduced by up to 10x compared to CPUs. For CIFAR-100, VGG-16 required 13,000 seconds on the GPU versus 130,000 seconds on the CPU, while ResNet-18, the most time-efficient model, completed training in 150 seconds on the GPU and 1,740 seconds on the CPU. ResNet-50 achieved the highest test accuracy (~80%) on CIFAR-100. On MNIST, ResNet-18 was the most efficient, with training times of 185 seconds on the GPU and 22,000 seconds on the CPU.

Discussion

This study highlights the clear advantage of GPUs over CPUs in reducing training times, particularly for complex models such as VGG-16 and ResNet-50. ResNet-50 achieved the highest accuracy, while ResNet-18 was the most time efficient. However, the use of simpler datasets (MNIST and CIFAR-100) may not fully capture the complexity of the real world.

Conclusion

This study emphasizes the importance of hardware selection for deep learning workflows. Using the CIFAR-100 and MNIST datasets, we demonstrated that GPUs significantly outperform CPUs, achieving up to a 10x reduction in training time while maintaining competitive accuracy. Among the architectures tested, ResNet-50 delivered the highest test accuracy (~80%) on CIFAR-100, demonstrating superior feature extraction capabilities compared to VGG-16 and ResNet-18. Meanwhile, ResNet-18 proved to be the most time-efficient architecture, completing training in 150 seconds on the GPU, a significant improvement over VGG-16's 13,000 seconds. These results highlight the advantage of residual connections in reducing training complexity and achieving higher performance. The results underscore the critical role of both architecture selection and hardware optimization in advancing deep learning workflows.

Loading

Article metrics loading...

/content/journals/raeeng/10.2174/0123520965348453250226080233
2025-01-16
2026-01-01
Loading full text...

Full text loading...

References

  1. SimonyanK. ZissermanA. Very deep convolutional networks for large-scale image recognitionarXiv preprint arXiv:1409.1556.201416
    [Google Scholar]
  2. HeK. ZhangX. RenS. SunJ. Deep residual learning for image recognitionProc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR) 2016+, pp. 770-778
    [Google Scholar]
  3. ElshawiR. WahabA. BarnawiA. SakrS. DLBench: A comprehensive experimental evaluation of deep learning frameworks.Cluster Comput.20212432017203810.1007/s10586‑021‑03240‑4
    [Google Scholar]
  4. RiahiA. SavadiA. NaghibzadehM. Comparison of analytical and ML-based models for predicting CPU–GPU data transfer time.Computing202010292099211610.1007/s00607‑019‑00780‑x
    [Google Scholar]
  5. JayasimhanA. PabithaP. A comparison between CPU and GPU for image classification using convolutional neural networksInternational Conference on Communication, Computing and Internet of Things (IC3IoT) IEEE, 2022. pp. 1-4.10.1109/IC3IOT53935.2022.9767990
    [Google Scholar]
  6. TodiA. NarulaN. SharmaM. GuptaU. ConvNext: A contemporary architecture for convolutional neural networks for image classificationProc. 2023 Int. Conf. Innovative Sustainable Comput. Technol. (CISCT) 2023, PP. 1-6.10.1109/CISCT57197.2023.10351320
    [Google Scholar]
  7. AzadR. Advances in medical image analysis with vision transformers: A comprehensive review.Med. Image Ana.202391103000
    [Google Scholar]
  8. MondalA. SahaP. Classification of brain tumor from magnetic resonance imaging using vision transformers ensembling,” Biomed. Signal Process. ControlOnline20227410348610.1016/j.bspc.2021.103486
    [Google Scholar]
  9. RoyS.K. Spectral–spatial morphological attention transformer for hyperspectral image classification., vol. 61, IEEE Trans. Geosci. Remote Sens.202311510.1109/TGRS.2023.3242346
    [Google Scholar]
  10. RasheedH. Efficient CNN architectures for edge computing environments.Open Comp. Sci. J.2023152201215
    [Google Scholar]
  11. BukhariT.A. Review of lightweight deep learning architectures for real-time object detection.Cur. Arti. Inte. Res.2022183219234
    [Google Scholar]
  12. KrizhevskyA. HintonG. Learning multiple layers of features from tiny images
    [Google Scholar]
  13. LiuZ. ConvNext: Revisiting ResNets for scalabilityProc. IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR)202219.
    [Google Scholar]
  14. TouvronT. Going deeper with image transformersProc. IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR)2023110.
    [Google Scholar]
  15. LeCunY. The MNIST database of handwritten digits.1998Available from: http://yann.lecun.com/exdb/mnist/
    [Google Scholar]
  16. LecunY. BottouL. BengioY. HaffnerP. Gradient-based learning applied to document recognition.Proc. IEEE199886112278232410.1109/5.726791
    [Google Scholar]
  17. CireganD. MeierU. SchmidhuberJ. Multi-column deep neural networks for image classificationProc. 2012 IEEE Conf. Computer Vision and Pattern Recognition. (CVPR) IEEE201210.1109/CVPR.2012.6248110
    [Google Scholar]
  18. LecunY. Handwritten digit recognition with a back-propagation network.Advances in Neural Information Processing Systems (NeurIPS)New Orleans1990396404
    [Google Scholar]
  19. KrizhevskyA. SutskeverI. HintonG.E. Imagenet classification with deep convolutional neural networks Adv. Neural Inf. Process. Syst.NIPS, 20122568490
    [Google Scholar]
  20. AlexA. Exploring the limitations of deep learning for robust image recognitionProc. IEEE Int. Conf. Comput. Vision (ICCV). 2012, pp. 99–106.
    [Google Scholar]
  21. SzegedyC. Going deeper with convolutionsProc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR) 2015, pp. 1–9.
    [Google Scholar]
  22. CholletF. Xception: Deep learning with depthwise separable convolutionsProc. IEEE Conf. Comput. Vision and Pattern Recognit. 2017, pp. 1251-1258.10.1109/CVPR.2017.195
    [Google Scholar]
  23. HuangG. LiuZ. van der MaatenL. WeinbergerK.Q. Densely connected convolutional networksProc. IEEE Conf. Comput. Vision and Pattern Recognit. 2017, pp. 4700–4708.10.1109/CVPR.2017.243
    [Google Scholar]
  24. HowardA.G. MobileNets: Efficient convolutional neural networks for mobile vision applicationsarXiv:1704.04861201716
    [Google Scholar]
  25. TanM. LeQ. EfficientNet: Rethinking model scaling for convolutional neural networksProc. Int. Conf. Mach. Learn. 2019, pp. 6105-6114.
    [Google Scholar]
  26. TouvronH. Training data-efficient image transformers and distillation through attentionProc. Int. Conf. Mach. Learn. 2021, pp. 10347-10357.
    [Google Scholar]
  27. NVIDIA, GeForce GTX 1660 Super2024Available from: http://www.nvidia.com/en-us/geforce/graphics-cards/gtx-1660-super/
    [Google Scholar]
  28. NVIDIA GeForce GTX 1660 SuperTechPowerUp2019Available from: www.techpowerup.com/gpu-specs/geforce-gtx-1660-super.c3450
    [Google Scholar]
  29. KrizhevskyA. GPU acceleration for deep learning: Analysis and insights.Adv. Mach. Lear.20221101118
    [Google Scholar]
  30. NVIDIA GeForce GTX 1660 Super ReviewTom's Hardware2019Available from: www.tomshardware.com/reviews/nvidia-geforce-gtx-1660-super-review/
    [Google Scholar]
  31. SharmaG. Comparative evaluation of GPUs for machine learning workloadsProc. Int. Conf. Comput. Arch. 2023, pp. 123-136.
    [Google Scholar]
  32. IntelIntel Core i5-10400 ProcessorIntel Official Website.2024Available from: www.intel.com/content/www/us/en/products/processors/core/i5-processors/i5-10400.html
    [Google Scholar]
  33. Intel Core i5-10400 ReviewTechSpot2020Available from: www.techspot.com/review/2011-intel-core-i5-10400/
    [Google Scholar]
  34. Intel Core i5-10400: Everything You Need to KnowTom's Hardware2020Available from: www.tomshardware.com/review/intel-core-i5-10400-review/
    [Google Scholar]
  35. BishopC.M. NasrabadiN.M. Pattern Recognition and Machine Learning.SpringerNew York20064738.4
    [Google Scholar]
  36. BrownleeJ. Deep Learning for Computer Vision: Image Classification, Object Detection, and Face Recognition in Python.Mach. Lear. Mast.y2019117
    [Google Scholar]
  37. OliverA. Realistic evaluation of deep semi-supervised learning algorithmsProc. NeurIPS20183132353246
    [Google Scholar]
  38. KohaviR. A study of cross-validation and bootstrap for accuracy estimation and model selectionProc. Int. Joint Conf. Artif. Intel.14199511371143
    [Google Scholar]
  39. SrivastavaN. Dropout: A simple way to prevent neural networks from overfitting.J. Mach. Learn. Res.20141519291958
    [Google Scholar]
  40. GayathriP. Exploring the potential of VGG-16 architecture for accurate brain tumor detection using deep learningJ. Comput., Mech. Manag.2023221710.57159/gadl.jcmm.2.2.23056
    [Google Scholar]
  41. SinghT. Early Stage Lung Cancer Detection Using Deep LearningMIT Art, Design and Technology School of Computing International Conference (MITADTSoCiCon). 2024, pp. 1–6.10.1109/MITADTSoCiCon60330.2024.10575345
    [Google Scholar]
  42. WangZ. Edge AI: On-demand accelerating deep neural network inference via edge computing.IEEE Trans. Neural Netw. Learn. Syst.20203193305331610.1109/TNNLS.2020.2966053 31613785
    [Google Scholar]
  43. JouppiN.P. In-datacenter Performance Analysis of a Tensor Processing Unit.Proc. 44th Annu. Int. Symp. Comput. Archit.Toronto, ON, Canada201711210.1145/3079856.3080246
    [Google Scholar]
  44. RussakovskyO. DengJ. SuH. KrauseJ. SatheeshS. MaS. HuangZ. KarpathyA. KhoslaA. BernsteinM. BergA.C. Fei-FeiL. ImageNet large scale visual recognition challenge.Int. J. Comput. Vis.2015115321125210.1007/s11263‑015‑0816‑y
    [Google Scholar]
  45. LinT. Microsoft COCO: Common objects in contextProc. IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR). 2014, pp. 748–755.10.1007/978‑3‑319‑10602‑1_48
    [Google Scholar]
  46. SrinivasanP.R. MRI scan classification using deep learning.J. Med. Imag.20182921910.1117/1.JMI.6.2.025001
    [Google Scholar]
  47. VijayaB.V.K. Hyperspectral image classification with deep learning techniquesProc. IEEE Int. Conf. Geosci. Remote Sens. (IGARSS). 2019, pp. 5313–5316.10.1109/IGARSS.2019.8898119
    [Google Scholar]
  48. DosovitskiyA. An image is worth 16x16 words: Transformers for image recognition at scalearXiv preprint, arXiv:2010.119292020116
    [Google Scholar]
  49. IoffeS. SzegedyC. Batch normalization: Accelerating deep network training by reducing internal covariate shiftProc. Int. Conf. Mach. Learn. 2015, pp. 448-456.
    [Google Scholar]
/content/journals/raeeng/10.2174/0123520965348453250226080233
Loading
/content/journals/raeeng/10.2174/0123520965348453250226080233
Loading

Data & Media loading...


  • Article Type:
    Research Article
Keyword(s): CNN architecture; CPU; deep learning; GPU; gpu-nvidia; image classification; training time
This is a required field
Please enter a valid email address
Approval was a Success
Invalid data
An Error Occurred
Approval was partially successful, following selected items could not be processed due to error
Please enter a valid_number test