DLBench: a comprehensive experimental evaluation of deep learning frameworks

Author:

Elshawi Radwa,Wahab Abdul,Barnawi Ahmed,Sakr SherifORCID

Abstract

AbstractDeep Learning (DL) has achieved remarkable progress over the last decade on various tasks such as image recognition, speech recognition, and natural language processing. In general, three main crucial aspects fueled this progress: the increasing availability of large amount of digitized data, the increasing availability of affordable parallel and powerful computing resources (e.g., GPU) and the growing number of open source deep learning frameworks that facilitate and ease the development process of deep learning architectures. In practice, the increasing popularity of deep learning frameworks calls for benchmarking studies that can effectively evaluate and understand the performance characteristics of these systems. In this paper, we conduct an extensive experimental evaluation and analysis of six popular deep learning frameworks, namely, TensorFlow, MXNet, PyTorch, Theano, Chainer, and Keras, using three types of DL architectures Convolutional Neural Networks (CNN), Faster Region-based Convolutional Neural Networks (Faster R-CNN), and Long Short Term Memory (LSTM). Our experimental evaluation considers different aspects for its comparison including accuracy, training time, convergence and resource consumption patterns. Our experiments have been conducted on both CPU and GPU environments using different datasets. We report and analyze the performance characteristics of the studied frameworks. In addition, we report a set of insights and important lessons that we have learned from conducting our experiments.

Publisher

Springer Science and Business Media LLC

Subject

Computer Networks and Communications,Software

Reference55 articles.

1. Abadi, M., et al.: Tensorflow: a system for large-scale machine learning. OSDI 16, 265–283 (2016)

2. Abugabah, A., AlZubi, A.A., Al-Obeidat, F.N., Alarifi, A., Alwadain, A.: Data mining techniques for analyzing healthcare conditions of urban space-person lung using meta-heuristic optimized neural networks. Clust. Comput. 23(3), 1781–1794 (2020)

3. Awan, A.A., Subramoni, H., Panda, D.K.: An in-depth performance characterization of CPU-and GPU-based DNN training on modern architectures. In: Proceedings of the Machine Learning on HPC Environments, p. 8. ACM, (2017)

4. Bahrampour, S., Ramakrishnan, N., Schott, L., Shah, M.: Comparative study of caffe, neon, theano, and torch for deep learning (2016)

5. Bastien, F., Lamblin, P., Pascanu, R., Bergstra, J., Goodfellow, I., Bergeron, A., Bouchard, N., Warde-Farley, D., Bengio, Y.: Theano: new features and speed improvements. arXiv preprint arXiv:1211.5590 (2012)

Cited by 38 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Recent advances in Machine Learning based Advanced Driver Assistance System applications;Microprocessors and Microsystems;2024-09

2. Uma metodologia para a avaliação de desempenho e custos do treinamento de redes neurais em ambientes de nuvem;Anais do XXIII Workshop em Desempenho de Sistemas Computacionais e de Comunicação (WPerformance 2024);2024-07-21

3. Studying the Impact of TensorFlow and PyTorch Bindings on Machine Learning Software Quality;ACM Transactions on Software Engineering and Methodology;2024-07-13

4. State-of-the-Art Machine Learning Frameworks for Training or Inference on Business Process Dataset;2024 47th MIPRO ICT and Electronics Convention (MIPRO);2024-05-20

5. Enhanced Small Drone Detection Using Optimized YOLOv8 With Attention Mechanisms;IEEE Access;2024

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3