Alzubaidi Laith, Zhang Jinglan, Humaidi Amjad J, Al-Dujaili Ayad, Duan Ye, Al-Shamma Omran, Santamaría J, Fadhel Mohammed A, Al-Amidie Muthana, Farhan Laith
School of Computer Science, Queensland University of Technology, Brisbane, QLD 4000 Australia.
AlNidhal Campus, University of Information Technology & Communications, Baghdad, 10001 Iraq.
J Big Data. 2021;8(1):53. doi: 10.1186/s40537-021-00444-8. Epub 2021 Mar 31.
In the last few years, the deep learning (DL) computing paradigm has been deemed the Gold Standard in the machine learning (ML) community. Moreover, it has gradually become the most widely used computational approach in the field of ML, thus achieving outstanding results on several complex cognitive tasks, matching or even beating those provided by human performance. One of the benefits of DL is the ability to learn massive amounts of data. The DL field has grown fast in the last few years and it has been extensively used to successfully address a wide range of traditional applications. More importantly, DL has outperformed well-known ML techniques in many domains, e.g., cybersecurity, natural language processing, bioinformatics, robotics and control, and medical information processing, among many others. Despite it has been contributed several works reviewing the State-of-the-Art on DL, all of them only tackled one aspect of the DL, which leads to an overall lack of knowledge about it. Therefore, in this contribution, we propose using a more holistic approach in order to provide a more suitable starting point from which to develop a full understanding of DL. Specifically, this review attempts to provide a more comprehensive survey of the most important aspects of DL and including those enhancements recently added to the field. In particular, this paper outlines the importance of DL, presents the types of DL techniques and networks. It then presents convolutional neural networks (CNNs) which the most utilized DL network type and describes the development of CNNs architectures together with their main features, e.g., starting with the AlexNet network and closing with the High-Resolution network (HR.Net). Finally, we further present the challenges and suggested solutions to help researchers understand the existing research gaps. It is followed by a list of the major DL applications. Computational tools including FPGA, GPU, and CPU are summarized along with a description of their influence on DL. The paper ends with the evolution matrix, benchmark datasets, and summary and conclusion.
在过去几年中,深度学习(DL)计算范式被视为机器学习(ML)社区的黄金标准。此外,它已逐渐成为ML领域中使用最广泛的计算方法,从而在一些复杂的认知任务上取得了出色的成果,达到甚至超越了人类的表现。DL的一个优点是能够学习大量数据。DL领域在过去几年中发展迅速,并已被广泛用于成功解决各种传统应用。更重要的是,DL在许多领域,如网络安全、自然语言处理、生物信息学、机器人与控制以及医学信息处理等,都优于知名的ML技术。尽管已经有一些关于DL的最新技术综述的著作,但它们都只涉及DL的一个方面,这导致对其整体缺乏了解。因此,在本论文中,我们建议采用一种更全面的方法,以便提供一个更合适的起点,从而全面理解DL。具体而言,本综述试图对DL的最重要方面进行更全面的调查,包括该领域最近的那些增强内容。特别是,本文概述了DL的重要性,介绍了DL技术和网络的类型。然后介绍了使用最广泛的DL网络类型——卷积神经网络(CNN),并描述了CNN架构的发展及其主要特征,例如从AlexNet网络开始,到高分辨率网络(HR.Net)结束。最后,我们进一步提出挑战和建议的解决方案,以帮助研究人员了解现有的研究差距。随后列出了主要的DL应用。总结了包括FPGA、GPU和CPU在内的计算工具及其对DL的影响。论文最后给出了演进矩阵、基准数据集以及总结和结论。