Nazim Sadia, Alam Muhammad Mansoor, Rizvi Syed Safdar, Mustapha Jawahir Che, Hussain Syed Shujaa, Suud Mazliham Mohd
Malaysian Institute of Information Technology, Universiti Kuala Lumpur, Kuala Lumpur, Malaysia.
Department of Computer Science, Bahria University Islamabad, Islamabad, Pakistan.
PLoS One. 2025 May 28;20(5):e0318542. doi: 10.1371/journal.pone.0318542. eCollection 2025.
Artificial Intelligence (AI) is being integrated into increasingly more domains of everyday activities. Whereas AI has countless benefits, its convoluted and sometimes vague internal operations can establish difficulties. Nowadays, AI is significantly employed for evaluations in cybersecurity that find it challenging to justify their proceedings; this absence of accountability is alarming. Additionally, over the last ten years, the fractional elevation in malware variants has directed scholars to utilize Machine Learning (ML) and Deep Learning (DL) approaches for detection. Although these methods yield exceptional accuracy, they are also difficult to understand. Thus, the advancement of interpretable and powerful AI models is indispensable to their reliability and trustworthiness. The trust of users in the models used for cybersecurity would be undermined by the ambiguous and indefinable nature of existing AI-based methods, specifically in light of the more complicated and diverse nature of cyberattacks in modern times. The present research addresses the comparative analysis of an ensemble deep neural network (DNNW) with different ensemble techniques like RUSBoost, Random Forest, Subspace, AdaBoost, and BagTree for the best prediction against imagery malware data. It determines the best-performing model, an ensemble DNNW, for which explainability is provided. There has been relatively little study on explainability, especially when dealing with malware imagery data, irrespective of the fact that DL/ML algorithms have revolutionized malware detection. Explainability techniques such as SHAP, LIME, and Grad-CAM approaches are employed to present a complete comprehension of feature significance and local or global predictive behavior of the model over various malware categories. A comprehensive investigation of significant characteristics and their impact on the decision-making process of the model and multiple query point visualizations are some of the contributions. This strategy promotes advanced transparency and trustworthy cybersecurity applications by improving the comprehension of malware detection techniques and integrating explainable AI observations with domain-specific knowledge.
人工智能(AI)正日益融入日常活动的越来越多领域。虽然人工智能有无数好处,但其复杂且有时模糊的内部运作可能会带来困难。如今,人工智能在网络安全评估中被大量使用,这些评估发现很难为其程序进行辩护;这种缺乏问责制的情况令人担忧。此外,在过去十年中,恶意软件变种数量的小幅上升促使学者们利用机器学习(ML)和深度学习(DL)方法进行检测。尽管这些方法具有极高的准确性,但也难以理解。因此,可解释且强大的人工智能模型的发展对于其可靠性和可信度至关重要。现有基于人工智能的方法的模糊性和不确定性会削弱用户对用于网络安全的模型的信任,特别是考虑到现代网络攻击的性质更加复杂和多样。本研究针对集成深度神经网络(DNNW)与不同的集成技术(如RUSBoost、随机森林、子空间、AdaBoost和BagTree)进行了比较分析,以针对图像恶意软件数据进行最佳预测。它确定了性能最佳的模型——集成DNNW,并提供了其可解释性。关于可解释性的研究相对较少,尤其是在处理恶意软件图像数据时,尽管DL/ML算法已经彻底改变了恶意软件检测。采用诸如SHAP、LIME和Grad-CAM方法等可解释性技术来全面理解模型在各种恶意软件类别上的特征重要性以及局部或全局预测行为。对重要特征及其对模型决策过程的影响进行全面调查以及多个查询点可视化是其中的一些贡献。这种策略通过提高对恶意软件检测技术的理解,并将可解释的人工智能观察结果与特定领域知识相结合,促进了先进的透明度和可信赖的网络安全应用。