Sadeghi Koosha, Banerjee Ayan, Gupta Sandeep K S
IMPACT lab (http://impact.asu.edu/), CIDSE, Arizona State University, Tempe, Arizona, USA, 85281.
IEEE Trans Emerg Top Comput Intell. 2020 Aug;4(4):450-467. doi: 10.1109/tetci.2020.2968933. Epub 2020 May 25.
Machine Learning (ML) algorithms, specifically supervised learning, are widely used in modern real-world applications, which utilize Computational Intelligence (CI) as their core technology, such as autonomous vehicles, assistive robots, and biometric systems. Attacks that cause misclassifications or mispredictions can lead to erroneous decisions resulting in unreliable operations. Designing robust ML with the ability to provide reliable results in the presence of such attacks has become a top priority in the field of adversarial machine learning. An essential characteristic for rapid development of robust ML is an arms race between attack and defense strategists. However, an important prerequisite for the arms race is access to a well-defined system model so that experiments can be repeated by independent researchers. This paper proposes a fine-grained system-driven taxonomy to specify ML applications and adversarial system models in an unambiguous manner such that independent researchers can replicate experiments and escalate the arms race to develop more evolved and robust ML applications. The paper provides taxonomies for: 1) the dataset, 2) the ML architecture, 3) the adversary's knowledge, capability, and goal, 4) adversary's strategy, and 5) the defense response. In addition, the relationships among these models and taxonomies are analyzed by proposing an adversarial machine learning cycle. The provided models and taxonomies are merged to form a comprehensive system-driven taxonomy, which represents the arms race between the ML applications and adversaries in recent years. The taxonomies encode best practices in the field and help evaluate and compare the contributions of research works and reveals gaps in the field.
机器学习(ML)算法,特别是监督学习,在现代实际应用中被广泛使用,这些应用以计算智能(CI)作为其核心技术,如自动驾驶车辆、辅助机器人和生物识别系统。导致错误分类或错误预测的攻击可能会导致错误决策,从而产生不可靠的操作。设计在存在此类攻击时能够提供可靠结果的鲁棒ML已成为对抗机器学习领域的首要任务。鲁棒ML快速发展的一个基本特征是攻击和防御策略制定者之间的军备竞赛。然而,军备竞赛的一个重要前提是能够获得一个定义明确的系统模型,以便独立研究人员能够重复实验。本文提出了一种细粒度的系统驱动分类法,以明确的方式指定ML应用和对抗系统模型,使独立研究人员能够复制实验并推动军备竞赛,以开发更先进、更鲁棒的ML应用。本文提供了以下分类法:1)数据集,2)ML架构,3)对手的知识、能力和目标,4)对手的策略,5)防御响应。此外,通过提出一个对抗机器学习周期来分析这些模型和分类法之间的关系。所提供的模型和分类法被合并形成一个全面的系统驱动分类法,它代表了近年来ML应用和对手之间的军备竞赛。这些分类法编码了该领域的最佳实践,有助于评估和比较研究工作的贡献,并揭示该领域的差距。