• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

深度强化学习中对对抗性状态不确定性的可验证鲁棒性

Certifiable Robustness to Adversarial State Uncertainty in Deep Reinforcement Learning.

作者信息

Everett Michael, Lutjens Bjorn, How Jonathan P

出版信息

IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4184-4198. doi: 10.1109/TNNLS.2021.3056046. Epub 2022 Aug 31.

DOI:10.1109/TNNLS.2021.3056046
PMID:33587714
Abstract

Deep neural network-based systems are now state-of-the-art in many robotics tasks, but their application in safety-critical domains remains dangerous without formal guarantees on network robustness. Small perturbations to sensor inputs (from noise or adversarial examples) are often enough to change network-based decisions, which was recently shown to cause an autonomous vehicle to swerve into another lane. In light of these dangers, numerous algorithms have been developed as defensive mechanisms from these adversarial inputs, some of which provide formal robustness guarantees or certificates. This work leverages research on certified adversarial robustness to develop an online certifiably robust for deep reinforcement learning algorithms. The proposed defense computes guaranteed lower bounds on state-action values during execution to identify and choose a robust action under a worst case deviation in input space due to possible adversaries or noise. Moreover, the resulting policy comes with a certificate of solution quality, even though the true state and optimal action are unknown to the certifier due to the perturbations. The approach is demonstrated on a deep Q-network (DQN) policy and is shown to increase robustness to noise and adversaries in pedestrian collision avoidance scenarios, a classic control task, and Atari Pong. This article extends our prior work with new performance guarantees, extensions to other reinforcement learning algorithms, expanded results aggregated across more scenarios, an extension into scenarios with adversarial behavior, comparisons with a more computationally expensive method, and visualizations that provide intuition about the robustness algorithm.

摘要

基于深度神经网络的系统目前在许多机器人任务中处于领先水平,但在没有网络鲁棒性的形式化保证的情况下,它们在安全关键领域的应用仍然存在危险。对传感器输入的微小扰动(来自噪声或对抗性示例)通常足以改变基于网络的决策,最近有研究表明,这会导致自动驾驶车辆驶入另一条车道。鉴于这些危险,人们开发了许多算法作为针对这些对抗性输入的防御机制,其中一些提供了形式化的鲁棒性保证或证书。这项工作利用了经认证的对抗鲁棒性研究,为深度强化学习算法开发了一种在线可认证鲁棒性方法。所提出的防御措施在执行过程中计算状态 - 动作值的有保证的下界,以便在由于可能的对手或噪声导致的输入空间最坏情况偏差下识别并选择一个鲁棒动作。此外,即使由于扰动认证器不知道真实状态和最优动作,所得到的策略也带有解决方案质量的证书。该方法在深度Q网络(DQN)策略上进行了演示,并在行人碰撞避免场景、经典控制任务和雅达利乒乓球游戏中显示出对噪声和对手的鲁棒性增强。本文通过新的性能保证、对其他强化学习算法的扩展、跨更多场景汇总的扩展结果、对具有对抗行为场景的扩展、与计算成本更高的方法的比较以及提供关于鲁棒性算法直观理解的可视化,扩展了我们之前的工作。

相似文献

1
Certifiable Robustness to Adversarial State Uncertainty in Deep Reinforcement Learning.深度强化学习中对对抗性状态不确定性的可验证鲁棒性
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4184-4198. doi: 10.1109/TNNLS.2021.3056046. Epub 2022 Aug 31.
2
Adversary Agnostic Robust Deep Reinforcement Learning.
IEEE Trans Neural Netw Learn Syst. 2023 Sep;34(9):6146-6157. doi: 10.1109/TNNLS.2021.3133537. Epub 2023 Sep 1.
3
Adversarial Training With Anti-Adversaries.
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):10210-10227. doi: 10.1109/TPAMI.2024.3432973. Epub 2024 Nov 6.
4
Safe Reinforcement Learning With Dual Robustness.具有双重稳健性的安全强化学习
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):10876-10890. doi: 10.1109/TPAMI.2024.3443916. Epub 2024 Nov 6.
5
Adversarial Robustness of Deep Reinforcement Learning Based Dynamic Recommender Systems.基于深度强化学习的动态推荐系统的对抗鲁棒性
Front Big Data. 2022 May 3;5:822783. doi: 10.3389/fdata.2022.822783. eCollection 2022.
6
Defense against adversarial attacks based on color space transformation.基于颜色空间变换的对抗攻击防御。
Neural Netw. 2024 May;173:106176. doi: 10.1016/j.neunet.2024.106176. Epub 2024 Feb 14.
7
Adversarial Attack and Defence through Adversarial Training and Feature Fusion for Diabetic Retinopathy Recognition.对抗训练和特征融合在糖尿病视网膜病变识别中的对抗攻击和防御。
Sensors (Basel). 2021 Jun 7;21(11):3922. doi: 10.3390/s21113922.
8
Adversarial Robustness with Partial Isometry.具有部分等距的对抗鲁棒性。
Entropy (Basel). 2024 Jan 24;26(2):103. doi: 10.3390/e26020103.
9
Interpreting and Improving Adversarial Robustness of Deep Neural Networks With Neuron Sensitivity.基于神经元敏感性的深度神经网络对抗鲁棒性解释与改进。
IEEE Trans Image Process. 2021;30:1291-1304. doi: 10.1109/TIP.2020.3042083. Epub 2020 Dec 23.
10
Improving the robustness and accuracy of biomedical language models through adversarial training.通过对抗训练提高生物医学语言模型的稳健性和准确性。
J Biomed Inform. 2022 Aug;132:104114. doi: 10.1016/j.jbi.2022.104114. Epub 2022 Jun 15.