• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

在集体风险困境中选择错误的人工代表比直接犯错要好。

Committing to the wrong artificial delegate in a collective-risk dilemma is better than directly committing mistakes.

机构信息

Department of Information Technology-IDLab, Ghent University-IMEC, Technologiepark Zwijnaarde 126, 9052, Ghent, Belgium.

Artificial Intelligence Lab, Computer Science Department, Vrije Universiteit Brussel, 1050, Brussels, Belgium.

出版信息

Sci Rep. 2024 May 7;14(1):10460. doi: 10.1038/s41598-024-61153-9.

DOI:10.1038/s41598-024-61153-9
PMID:38714713
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11076577/
Abstract

While autonomous artificial agents are assumed to perfectly execute the strategies they are programmed with, humans who design them may make mistakes. These mistakes may lead to a misalignment between the humans' intended goals and their agents' observed behavior, a problem of value alignment. Such an alignment problem may have particularly strong consequences when these autonomous systems are used in social contexts that involve some form of collective risk. By means of an evolutionary game theoretical model, we investigate whether errors in the configuration of artificial agents change the outcome of a collective-risk dilemma, in comparison to a scenario with no delegation. Delegation is here distinguished from no-delegation simply by the moment at which a mistake occurs: either when programming/choosing the agent (in case of delegation) or when executing the actions at each round of the game (in case of no-delegation). We find that, while errors decrease success rate, it is better to delegate and commit to a somewhat flawed strategy, perfectly executed by an autonomous agent, than to commit execution errors directly. Our model also shows that in the long-term, delegation strategies should be favored over no-delegation, if given the choice.

摘要

虽然自主人工智能代理被假设能够完美执行它们所编程的策略,但设计它们的人类可能会犯错。这些错误可能导致人类的预期目标与其代理的观察行为之间的不一致,这是一个价值对齐的问题。当这些自主系统被用于涉及某种形式的集体风险的社交环境中时,这种对齐问题可能会产生特别强烈的后果。通过一个进化博弈论模型,我们研究了在没有委托的情况下,与配置人工代理的错误相比,这种错误是否会改变集体风险困境的结果。这里的委托与非委托的区别仅仅在于错误发生的时刻:要么是在编程/选择代理时(如果是委托),要么是在游戏的每一轮执行行动时(如果是无委托)。我们发现,虽然错误会降低成功率,但最好是委托给一个自主代理来执行一个略有缺陷的策略,而不是直接犯执行错误。我们的模型还表明,如果有选择的话,从长远来看,应该优先选择委托策略而不是无委托策略。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/08fb407925b2/41598_2024_61153_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/32cb034e1c1a/41598_2024_61153_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/c242a4466cca/41598_2024_61153_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/8ccc6214147e/41598_2024_61153_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/277a8c67f110/41598_2024_61153_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/08fb407925b2/41598_2024_61153_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/32cb034e1c1a/41598_2024_61153_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/c242a4466cca/41598_2024_61153_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/8ccc6214147e/41598_2024_61153_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/277a8c67f110/41598_2024_61153_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2265/11076577/08fb407925b2/41598_2024_61153_Fig5_HTML.jpg

相似文献

1
Committing to the wrong artificial delegate in a collective-risk dilemma is better than directly committing mistakes.在集体风险困境中选择错误的人工代表比直接犯错要好。
Sci Rep. 2024 May 7;14(1):10460. doi: 10.1038/s41598-024-61153-9.
2
Delegation to artificial agents fosters prosocial behaviors in the collective risk dilemma.授权给人工智能代理有助于促进集体风险困境中的亲社会行为。
Sci Rep. 2022 May 19;12(1):8492. doi: 10.1038/s41598-022-11518-9.
3
Medication errors by caregivers in the homes of children discharged from a pediatric department in Ghana.加纳一家儿科科室出院儿童家中护理人员的用药错误。
Ther Adv Drug Saf. 2024 Jan 27;15:20420986231225850. doi: 10.1177/20420986231225850. eCollection 2024.
4
Delegation: developing the habit.授权:养成习惯。
Radiol Manage. 2001 Jul-Aug;23(4):16-20, 22, 24.
5
Coevolutionary dynamics via adaptive feedback in collective-risk social dilemma game.通过集体风险社会困境博弈中的自适应反馈进行共同进化动力学。
Elife. 2023 May 19;12:e82954. doi: 10.7554/eLife.82954.
6
Difference of reciprocity effect in two coevolutionary models of presumed two-player and multiplayer games.假定的两人及多人博弈的两种协同进化模型中互惠效应的差异。
Phys Rev E Stat Nonlin Soft Matter Phys. 2013 Jun;87(6):062136. doi: 10.1103/PhysRevE.87.062136. Epub 2013 Jun 25.
7
The art of compensation: How hybrid teams solve collective-risk dilemmas.补偿的艺术:混合团队如何解决集体风险困境。
PLoS One. 2024 Feb 9;19(2):e0297213. doi: 10.1371/journal.pone.0297213. eCollection 2024.
8
Strategies that enforce linear payoff relationships under observation errors in Repeated Prisoner's Dilemma game.在重复囚徒困境博弈中观察误差下强制线性收益关系的策略。
J Theor Biol. 2019 Sep 21;477:63-76. doi: 10.1016/j.jtbi.2019.06.009. Epub 2019 Jun 12.
9
Stochastic evolutionary dynamics resolve the Traveler's Dilemma.随机进化动力学解决了旅行者困境。
J Theor Biol. 2012 Jun 21;303:119-27. doi: 10.1016/j.jtbi.2012.03.014. Epub 2012 Mar 23.
10
Supporting the distributed execution of clinical guidelines by multiple agents.支持多个代理分布式执行临床指南。
Artif Intell Med. 2019 Jul;98:87-108. doi: 10.1016/j.artmed.2019.05.001. Epub 2019 May 18.

引用本文的文献

1
Emergence of cooperation in the one-shot Prisoner's dilemma through Discriminatory and Samaritan AIs.通过歧视性和利他性 AI,实现一次性囚徒困境中的合作。
J R Soc Interface. 2024 Sep;21(218):20240212. doi: 10.1098/rsif.2024.0212. Epub 2024 Sep 25.

本文引用的文献

1
EGTtools: Evolutionary game dynamics in Python.EGTtools:用Python实现的进化博弈动力学
iScience. 2023 Mar 17;26(4):106419. doi: 10.1016/j.isci.2023.106419. eCollection 2023 Apr 21.
2
Human-level play in the game of by combining language models with strategic reasoning.通过将语言模型与策略推理相结合,在游戏中实现人类级别的表现。
Science. 2022 Dec 9;378(6624):1067-1074. doi: 10.1126/science.ade9097. Epub 2022 Nov 22.
3
Delegation to artificial agents fosters prosocial behaviors in the collective risk dilemma.授权给人工智能代理有助于促进集体风险困境中的亲社会行为。
Sci Rep. 2022 May 19;12(1):8492. doi: 10.1038/s41598-022-11518-9.
4
Cooperative AI: machines must learn to find common ground.协作式人工智能:机器必须学会找到共同点。
Nature. 2021 May;593(7857):33-36. doi: 10.1038/d41586-021-01170-0.
5
Timing Uncertainty in Collective Risk Dilemmas Encourages Group Reciprocation and Polarization.集体风险困境中的时间不确定性会促使群体互惠与两极分化。
iScience. 2020 Oct 31;23(12):101752. doi: 10.1016/j.isci.2020.101752. eCollection 2020 Dec 18.
6
Signalling boosts the evolution of cooperation in repeated group interactions.信号促进了重复群体相互作用中合作的进化。
J R Soc Interface. 2020 Nov;17(172):20200635. doi: 10.1098/rsif.2020.0635. Epub 2020 Nov 4.
7
Participation in and provision of public goods: Does granularity matter?参与和提供公共物品:粒度重要吗?
J Econ Interact Coord. 2021;16(2):265-285. doi: 10.1007/s11403-020-00293-3. Epub 2020 Jul 17.
8
AI for social good: unlocking the opportunity for positive impact.人工智能造福社会:释放积极影响的机遇。
Nat Commun. 2020 May 18;11(1):2468. doi: 10.1038/s41467-020-15871-z.
9
Computation and Simulation of Evolutionary Game Dynamics in Finite Populations.有限群体中进化博弈动力学的计算与模拟。
Sci Rep. 2019 May 6;9(1):6946. doi: 10.1038/s41598-019-43102-z.
10
Locally noisy autonomous agents improve global human coordination in network experiments.在网络实验中,局部有噪声的自主主体可改善全球人类协调。
Nature. 2017 May 17;545(7654):370-374. doi: 10.1038/nature22332.