Subasri Vallijah, Baghbanzadeh Negin, Celi Leo Anthony, Seyyed-Kalantari Laleh
Peter Munk Cardiac Centre, University Health Network, Toronto, Ontario, Canada.
Vector Institute, Toronto, Ontario, Canada.
BMJ Health Care Inform. 2025 Apr 27;32(1):e101130. doi: 10.1136/bmjhci-2024-101130.
The recent emergence of foundation model-based chatbots, such as ChatGPT (OpenAI, San Francisco, CA, USA), has showcased remarkable language mastery and intuitive comprehension capabilities. Despite significant efforts to identify and address the near-term risks associated with artificial intelligence (AI), our understanding of the existential threats they pose remains limited. Near-term risks stem from AI that already exist or are under active development with a clear trajectory towards deployment. Existential risks of AI can be an extension of the near-term risks studied by the fairness, accountability, transparency and ethics community, and are characterised by a potential to threaten humanity's long-term potential. In this paper, we delve into the ways AI can give rise to existential harm and explore potential risk mitigation strategies. This involves further investigation of critical domains, including AI alignment, overtrust in AI, AI safety, open-sourcing, the implications of AI to healthcare and the broader societal risks.
最近出现的基于基础模型的聊天机器人,如ChatGPT(美国加利福尼亚州旧金山的OpenAI公司),展现出了卓越的语言掌握能力和直观理解能力。尽管人们付出了巨大努力来识别和应对与人工智能(AI)相关的短期风险,但我们对它们所构成的生存威胁的理解仍然有限。短期风险源于已经存在或正在积极开发且有明确部署轨迹的人工智能。人工智能的生存风险可能是公平、问责、透明和伦理领域所研究的短期风险的延伸,其特点是有可能威胁人类的长期潜力。在本文中,我们深入探讨人工智能可能造成生存危害的方式,并探索潜在的风险缓解策略。这涉及对关键领域的进一步研究,包括人工智能对齐、对人工智能的过度信任、人工智能安全、开源、人工智能对医疗保健的影响以及更广泛的社会风险。