• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过基于大语言模型训练的交叉注意力网络增强生成式智能体中的记忆检索。

Enhancing memory retrieval in generative agents through LLM-trained cross attention networks.

作者信息

Hong Chuanyang, He Qingyun

机构信息

School of Computing and Artificial Intelligence, Southwestern University of Finance and Economics, Chengdu, China.

School of Finance and Economics, Anhui Science and Technology University, Bengbu, China.

出版信息

Front Psychol. 2025 May 7;16:1591618. doi: 10.3389/fpsyg.2025.1591618. eCollection 2025.

DOI:10.3389/fpsyg.2025.1591618
PMID:40400752
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12092450/
Abstract

INTRODUCTION

The surge in the capabilities of large language models (LLMs) has propelled the development of Artificial General Intelligence (AGI), highlighting generative agents as pivotal components for emulating complex AI behaviors. Given the high costs associated with individually training LLMs for each AI agent, there is a critical need for advanced memory retrieval mechanisms to maintain the unique characteristics and memories of individual AI agents.

METHODS

In this research, we developed a text-based simulation of a generative agent world, constructing a community with multiple agents and locations in which certain levels of interaction were enabled. Within this framework, we introduced a novel memory retrieval system using an Auxiliary Cross Attention Network (ACAN). This system calculates and ranks attention weights between an agent's current state and stored memories, selecting the most relevant memories for any given situation. In a novel approach, we incorporated LLM assistance, comparing memories retrieved by our model with those extracted using a base method during training, and constructing a novel loss function based on these comparisons to optimize the training process effectively. To our knowledge, this is the first study to utilize LLMs to train a dedicated agent memory retrieval network.

RESULTS

Our empirical evaluations demonstrate that this approach substantially enhances the quality of memory retrieval, thereby increasing the adaptability and behavioral consistency of agents in fluctuating environments.

DISCUSSION

Our findings not only introduce new perspectives and methodologies for memory retrieval in generative agents but also extend the utility of LLMs in memory management across varied AI agent applications.

摘要

引言

大语言模型(LLMs)能力的激增推动了通用人工智能(AGI)的发展,凸显了生成式智能体作为模拟复杂人工智能行为的关键组件。鉴于为每个人工智能智能体单独训练大语言模型成本高昂,迫切需要先进的内存检索机制来维护单个人工智能智能体的独特特征和记忆。

方法

在本研究中,我们开发了一个基于文本的生成式智能体世界模拟,构建了一个具有多个智能体和地点的社区,其中允许一定程度的交互。在此框架内,我们引入了一种使用辅助交叉注意力网络(ACAN)的新型内存检索系统。该系统计算并对智能体当前状态与存储记忆之间的注意力权重进行排序,为任何给定情况选择最相关的记忆。采用一种新颖的方法,我们纳入了大语言模型的辅助,将我们模型检索到的记忆与训练期间使用基本方法提取的记忆进行比较,并基于这些比较构建一个新颖的损失函数,以有效优化训练过程。据我们所知,这是第一项利用大语言模型训练专用智能体内存检索网络的研究。

结果

我们的实证评估表明,这种方法显著提高了内存检索的质量,从而提高了智能体在波动环境中的适应性和行为一致性。

讨论

我们的研究结果不仅为生成式智能体的内存检索引入了新的视角和方法,还扩展了大语言模型在各种人工智能智能体应用中的内存管理效用。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2492/12092450/e2e4f69e2a1b/fpsyg-16-1591618-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2492/12092450/9973d351baa8/fpsyg-16-1591618-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2492/12092450/ee1316ca2ab5/fpsyg-16-1591618-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2492/12092450/e79652d6f000/fpsyg-16-1591618-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2492/12092450/e2e4f69e2a1b/fpsyg-16-1591618-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2492/12092450/9973d351baa8/fpsyg-16-1591618-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2492/12092450/ee1316ca2ab5/fpsyg-16-1591618-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2492/12092450/e79652d6f000/fpsyg-16-1591618-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2492/12092450/e2e4f69e2a1b/fpsyg-16-1591618-g0004.jpg

相似文献

1
Enhancing memory retrieval in generative agents through LLM-trained cross attention networks.通过基于大语言模型训练的交叉注意力网络增强生成式智能体中的记忆检索。
Front Psychol. 2025 May 7;16:1591618. doi: 10.3389/fpsyg.2025.1591618. eCollection 2025.
2
Generative Large Language Model-Powered Conversational AI App for Personalized Risk Assessment: Case Study in COVID-19.用于个性化风险评估的生成式大语言模型驱动的对话式人工智能应用程序:COVID-19案例研究
JMIR AI. 2025 Mar 27;4:e67363. doi: 10.2196/67363.
3
The Accuracy and Capability of Artificial Intelligence Solutions in Health Care Examinations and Certificates: Systematic Review and Meta-Analysis.人工智能解决方案在医疗检查和证书中的准确性和能力:系统评价和荟萃分析。
J Med Internet Res. 2024 Nov 5;26:e56532. doi: 10.2196/56532.
4
Using Generative Artificial Intelligence in Health Economics and Outcomes Research: A Primer on Techniques and Breakthroughs.在卫生经济学与结果研究中使用生成式人工智能:技术与突破入门
Pharmacoecon Open. 2025 Apr 29. doi: 10.1007/s41669-025-00580-4.
5
Evaluating the Capabilities of Generative AI Tools in Understanding Medical Papers: Qualitative Study.评估生成式人工智能工具理解医学论文的能力:定性研究
JMIR Med Inform. 2024 Sep 4;12:e59258. doi: 10.2196/59258.
6
Utilizing large language models for gastroenterology research: a conceptual framework.利用大语言模型进行胃肠病学研究:一个概念框架。
Therap Adv Gastroenterol. 2025 Apr 1;18:17562848251328577. doi: 10.1177/17562848251328577. eCollection 2025.
7
Generative Retrieval-Augmented Ontologic Graph and Multiagent Strategies for Interpretive Large Language Model-Based Materials Design.基于生成式检索增强本体图和多智能体策略的基于解释性大语言模型的材料设计
ACS Eng Au. 2024 Jan 12;4(2):241-277. doi: 10.1021/acsengineeringau.3c00058. eCollection 2024 Apr 17.
8
Improving Dietary Supplement Information Retrieval: Development of a Retrieval-Augmented Generation System With Large Language Models.改善膳食补充剂信息检索:利用大语言模型开发检索增强生成系统
J Med Internet Res. 2025 Mar 19;27:e67677. doi: 10.2196/67677.
9
Large Language Models in Worldwide Medical Exams: Platform Development and Comprehensive Analysis.全球医学考试中的大语言模型:平台开发与综合分析
J Med Internet Res. 2024 Dec 27;26:e66114. doi: 10.2196/66114.
10
The Role of Large Language Models in Transforming Emergency Medicine: Scoping Review.大型语言模型在变革急诊医学中的作用:范围综述
JMIR Med Inform. 2024 May 10;12:e53787. doi: 10.2196/53787.

本文引用的文献

1
Reconsolidation of human memory: brain mechanisms and clinical relevance.人类记忆的再巩固:大脑机制与临床相关性。
Biol Psychiatry. 2014 Aug 15;76(4):274-80. doi: 10.1016/j.biopsych.2014.03.008. Epub 2014 Mar 15.
2
Mechanisms of memory.记忆机制。
Science. 1986 Jun 27;232(4758):1612-9. doi: 10.1126/science.3086978.