• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

Real-Time Online Goal Recognition in Continuous Domains via Deep Reinforcement Learning.

作者信息

Fang Zihao, Chen Dejun, Zeng Yunxiu, Wang Tao, Xu Kai

机构信息

College of Systems Engineering, National University of Defense Technology, Changsha 410000, China.

出版信息

Entropy (Basel). 2023 Oct 4;25(10):1415. doi: 10.3390/e25101415.

DOI:10.3390/e25101415
PMID:37895536
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10606411/
Abstract

The problem of goal recognition involves inferring the high-level task goals of an agent based on observations of its behavior in an environment. Current methods for achieving this task rely on offline comparison inference of observed behavior in discrete environments, which presents several challenges. First, accurately modeling the behavior of the observed agent requires significant computational resources. Second, continuous simulation environments cannot be accurately recognized using existing methods. Finally, real-time computing power is required to infer the likelihood of each potential goal. In this paper, we propose an advanced and efficient real-time online goal recognition algorithm based on deep reinforcement learning in continuous domains. By leveraging the offline modeling of the observed agent's behavior with deep reinforcement learning, our algorithm achieves real-time goal recognition. We evaluate the algorithm's online goal recognition accuracy and stability in continuous simulation environments under communication constraints.

摘要
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/38f9434bd11b/entropy-25-01415-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/c99992966975/entropy-25-01415-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/7bf8dfbd10ff/entropy-25-01415-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/c28d58f23d2c/entropy-25-01415-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/b59b1380dcb3/entropy-25-01415-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/6740939a78ed/entropy-25-01415-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/d13d02be3472/entropy-25-01415-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/38f9434bd11b/entropy-25-01415-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/c99992966975/entropy-25-01415-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/7bf8dfbd10ff/entropy-25-01415-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/c28d58f23d2c/entropy-25-01415-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/b59b1380dcb3/entropy-25-01415-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/6740939a78ed/entropy-25-01415-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/d13d02be3472/entropy-25-01415-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/24e4/10606411/38f9434bd11b/entropy-25-01415-g007.jpg

相似文献

1
Real-Time Online Goal Recognition in Continuous Domains via Deep Reinforcement Learning.
Entropy (Basel). 2023 Oct 4;25(10):1415. doi: 10.3390/e25101415.
2
Inverse Value Iteration and Q-Learning: Algorithms, Stability, and Robustness.逆值迭代与Q学习:算法、稳定性及鲁棒性
IEEE Trans Neural Netw Learn Syst. 2025 Apr;36(4):6970-6980. doi: 10.1109/TNNLS.2024.3409182. Epub 2025 Apr 4.
3
Quantifying Reinforcement-Learning Agent's Autonomy, Reliance on Memory and Internalisation of the Environment.量化强化学习智能体的自主性、对记忆的依赖以及环境内化程度。
Entropy (Basel). 2022 Mar 13;24(3):401. doi: 10.3390/e24030401.
4
Towards a Broad-Persistent Advising Approach for Deep Interactive Reinforcement Learning in Robotic Environments.迈向机器人环境中深度交互式强化学习的广泛持久咨询方法。
Sensors (Basel). 2023 Mar 1;23(5):2681. doi: 10.3390/s23052681.
5
Research on a Task Offloading Strategy for the Internet of Vehicles Based on Reinforcement Learning.基于强化学习的车联网任务卸载策略研究。
Sensors (Basel). 2021 Sep 9;21(18):6058. doi: 10.3390/s21186058.
6
Goal-oriented inference of environment from redundant observations.从冗余观测中进行面向目标的环境推断。
Neural Netw. 2024 Jun;174:106246. doi: 10.1016/j.neunet.2024.106246. Epub 2024 Mar 15.
7
Masked and Inverse Dynamics Modeling for Data-Efficient Reinforcement Learning.用于数据高效强化学习的掩码与逆动力学建模
IEEE Trans Neural Netw Learn Syst. 2025 May;36(5):8814-8827. doi: 10.1109/TNNLS.2024.3439261. Epub 2025 May 2.
8
Bayesian reinforcement learning for navigation planning in unknown environments.用于未知环境中导航规划的贝叶斯强化学习
Front Artif Intell. 2024 Jul 4;7:1308031. doi: 10.3389/frai.2024.1308031. eCollection 2024.
9
PaCAR: COVID-19 Pandemic Control Decision Making via Large-Scale Agent-Based Modeling and Deep Reinforcement Learning.PaCAR:通过大规模基于代理的建模和深度强化学习进行 COVID-19 大流行控制决策。
Med Decis Making. 2022 Nov;42(8):1064-1077. doi: 10.1177/0272989X221107902. Epub 2022 Jul 1.
10
Dynamics-Adaptive Continual Reinforcement Learning via Progressive Contextualization.通过渐进式情境化实现动态自适应持续强化学习
IEEE Trans Neural Netw Learn Syst. 2024 Oct;35(10):14588-14602. doi: 10.1109/TNNLS.2023.3280085. Epub 2024 Oct 7.

本文引用的文献

1
Behaviour Recognition with Kinodynamic Planning Over Continuous Domains.
Front Artif Intell. 2021 Oct 25;4:717003. doi: 10.3389/frai.2021.717003. eCollection 2021.
2
Activity, Plan, and Goal Recognition: A Review.活动、计划与目标识别:综述
Front Robot AI. 2021 May 10;8:643010. doi: 10.3389/frobt.2021.643010. eCollection 2021.
3
Single Real Goal, Magnitude-Based Deceptive Path-Planning.单一实际目标,基于量级的欺骗性路径规划。
Entropy (Basel). 2020 Jan 10;22(1):88. doi: 10.3390/e22010088.
4
Goal Identification Control Using an Information Entropy-Based Goal Uncertainty Metric.使用基于信息熵的目标不确定性度量进行目标识别控制
Entropy (Basel). 2019 Mar 20;21(3):299. doi: 10.3390/e21030299.