• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

带有时间先验的自注意力机制:我们能否从时间之箭中学到更多?

Self-attention with temporal prior: can we learn more from the arrow of time?

作者信息

Kim Kyung Geun, Lee Byeong Tak

机构信息

VUNO Inc., Seoul, Republic of Korea.

Medical AI Co., Ltd., Seoul, Republic of Korea.

出版信息

Front Artif Intell. 2024 Aug 6;7:1397298. doi: 10.3389/frai.2024.1397298. eCollection 2024.

DOI:10.3389/frai.2024.1397298
PMID:39165902
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11333831/
Abstract

Many diverse phenomena in nature often inherently encode both short- and long-term temporal dependencies, which especially result from the direction of the flow of time. In this respect, we discovered experimental evidence suggesting that of these events are higher for closer time stamps. However, to be able for attention-based models to learn these regularities in short-term dependencies, it requires large amounts of data, which are often infeasible. This is because, while they are good at learning piece-wise temporal dependencies, attention-based models lack structures that encode biases in time series. As a resolution, we propose a simple and efficient method that enables attention layers to better encode the short-term temporal bias of these data sets by applying learnable, adaptive kernels directly to the attention matrices. We chose various prediction tasks for the experiments using Electronic Health Records (EHR) data sets since they are great examples with underlying long- and short-term temporal dependencies. Our experiments show exceptional classification results compared to best-performing models on most tasks and data sets.

摘要

自然界中许多不同的现象往往内在地编码了短期和长期的时间依赖性,这尤其源于时间流动的方向。在这方面,我们发现实验证据表明,对于时间戳越接近的这些事件,其发生率越高。然而,为了使基于注意力的模型能够学习短期依赖性中的这些规律,需要大量数据,而这通常是不可行的。这是因为,虽然基于注意力的模型擅长学习分段时间依赖性,但它们缺乏对时间序列中的偏差进行编码的结构。作为一种解决方案,我们提出了一种简单有效的方法,通过将可学习的自适应内核直接应用于注意力矩阵,使注意力层能够更好地编码这些数据集的短期时间偏差。我们选择使用电子健康记录(EHR)数据集进行各种预测任务的实验,因为它们是具有潜在长期和短期时间依赖性的很好例子。与大多数任务和数据集上表现最佳的模型相比,我们的实验显示出了卓越的分类结果。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bdb2/11333831/ed5b448f34ab/frai-07-1397298-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bdb2/11333831/20eed1475028/frai-07-1397298-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bdb2/11333831/bf2cbceb4803/frai-07-1397298-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bdb2/11333831/7454053ace68/frai-07-1397298-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bdb2/11333831/ed5b448f34ab/frai-07-1397298-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bdb2/11333831/20eed1475028/frai-07-1397298-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bdb2/11333831/bf2cbceb4803/frai-07-1397298-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bdb2/11333831/7454053ace68/frai-07-1397298-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bdb2/11333831/ed5b448f34ab/frai-07-1397298-g0004.jpg

相似文献

1
Self-attention with temporal prior: can we learn more from the arrow of time?带有时间先验的自注意力机制:我们能否从时间之箭中学到更多?
Front Artif Intell. 2024 Aug 6;7:1397298. doi: 10.3389/frai.2024.1397298. eCollection 2024.
2
ConTraNet: A hybrid network for improving the classification of EEG and EMG signals with limited training data.ConTraNet:一种混合网络,用于在有限的训练数据下提高 EEG 和 EMG 信号的分类。
Comput Biol Med. 2024 Jan;168:107649. doi: 10.1016/j.compbiomed.2023.107649. Epub 2023 Nov 2.
3
Deep learning prediction models based on EHR trajectories: A systematic review.基于电子健康记录轨迹的深度学习预测模型:系统评价。
J Biomed Inform. 2023 Aug;144:104430. doi: 10.1016/j.jbi.2023.104430. Epub 2023 Jun 26.
4
Temporal-based Swin Transformer network for workflow recognition of surgical video.用于手术视频工作流识别的基于时间的Swin Transformer网络
Int J Comput Assist Radiol Surg. 2023 Jan;18(1):139-147. doi: 10.1007/s11548-022-02785-y. Epub 2022 Nov 4.
5
Learning global dependencies based on hierarchical full connection for brain tumor segmentation.基于层次全连接的脑肿瘤分割的全局依赖关系学习。
Comput Methods Programs Biomed. 2022 Jun;221:106925. doi: 10.1016/j.cmpb.2022.106925. Epub 2022 May 30.
6
A Temporal Window Attention-Based Window-Dependent Long Short-Term Memory Network for Multivariate Time Series Prediction.一种基于时间窗口注意力的窗口相关长短期记忆网络用于多变量时间序列预测。
Entropy (Basel). 2022 Dec 21;25(1):10. doi: 10.3390/e25010010.
7
MR-Transformer: Multiresolution Transformer for Multivariate Time Series Prediction.MR-Transformer:用于多元时间序列预测的多分辨率Transformer
IEEE Trans Neural Netw Learn Syst. 2025 Jan;36(1):1171-1183. doi: 10.1109/TNNLS.2023.3327416. Epub 2025 Jan 7.
8
TransformerG2G: Adaptive time-stepping for learning temporal graph embeddings using transformers.TransformerG2G:使用转换器学习时间图嵌入的自适应时间步长。
Neural Netw. 2024 Apr;172:106086. doi: 10.1016/j.neunet.2023.12.040. Epub 2023 Dec 26.
9
Continuous patient state attention model for addressing irregularity in electronic health records.用于解决电子健康记录中不规则性的连续患者状态注意模型。
BMC Med Inform Decis Mak. 2024 May 3;24(1):117. doi: 10.1186/s12911-024-02514-2.
10
Disentangled Dynamic Deviation Transformer Networks for Multivariate Time Series Anomaly Detection.解缠动态偏差变换网络在多元时间序列异常检测中的应用。
Sensors (Basel). 2023 Jan 18;23(3):1104. doi: 10.3390/s23031104.

本文引用的文献

1
TransformEHR: transformer-based encoder-decoder generative model to enhance prediction of disease outcomes using electronic health records.TransformEHR:基于转换器的编解码器生成模型,用于使用电子健康记录增强疾病结局预测。
Nat Commun. 2023 Nov 29;14(1):7857. doi: 10.1038/s41467-023-43715-z.
2
Hi-BEHRT: Hierarchical Transformer-Based Model for Accurate Prediction of Clinical Events Using Multimodal Longitudinal Electronic Health Records.Hi-BEHRT:基于分层转换器的模型,用于使用多模态纵向电子健康记录准确预测临床事件。
IEEE J Biomed Health Inform. 2023 Feb;27(2):1106-1117. doi: 10.1109/JBHI.2022.3224727. Epub 2023 Feb 3.
3
Bidirectional Representation Learning From Transformers Using Multimodal Electronic Health Record Data to Predict Depression.
利用多模态电子健康记录数据从转换器中进行双向表示学习以预测抑郁。
IEEE J Biomed Health Inform. 2021 Aug;25(8):3121-3129. doi: 10.1109/JBHI.2021.3063721. Epub 2021 Aug 5.
4
Early Prediction of Sepsis From Clinical Data: The PhysioNet/Computing in Cardiology Challenge 2019.从临床数据中早期预测脓毒症:PhysioNet/Computing in Cardiology 挑战赛 2019。
Crit Care Med. 2020 Feb;48(2):210-217. doi: 10.1097/CCM.0000000000004145.
5
Multitask learning and benchmarking with clinical time series data.多任务学习与临床时间序列数据的基准测试。
Sci Data. 2019 Jun 17;6(1):96. doi: 10.1038/s41597-019-0103-9.
6
The eICU Collaborative Research Database, a freely available multi-center database for critical care research.eICU 协作研究数据库,一个免费的多中心重症监护研究数据库。
Sci Data. 2018 Sep 11;5:180178. doi: 10.1038/sdata.2018.178.
7
Recurrent Neural Networks for Multivariate Time Series with Missing Values.具有缺失值的多元时间序列的递归神经网络。
Sci Rep. 2018 Apr 17;8(1):6085. doi: 10.1038/s41598-018-24271-9.
8
MIMIC-III, a freely accessible critical care database.MIMIC-III,一个免费获取的重症监护数据库。
Sci Data. 2016 May 24;3:160035. doi: 10.1038/sdata.2016.35.
9
The Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3).《脓毒症及脓毒性休克第三次国际共识定义(脓毒症-3)》
JAMA. 2016 Feb 23;315(8):801-10. doi: 10.1001/jama.2016.0287.
10
Why try to predict ICU outcomes?为什么要尝试预测重症监护病房的治疗结果?
Curr Opin Crit Care. 2014 Oct;20(5):544-9. doi: 10.1097/MCC.0000000000000136.