Suppr超能文献

移动健康应用中基于离策略估计的长期平均结果

Off-Policy Estimation of Long-Term Average Outcomes with Applications to Mobile Health.

作者信息

Liao Peng, Klasnja Predrag, Murphy Susan

机构信息

Department of Statistics, University of Michigan.

School of Information, University of Michigan.

出版信息

J Am Stat Assoc. 2021;116(533):382-391. doi: 10.1080/01621459.2020.1807993. Epub 2020 Oct 1.

Abstract

Due to the recent advancements in wearables and sensing technology, health scientists are increasingly developing mobile health (mHealth) interventions. In mHealth interventions, mobile devices are used to deliver treatment to individuals as they go about their daily lives. These treatments are generally designed to impact a near time, proximal outcome such as stress or physical activity. The mHealth intervention policies, often called just-in-time adaptive interventions, are decision rules that map a individual's current state (e.g., individual's past behaviors as well as current observations of time, location, social activity, stress and urges to smoke) to a particular treatment at each of many time points. The vast majority of current mHealth interventions deploy expert-derived policies. In this paper, we provide an approach for conducting inference about the performance of one or more such policies using historical data collected under a possibly different policy. Our measure of performance is the average of proximal outcomes over a long time period should the particular mHealth policy be followed. We provide an estimator as well as confidence intervals. This work is motivated by HeartSteps, an mHealth physical activity intervention.

摘要

由于可穿戴设备和传感技术最近的进展,健康科学家们越来越多地开发移动健康(mHealth)干预措施。在移动健康干预中,移动设备用于在个体日常生活过程中为其提供治疗。这些治疗通常旨在影响近期、近端的结果,如压力或身体活动。移动健康干预政策,通常称为即时自适应干预,是一种决策规则,它将个体的当前状态(例如,个体过去的行为以及当前对时间、地点、社交活动、压力和吸烟冲动的观察)映射到多个时间点中每个时间点的特定治疗。当前绝大多数移动健康干预措施采用专家制定的政策。在本文中,我们提供了一种方法,用于使用在可能不同的政策下收集的历史数据对一个或多个此类政策的性能进行推断。我们的性能衡量标准是如果遵循特定的移动健康政策,在很长一段时间内近端结果的平均值。我们提供了一个估计器以及置信区间。这项工作的灵感来自于移动健康身体活动干预措施“心脏步数”(HeartSteps)。

相似文献

1
Off-Policy Estimation of Long-Term Average Outcomes with Applications to Mobile Health.
J Am Stat Assoc. 2021;116(533):382-391. doi: 10.1080/01621459.2020.1807993. Epub 2020 Oct 1.
2
Personalized HeartSteps: A Reinforcement Learning Algorithm for Optimizing Physical Activity.
Proc ACM Interact Mob Wearable Ubiquitous Technol. 2020 Mar;4(1). doi: 10.1145/3381007.
5
Data-driven Interpretable Policy Construction for Personalized Mobile Health.
2022 IEEE Int Conf Digit Health IEEE IDCH 2022 (2022). 2022 Jul;2022:13-22. doi: 10.1109/ICDH55609.2022.00010. Epub 2022 Aug 24.
7
Mobile Health Technologies in Cardiopulmonary Disease.
Chest. 2020 Mar;157(3):654-664. doi: 10.1016/j.chest.2019.10.015. Epub 2019 Oct 31.
10

引用本文的文献

1
Optimizing Contingency Management with Reinforcement Learning.
medRxiv. 2024 Mar 29:2024.03.28.24305031. doi: 10.1101/2024.03.28.24305031.
2
BATCH POLICY LEARNING IN AVERAGE REWARD MARKOV DECISION PROCESSES.
Ann Stat. 2022 Dec;50(6):3364-3387. doi: 10.1214/22-aos2231. Epub 2022 Dec 21.
3
Analyzing Patient Trajectories With Artificial Intelligence.
J Med Internet Res. 2021 Dec 3;23(12):e29812. doi: 10.2196/29812.

本文引用的文献

2
Estimating Dynamic Treatment Regimes in Mobile Health Using V-learning.
J Am Stat Assoc. 2020;115(530):692-706. doi: 10.1080/01621459.2018.1537919. Epub 2019 Apr 17.
3
Just-in-Time but Not Too Much: Determining Treatment Timing in Mobile Health.
Proc ACM Interact Mob Wearable Ubiquitous Technol. 2018 Dec;2(4). doi: 10.1145/3287057.
7
A PARTIALLY LINEAR FRAMEWORK FOR MASSIVE HETEROGENEOUS DATA.
Ann Stat. 2016 Aug;44(4):1400-1437. doi: 10.1214/15-AOS1410. Epub 2016 Jul 7.
9
Sample size calculations for micro-randomized trials in mHealth.
Stat Med. 2016 May 30;35(12):1944-71. doi: 10.1002/sim.6847. Epub 2015 Dec 28.
10

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验