Department of Statistics, University of Michigan, Ann Arbor, MI, United States.
Department of Statistics and Data Science, Carnegie Mellon University, Pittsburgh, PA, United States.
J Med Internet Res. 2023 Apr 20;25:e43664. doi: 10.2196/43664.
Although evidence supporting the feasibility of large-scale mobile health (mHealth) systems continues to grow, privacy protection remains an important implementation challenge. The potential scale of publicly available mHealth applications and the sensitive nature of the data involved will inevitably attract unwanted attention from adversarial actors seeking to compromise user privacy. Although privacy-preserving technologies such as federated learning (FL) and differential privacy (DP) offer strong theoretical guarantees, it is not clear how such technologies actually perform under real-world conditions.
Using data from the University of Michigan Intern Health Study (IHS), we assessed the privacy protection capabilities of FL and DP against the trade-offs in the associated model's accuracy and training time. Using a simulated external attack on a target mHealth system, we aimed to measure the effectiveness of such an attack under various levels of privacy protection on the target system and measure the costs to the target system's performance associated with the chosen levels of privacy protection.
A neural network classifier that attempts to predict IHS participant daily mood ecological momentary assessment score from sensor data served as our target system. An external attacker attempted to identify participants whose average mood ecological momentary assessment score is lower than the global average. The attack followed techniques in the literature, given the relevant assumptions about the abilities of the attacker. For measuring attack effectiveness, we collected attack success metrics (area under the curve [AUC], positive predictive value, and sensitivity), and for measuring privacy costs, we calculated the target model training time and measured the model utility metrics. Both sets of metrics are reported under varying degrees of privacy protection on the target.
We found that FL alone does not provide adequate protection against the privacy attack proposed above, where the attacker's AUC in determining which participants exhibit lower than average mood is over 0.90 in the worst-case scenario. However, under the highest level of DP tested in this study, the attacker's AUC fell to approximately 0.59 with only a 10% point decrease in the target's R and a 43% increase in model training time. Attack positive predictive value and sensitivity followed similar trends. Finally, we showed that participants in the IHS most likely to require strong privacy protection are also most at risk from this particular privacy attack and subsequently stand to benefit the most from these privacy-preserving technologies.
Our results demonstrated both the necessity of proactive privacy protection research and the feasibility of the current FL and DP methods implemented in a real mHealth scenario. Our simulation methods characterized the privacy-utility trade-off in our mHealth setup using highly interpretable metrics, providing a framework for future research into privacy-preserving technologies in data-driven health and medical applications.
尽管越来越多的证据支持大规模移动医疗(mHealth)系统的可行性,但隐私保护仍然是一个重要的实施挑战。公开的 mHealth 应用程序的潜在规模和所涉及数据的敏感性将不可避免地吸引敌对行为者的关注,这些行为者试图损害用户隐私。尽管联邦学习(FL)和差分隐私(DP)等隐私保护技术提供了强有力的理论保证,但在实际情况下,这些技术的实际表现尚不清楚。
我们使用密歇根大学实习健康研究(IHS)的数据,评估了 FL 和 DP 在权衡相关模型的准确性和训练时间方面的隐私保护能力。通过对目标 mHealth 系统进行模拟的外部攻击,我们旨在衡量在目标系统的各种隐私保护级别下,这种攻击的有效性,并衡量与所选隐私保护级别相关的目标系统性能的成本。
一个试图从传感器数据预测 IHS 参与者每日情绪生态瞬间评估得分的神经网络分类器作为我们的目标系统。一个外部攻击者试图识别平均情绪生态瞬间评估得分低于全球平均水平的参与者。攻击采用了文献中的技术,考虑到攻击者能力的相关假设。为了衡量攻击的有效性,我们收集了攻击成功指标(曲线下面积[AUC]、阳性预测值和敏感性),为了衡量隐私成本,我们计算了目标模型的训练时间,并衡量了模型的实用指标。在目标系统的不同隐私保护级别下,报告了这两组指标。
我们发现,仅 FL 并不能为上述隐私攻击提供充分的保护,在最糟糕的情况下,攻击者确定哪些参与者表现出较低情绪的 AUC 超过 0.90。然而,在本研究中测试的最高 DP 级别下,攻击者的 AUC 降至约 0.59,仅使目标的 R 值降低了 10 个百分点,模型训练时间增加了 43%。攻击阳性预测值和敏感性也呈现出类似的趋势。最后,我们表明,IHS 中最需要强隐私保护的参与者也最容易受到这种特定隐私攻击的威胁,因此最能从这些隐私保护技术中受益。
我们的结果既证明了主动隐私保护研究的必要性,也证明了当前在真实 mHealth 场景中实施的 FL 和 DP 方法的可行性。我们的模拟方法使用高度可解释的指标来描述我们的 mHealth 设置中的隐私-效用权衡,为未来数据驱动的健康和医疗应用中的隐私保护技术研究提供了框架。