Burr Tom, Graves Todd, Klamann Richard, Michalak Sarah, Picard Richard, Hengartner Nicolas
Statistical Sciences, Mail Stop F600, Los Alamos National Laboratory, Los Alamos, NM 87545, USA.
BMC Med Inform Decis Mak. 2006 Dec 4;6:40. doi: 10.1186/1472-6947-6-40.
Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from year to year. Our goal is to evaluate the impact of inconsistent seasonal effects on performance assessments (false and true positive rates) in the context of detecting anomalous counts in data that exhibit seasonal variation.
To evaluate the impact of inconsistent seasonal effects, we injected synthetic outbreaks into real data and into data simulated from each of two models fit to the same real data. Using real respiratory syndrome counts collected in an emergency department from 2/1/94-5/31/03, we varied the length of training data from one to eight years, applied a sequential test to the forecast errors arising from each of eight forecasting methods, and evaluated their detection probabilities (DP) on the basis of 1000 injected synthetic outbreaks. We did the same for each of two corresponding simulated data sets. The less realistic, nonhierarchical model's simulated data set assumed that "one season fits all," meaning that each year's seasonal peak has the same onset, duration, and magnitude. The more realistic simulated data set used a hierarchical model to capture violation of the "one season fits all" assumption.
This experiment demonstrated optimistic bias in DP estimates for some of the methods when data simulated from the nonhierarchical model was used for DP estimation, thus suggesting that at least for some real data sets and methods, it is not adequate to assume that "one season fits all."
For the data we analyze, the "one season fits all " assumption is violated, and DP performance claims based on simulated data that assume "one season fits all," for the forecast methods considered, except for moving average methods, tend to be optimistic. Moving average methods based on relatively short amounts of training data are competitive on all three data sets, but are particularly competitive on the real data and on data from the hierarchical model, which are the two data sets that violate the "one season fits all" assumption.
症状监测(SS)通过提供及时、新颖的数据源,可能有助于提高疫情检测能力。症状监测面临的一个挑战是,某些症状计数随季节变化,且每年的变化方式不尽相同。我们的目标是在检测呈现季节性变化的数据中的异常计数时,评估不一致的季节效应对性能评估(假阳性率和真阳性率)的影响。
为评估不一致的季节效应的影响,我们将合成疫情注入真实数据以及从拟合同一真实数据的两个模型之一模拟出的数据中。利用1994年2月1日至2003年5月31日在急诊科收集的真实呼吸道症状计数,我们将训练数据的长度从一年变化到八年,对八种预测方法各自产生的预测误差应用序贯检验,并基于1000次注入的合成疫情评估它们的检测概率(DP)。我们对两个相应的模拟数据集也进行了同样的操作。不太现实的非分层模型的模拟数据集假定“一个季节适用于所有情况”,即每年的季节高峰具有相同的起始时间、持续时间和幅度。更现实的模拟数据集使用分层模型来捕捉对“一个季节适用于所有情况”这一假设的违背。
当使用从非分层模型模拟出的数据进行DP估计时,该实验证明了某些方法的DP估计存在乐观偏差,这表明至少对于一些真实数据集和方法而言,假定“一个季节适用于所有情况”是不够的。
对于我们分析的数据,“一个季节适用于所有情况”的假设不成立,对于所考虑的预测方法,基于假定“一个季节适用于所有情况”的模拟数据的DP性能声明往往过于乐观,移动平均方法除外。基于相对少量训练数据的移动平均方法在所有三个数据集上都具有竞争力,但在真实数据和分层模型数据上尤其具有竞争力,这两个数据集违背了“一个季节适用于所有情况”的假设。