Suppr超能文献

基于剑桥神经心理测试自动化电池的网络与实验室认知评估比较:一项被试内平衡研究。

Comparing Web-Based and Lab-Based Cognitive Assessment Using the Cambridge Neuropsychological Test Automated Battery: A Within-Subjects Counterbalanced Study.

机构信息

Cambridge Cognition Ltd, Cambridge, United Kingdom.

School of Psychological Science, University of Bristol, Bristol, United Kingdom.

出版信息

J Med Internet Res. 2020 Aug 4;22(8):e16792. doi: 10.2196/16792.

Abstract

BACKGROUND

Computerized assessments are already used to derive accurate and reliable measures of cognitive function. Web-based cognitive assessment could improve the accessibility and flexibility of research and clinical assessment, widen participation, and promote research recruitment while simultaneously reducing costs. However, differences in context may influence task performance.

OBJECTIVE

This study aims to determine the comparability of an unsupervised, web-based administration of the Cambridge Neuropsychological Test Automated Battery (CANTAB) against a typical in-person lab-based assessment, using a within-subjects counterbalanced design. The study aims to test (1) reliability, quantifying the relationship between measurements across settings using correlational approaches; (2) equivalence, the extent to which test results in different settings produce similar overall results; and (3) agreement, by quantifying acceptable limits to bias and differences between measurement environments.

METHODS

A total of 51 healthy adults (32 women and 19 men; mean age 36.8, SD 15.6 years) completed 2 testing sessions, which were completed on average 1 week apart (SD 4.5 days). Assessments included equivalent tests of emotion recognition (emotion recognition task [ERT]), visual recognition (pattern recognition memory [PRM]), episodic memory (paired associate learning [PAL]), working memory and spatial planning (spatial working memory [SWM] and one touch stockings of Cambridge), and sustained attention (rapid visual information processing [RVP]). Participants were randomly allocated to one of the two groups, either assessed in-person in the laboratory first (n=33) or with unsupervised web-based assessments on their personal computing systems first (n=18). Performance indices (errors, correct trials, and response sensitivity) and median reaction times were extracted. Intraclass and bivariate correlations examined intersetting reliability, linear mixed models and Bayesian paired sample t tests tested for equivalence, and Bland-Altman plots examined agreement.

RESULTS

Intraclass correlation (ICC) coefficients ranged from ρ=0.23-0.67, with high correlations in 3 performance indices (from PAL, SWM, and RVP tasks; ρ≥0.60). High ICC values were also seen for reaction time measures from 2 tasks (PRM and ERT tasks; ρ≥0.60). However, reaction times were slower during web-based assessments, which undermined both equivalence and agreement for reaction time measures. Performance indices did not differ between assessment settings and generally showed satisfactory agreement.

CONCLUSIONS

Our findings support the comparability of CANTAB performance indices (errors, correct trials, and response sensitivity) in unsupervised, web-based assessments with in-person and laboratory tests. Reaction times are not as easily translatable from in-person to web-based testing, likely due to variations in computer hardware. The results underline the importance of examining more than one index to ascertain comparability, as high correlations can present in the context of systematic differences, which are a product of differences between measurement environments. Further work is now needed to examine web-based assessments in clinical populations and in larger samples to improve sensitivity for detecting subtler differences between test settings.

摘要

背景

计算机化评估已经用于得出准确可靠的认知功能测量结果。基于网络的认知评估可以提高研究和临床评估的可及性和灵活性,扩大参与度,并促进研究招募,同时降低成本。然而,不同的背景可能会影响任务表现。

目的

本研究旨在使用被试内对照平衡设计,确定无监督的基于网络的剑桥神经心理学测试自动化电池(CANTAB)与典型的现场基于实验室的评估之间的可比性。该研究旨在测试(1)可靠性,使用相关方法量化不同设置之间测量的关系;(2)等效性,即不同设置下的测试结果产生相似总体结果的程度;以及(3)一致性,通过量化偏差和测量环境之间差异的可接受极限来进行。

方法

共有 51 名健康成年人(32 名女性和 19 名男性;平均年龄 36.8 岁,标准差 15.6 岁)完成了 2 次测试,平均相隔 1 周(标准差 4.5 天)完成。评估包括等效的情绪识别测试(情绪识别任务[ERT])、视觉识别测试(模式识别记忆[PRM])、情景记忆测试(配对联想学习[PAL])、工作记忆和空间规划测试(空间工作记忆[SWM]和剑桥的一次触摸袜子)以及持续注意力测试(快速视觉信息处理[RVP])。参与者被随机分配到两个组中的一个,要么先在实验室进行现场评估(n=33),要么先使用基于网络的个人计算机系统进行无监督的在线评估(n=18)。提取性能指标(错误、正确试验和反应敏感性)和中位数反应时间。使用组内相关和双变量相关性检验设置间可靠性,使用线性混合模型和贝叶斯配对样本 t 检验检验等效性,使用 Bland-Altman 图检验一致性。

结果

组内相关(ICC)系数范围为ρ=0.23-0.67,3 项性能指标(来自 PAL、SWM 和 RVP 任务;ρ≥0.60)具有高相关性。来自 2 项任务(PRM 和 ERT 任务)的反应时间测量也具有较高的 ICC 值(ρ≥0.60)。然而,基于网络的评估中的反应时间较慢,这破坏了反应时间测量的等效性和一致性。性能指标在评估设置之间没有差异,并且通常具有令人满意的一致性。

结论

我们的研究结果支持 CANTAB 性能指标(错误、正确试验和反应敏感性)在无监督的基于网络的评估与现场和实验室测试之间的可比性。反应时间不容易从现场测试转换为基于网络的测试,可能是由于计算机硬件的变化所致。研究结果强调了检查多个指标以确定可比性的重要性,因为在存在系统差异的情况下,高相关性可能会出现,而系统差异是测量环境差异的产物。现在需要进一步研究在临床人群和更大的样本中进行基于网络的评估,以提高检测测试设置之间更细微差异的敏感性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cb57/7435628/d94cbc6ed9cc/jmir_v22i8e16792_fig1.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验