Suppr超能文献

在专家会议环境下对医疗软件进行可用性评估。

A usability evaluation of medical software at an expert conference setting.

机构信息

University of Ulster, Northern Ireland, UK.

出版信息

Comput Methods Programs Biomed. 2014;113(1):383-95. doi: 10.1016/j.cmpb.2013.10.006. Epub 2013 Oct 21.

Abstract

INTRODUCTION

A usability test was employed to evaluate two medical software applications at an expert conference setting. One software application is a medical diagnostic tool (electrocardiogram [ECG] viewer) and the other is a medical research tool (electrode misplacement simulator [EMS]). These novel applications have yet to be adopted by the healthcare domain, thus, (1) we wanted to determine the potential user acceptance of these applications and (2) we wanted to determine the feasibility of evaluating medical diagnostic and medical research software at a conference setting as opposed to the conventional laboratory setting.

METHODS

The medical diagnostic tool (ECG viewer) was evaluated using seven delegates and the medical research tool (EMS) was evaluated using 17 delegates that were recruited at the 2010 International Conference on Computing in Cardiology. Each delegate/participant was required to use the software and undertake a set of predefined tasks during the session breaks at the conference. User interactions with the software were recorded using screen-recording software. The 'think-aloud' protocol was also used to elicit verbal feedback from the participants whilst they attempted the pre-defined tasks. Before and after each session, participants completed a pre-test and a post-test questionnaire respectively.

RESULTS

The average duration of a usability session at the conference was 34.69 min (SD=10.28). However, taking into account that 10 min was dedicated to the pre-test and post-test questionnaires, the average time dedication to user interaction of the medical software was 24.69 min (SD=10.28). Given we have shown that usability data can be collected at conferences, this paper details the advantages of conference-based usability studies over the laboratory-based approach. For example, given delegates gather at one geographical location, a conference-based usability evaluation facilitates recruitment of a convenient sample of international subject experts. This would otherwise be very expensive to arrange. A conference-based approach also allows for data to be collected over a few days as opposed to months by avoiding administration duties normally involved in laboratory based approach, e.g. mailing invitation letters as part of a recruitment campaign. Following analysis of the user video recordings, 41 (previously unknown) use errors were identified in the advanced ECG viewer and 29 were identified in the EMS application. All use errors were given a consensus severity rating from two independent usability experts. Out of a rating scale of 4 (where 1=cosmetic and 4=critical), the average severity rating for the ECG viewer was 2.24 (SD=1.09) and the average severity rating for the EMS application was 2.34 (SD=0.97). We were also able to extract task completion rates and times from the video recordings to determine the effectiveness of the software applications. For example, six out of seven tasks were completed by all participants when using both applications. This statistic alone suggests both applications already have a high degree of usability. As well as extracting data from the video recordings, we were also able to extract data from the questionnaires. Using a semantic differential scale (where 1=poor and 5=excellent), delegates highly rated the 'responsiveness', 'usefulness', 'learnability' and the 'look and feel' of both applications.

CONCLUSION

This study has shown the potential user acceptance and user-friendliness of the novel EMS and the ECG viewer applications within the healthcare domain. It has also shown that both medical diagnostic software and medical research software can be evaluated for their usability at an expert conference setting. The primary advantage of a conference-based usability evaluation over a laboratory-based evaluation is the high concentration of experts at one location, which is convenient, less time consuming and less expensive.

摘要

简介

在专家会议环境中,我们采用可用性测试来评估两种医学软件应用程序。一种软件应用程序是一种医学诊断工具(心电图 [ECG] 查看器),另一种是一种医学研究工具(电极放置错误模拟器 [EMS])。这些新应用程序尚未被医疗保健领域采用,因此,(1)我们想确定这些应用程序的潜在用户接受程度,(2)我们想确定在会议环境中评估医学诊断和医学研究软件的可行性,而不是传统的实验室环境。

方法

使用 7 名代表评估医学诊断工具(ECG 查看器),使用在 2010 年国际心脏病学计算会议上招募的 17 名代表评估医学研究工具(EMS)。要求每位代表/参与者在会议休息期间使用软件并完成一组预定义任务。使用屏幕录制软件记录软件的用户交互。还使用“出声思维”协议从参与者那里获取口头反馈,因为他们尝试预定义的任务。在每次会议前后,参与者分别完成预测试和后测试问卷。

结果

会议上的可用性会议平均持续时间为 34.69 分钟(SD=10.28)。然而,考虑到 10 分钟用于预测试和后测试问卷,用于医疗软件用户交互的平均时间为 24.69 分钟(SD=10.28)。由于我们已经表明可以在会议上收集可用性数据,因此本文详细介绍了基于会议的可用性研究相对于基于实验室的方法的优势。例如,由于代表们聚集在一个地理位置,基于会议的可用性评估有助于方便地招募国际专家作为研究对象。否则,这将非常昂贵。基于会议的方法还可以在几天内收集数据,而不是通过避免实验室方法通常涉及的管理职责来收集几个月的数据,例如作为招聘活动的一部分发送邀请信。在分析用户视频记录后,在高级 ECG 查看器中发现了 41 个(以前未知的)使用错误,在 EMS 应用程序中发现了 29 个。所有使用错误都由两位独立的可用性专家进行了共识严重程度评分。在 4 分制(其中 1=轻微,4=严重)中,ECG 查看器的平均严重程度评分为 2.24(SD=1.09),EMS 应用程序的平均严重程度评分为 2.34(SD=0.97)。我们还能够从视频记录中提取任务完成率和时间,以确定软件应用程序的有效性。例如,当使用这两个应用程序时,七名参与者中的六名完成了所有任务。仅这一统计数据表明这两个应用程序已经具有很高的可用性。除了从视频记录中提取数据外,我们还能够从问卷中提取数据。使用语义差异量表(其中 1=差,5=优),代表们高度评价了这两个应用程序的“响应性”、“有用性”、“易学性”和“外观和感觉”。

结论

本研究表明,在医疗保健领域,新型 EMS 和 ECG 查看器应用程序具有潜在的用户接受度和用户友好性。它还表明,医学诊断软件和医学研究软件都可以在专家会议环境中评估其可用性。基于会议的可用性评估相对于基于实验室的评估的主要优势是专家在一个地点的高度集中,这既方便、省时又省钱。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验