School of Health and Related Research (ScHARR), Regent Court, Sheffield, UK.
Health Technol Assess. 2010 May;14(25):iii-iv, ix-xii, 1-107. doi: 10.3310/hta14250.
Health policy decisions must be relevant, evidence-based and transparent. Decision-analytic modelling supports this process but its role is reliant on its credibility. Errors in mathematical decision models or simulation exercises are unavoidable but little attention has been paid to processes in model development. Numerous error avoidance/identification strategies could be adopted but it is difficult to evaluate the merits of strategies for improving the credibility of models without first developing an understanding of error types and causes.
The study aims to describe the current comprehension of errors in the HTA modelling community and generate a taxonomy of model errors. Four primary objectives are to: (1) describe the current understanding of errors in HTA modelling; (2) understand current processes applied by the technology assessment community for avoiding errors in development, debugging and critically appraising models for errors; (3) use HTA modellers' perceptions of model errors with the wider non-HTA literature to develop a taxonomy of model errors; and (4) explore potential methods and procedures to reduce the occurrence of errors in models. It also describes the model development process as perceived by practitioners working within the HTA community.
A methodological review was undertaken using an iterative search methodology. Exploratory searches informed the scope of interviews; later searches focused on issues arising from the interviews. Searches were undertaken in February 2008 and January 2009. In-depth qualitative interviews were performed with 12 HTA modellers from academic and commercial modelling sectors.
All qualitative data were analysed using the Framework approach. Descriptive and explanatory accounts were used to interrogate the data within and across themes and subthemes: organisation, roles and communication; the model development process; definition of error; types of model error; strategies for avoiding errors; strategies for identifying errors; and barriers and facilitators.
There was no common language in the discussion of modelling errors and there was inconsistency in the perceived boundaries of what constitutes an error. Asked about the definition of model error, there was a tendency for interviewees to exclude matters of judgement from being errors and focus on 'slips' and 'lapses', but discussion of slips and lapses comprised less than 20% of the discussion on types of errors. Interviewees devoted 70% of the discussion to softer elements of the process of defining the decision question and conceptual modelling, mostly the realms of judgement, skills, experience and training. The original focus concerned model errors, but it may be more useful to refer to modelling risks. Several interviewees discussed concepts of validation and verification, with notable consistency in interpretation: verification meaning the process of ensuring that the computer model correctly implemented the intended model, whereas validation means the process of ensuring that a model is fit for purpose. Methodological literature on verification and validation of models makes reference to the Hermeneutic philosophical position, highlighting that the concept of model validation should not be externalized from the decision-makers and the decision-making process. Interviewees demonstrated examples of all major error types identified in the literature: errors in the description of the decision problem, in model structure, in use of evidence, in implementation of the model, in operation of the model, and in presentation and understanding of results. The HTA error classifications were compared against existing classifications of model errors in the literature. A range of techniques and processes are currently used to avoid errors in HTA models: engaging with clinical experts, clients and decision-makers to ensure mutual understanding, producing written documentation of the proposed model, explicit conceptual modelling, stepping through skeleton models with experts, ensuring transparency in reporting, adopting standard housekeeping techniques, and ensuring that those parties involved in the model development process have sufficient and relevant training. Clarity and mutual understanding were identified as key issues. However, their current implementation is not framed within an overall strategy for structuring complex problems.
Some of the questioning may have biased interviewees responses but as all interviewees were represented in the analysis no rebalancing of the report was deemed necessary. A potential weakness of the literature review was its focus on spreadsheet and program development rather than specifically on model development. It should also be noted that the identified literature concerning programming errors was very narrow despite broad searches being undertaken.
Published definitions of overall model validity comprising conceptual model validation, verification of the computer model, and operational validity of the use of the model in addressing the real-world problem are consistent with the views expressed by the HTA community and are therefore recommended as the basis for further discussions of model credibility. Such discussions should focus on risks, including errors of implementation, errors in matters of judgement and violations. Discussions of modelling risks should reflect the potentially complex network of cognitive breakdowns that lead to errors in models and existing research on the cognitive basis of human error should be included in an examination of modelling errors. There is a need to develop a better understanding of the skills requirements for the development, operation and use of HTA models. Interaction between modeller and client in developing mutual understanding of a model establishes that model's significance and its warranty. This highlights that model credibility is the central concern of decision-makers using models so it is crucial that the concept of model validation should not be externalized from the decision-makers and the decision-making process. Recommendations for future research would be studies of verification and validation; the model development process; and identification of modifications to the modelling process with the aim of preventing the occurrence of errors and improving the identification of errors in models.
卫生政策决策必须具有相关性、基于证据且透明。决策分析模型支持这一过程,但它的作用取决于其可信度。数学决策模型或模拟练习中的错误是不可避免的,但很少有人关注模型开发过程中的错误。可以采用许多避免/识别策略,但如果不首先了解错误类型和原因,就很难评估提高模型可信度的策略的优点。
本研究旨在描述 HTA 建模界中错误的当前理解,并生成模型错误分类法。四项主要目标是:(1)描述 HTA 建模中错误的当前理解;(2)了解技术评估界避免模型开发、调试和批判性评估模型错误过程中应用的当前流程;(3)使用 HTA 建模者对模型错误的看法以及更广泛的非 HTA 文献,开发模型错误分类法;(4)探索减少模型中错误发生的潜在方法和程序。它还描述了在 HTA 社区工作的从业者对模型开发过程的看法。
采用迭代搜索方法进行了方法学综述。探索性搜索为访谈的范围提供了信息;后来的搜索集中在访谈中出现的问题上。搜索于 2008 年 2 月和 2009 年 1 月进行。对来自学术和商业建模部门的 12 名 HTA 建模者进行了深入的定性访谈。
所有定性数据均使用框架方法进行分析。使用描述性和解释性方法在主题和子主题内部和之间进行了数据的询问:组织、角色和沟通;模型开发过程;错误的定义;模型错误的类型;避免错误的策略;识别错误的策略;以及障碍和促进因素。
在讨论建模错误时没有共同的语言,并且对构成错误的界限存在不一致的看法。当被问及模型错误的定义时,受访者倾向于将判断排除在错误之外,而将重点放在“失误”和“疏忽”上,但对失误和疏忽的讨论仅占错误类型讨论的不到 20%。受访者将 70%的讨论集中在定义决策问题和概念建模的更软的过程上,主要是判断、技能、经验和培训领域。最初的重点是模型错误,但更有用的可能是参考建模风险。几位受访者讨论了验证和验证的概念,解释上存在显著的一致性:验证意味着确保计算机模型正确实现预期模型的过程,而验证意味着确保模型适合目的的过程。关于模型验证和验证的方法学文献提到了诠释哲学立场,强调模型验证的概念不应从决策者和决策过程中排除。受访者展示了文献中所有主要错误类型的示例:决策问题描述中的错误、模型结构中的错误、证据使用中的错误、模型实施中的错误、模型操作中的错误以及结果的呈现和理解中的错误。HTA 错误分类与文献中现有的模型错误分类法进行了比较。目前在 HTA 模型中采用了一系列技术和流程来避免错误:与临床专家、客户和决策者进行接触,以确保相互理解;制定拟议模型的书面文件;明确概念建模;与专家一起逐步建立模型;确保报告的透明度;采用标准的内务管理技术;并确保参与模型开发过程的各方具有足够和相关的培训。受访者认为清晰度和相互理解是关键问题。然而,它们目前的实施并未纳入解决复杂问题的总体战略框架中。
一些提问可能会影响受访者的回答,但由于所有受访者都在分析中得到了代表,因此不需要重新平衡报告。文献综述的一个潜在弱点是它侧重于电子表格和程序开发,而不是专门的模型开发。还应该注意的是,尽管进行了广泛的搜索,但确定的有关编程错误的文献非常有限。
发表的整体模型有效性定义包括概念模型验证、计算机模型验证以及在解决实际问题中使用模型的操作性有效性,与 HTA 社区表达的观点一致,因此建议作为进一步讨论模型可信度的基础。此类讨论应侧重于风险,包括实施错误、判断失误和违规行为。建模风险的讨论应反映导致模型错误的潜在复杂认知崩溃网络,并且应包括对人类错误认知基础的现有研究,以检查建模错误。需要更好地了解 HTA 模型的开发、操作和使用所需的技能要求。建模者和客户在开发对模型的共同理解方面的互动建立了模型的重要性及其保证。这强调了决策者使用模型的关键关注点是模型的可信度,因此模型验证的概念不应从决策者和决策过程中排除。未来研究的建议将是验证和验证研究;模型开发过程;以及旨在防止错误发生和提高模型中错误识别的建模过程的修改。