IEEE Trans Med Imaging. 2022 Oct;41(10):2814-2827. doi: 10.1109/TMI.2022.3170701. Epub 2022 Sep 30.
Constructing and analyzing functional brain networks (FBN) has become a promising approach to brain disorder classification. However, the conventional successive construct-and-analyze process would limit the performance due to the lack of interactions and adaptivity among the subtasks in the process. Recently, Transformer has demonstrated remarkable performance in various tasks, attributing to its effective attention mechanism in modeling complex feature relationships. In this paper, for the first time, we develop Transformer for integrated FBN modeling, analysis and brain disorder classification with rs-fMRI data by proposing a Diffusion Kernel Attention Network to address the specific challenges. Specifically, directly applying Transformer does not necessarily admit optimal performance in this task due to its extensive parameters in the attention module against the limited training samples usually available. Looking into this issue, we propose to use kernel attention to replace the original dot-product attention module in Transformer. This significantly reduces the number of parameters to train and thus alleviates the issue of small sample while introducing a non-linear attention mechanism to model complex functional connections. Another limit of Transformer for FBN applications is that it only considers pair-wise interactions between directly connected brain regions but ignores the important indirect connections. Therefore, we further explore diffusion process over the kernel attention to incorporate wider interactions among indirectly connected brain regions. Extensive experimental study is conducted on ADHD-200 data set for ADHD classification and on ADNI data set for Alzheimer's disease classification, and the results demonstrate the superior performance of the proposed method over the competing methods.
构建和分析功能脑网络(FBN)已成为一种有前途的脑疾病分类方法。然而,由于过程中各子任务之间缺乏交互和适应性,传统的连续构建和分析过程会限制性能。最近,Transformer 在各种任务中表现出色,这归因于其在建模复杂特征关系方面的有效注意力机制。在本文中,我们首次通过提出扩散核注意力网络,为 rs-fMRI 数据的集成 FBN 建模、分析和脑疾病分类开发了 Transformer,以解决特定挑战。具体来说,由于注意力模块中的广泛参数与通常可用的有限训练样本相冲突,直接应用 Transformer 不一定能在这项任务中获得最佳性能。针对这个问题,我们建议使用核注意力来替代 Transformer 中的原始点积注意力模块。这大大减少了需要训练的参数数量,从而缓解了小样本量的问题,同时引入了一种非线性注意力机制来模拟复杂的功能连接。Transformer 在 FBN 应用中的另一个限制是,它仅考虑直接连接的脑区之间的两两相互作用,而忽略了重要的间接连接。因此,我们进一步探索了核注意力上的扩散过程,以纳入间接连接的脑区之间更广泛的相互作用。我们在 ADHD-200 数据集上进行了 ADHD 分类的广泛实验研究,并在 ADNI 数据集上进行了阿尔茨海默病分类的实验研究,结果表明,所提出的方法优于竞争方法。