Wu Peiliang, Zhang Haozhe, Li Yao, Chen Wenbai, Gao Guowei
IEEE Trans Haptics. 2024 Sep 27;PP. doi: 10.1109/TOH.2024.3449411.
Current issues with neuromorphic visual-tactile perception include limited training network representation and inadequate cross-modal fusion. To address these two issues, we proposed a dual network called visual-tactile spiking graph neural network (VT-SGN) that combines graph neural networks and spiking neural networks to jointly utilize the neuromorphic visual and tactile source data. First, the neuromorphic visual-tactile data were expanded spatiotemporally to create a taxel-based tactile graph in the spatial domain, enabling the complete exploitation of the irregular spatial structure properties of tactile information. Subsequently, a method for converting images into graph structures was proposed, allowing the vision to be trained alongside graph neural networks and extracting graph-level features from the vision for fusion with tactile data. Finally, the data were expanded into the time domain using a spiking neural network to train the model and propagate it backwards. This framework effectively utilizes the structural differences between sample instances in the spatial dimension to improve the representational power of spiking neurons, while preserving the biodynamic mechanism of the spiking neural network. Additionally, it effectively solves the morphological variance between the two perceptions and further uses complementary data between visual and tactile. To demonstrate that our approach can improve the learning of neuromorphic perceptual information, we conducted comprehensive comparative experiments on three datasets to validate the benefits of the proposed VT-SGN framework by comparing it with state-of-the-art studies.
当前神经形态视觉触觉感知存在的问题包括训练网络表示有限和跨模态融合不足。为了解决这两个问题,我们提出了一种名为视觉触觉脉冲图神经网络(VT-SGN)的双网络,它结合了图神经网络和脉冲神经网络,以联合利用神经形态视觉和触觉源数据。首先,对神经形态视觉触觉数据进行时空扩展,在空间域中创建基于体素的触觉图,从而能够充分利用触觉信息的不规则空间结构特性。随后,提出了一种将图像转换为图结构的方法,使视觉能够与图神经网络一起训练,并从视觉中提取图级特征以与触觉数据融合。最后,使用脉冲神经网络将数据扩展到时间域来训练模型并进行反向传播。该框架有效地利用了空间维度中样本实例之间的结构差异,以提高脉冲神经元的表示能力,同时保留脉冲神经网络的生物动力学机制。此外,它有效地解决了两种感知之间的形态差异,并进一步利用视觉和触觉之间的互补数据。为了证明我们的方法可以改善神经形态感知信息的学习,我们在三个数据集上进行了全面的对比实验,通过与现有研究进行比较来验证所提出的VT-SGN框架的优势。