Zhan Zhenye, Gao Yulu, Liao Yue, Xie Weiguang, Liu Si, Wang Xiaomu
Siyuan Laboratory, Guangdong Provincial Engineering Technology Research Center of Vacuum Coating Technologies and New Energy Materials, Department of Physics, Jinan University, Guangzhou, Guangdong, 510632, China.
Institute of Artificial Intelligence, Beihang University, Beijing, 100191, China.
Adv Sci (Weinh). 2025 Sep;12(33):e04706. doi: 10.1002/advs.202504706. Epub 2025 Jul 13.
Memristive computing refers to the hardware implementation of artificial neural networks (ANNs) by employing memristive devices. It supports analog multiply-and-accumulation (MAC) operation in a compact and highly parallel manner, which can significantly enhance computing efficiency. However, applying memristive computing in advanced network structures, such as deep neural networks and multimodal networks, is inefficient because the partial analog computing requires frequently exchanging data between analog and digital domains. Here, a perovskite memristive computing unit with flexible reconfigurability and desired nonlinearity through fully vapor deposition is reported. It enables performing all the mathematical operations necessary for Transformer ANNs completely in the analog domain. A prototypical attention module is implemented by combining cells configured in different operators of dynamic MAC, activation, and softmax functions. By cascading the modules in a multi-layer Transformer network, a neuromorphic engine is fabricated and tested RGB-T tracking and visual question answering tasks, fully considering device non-idealities. It is found that the network performance is close to that of operating on a graphics processing unit (GPU)-accelerated workstation, but it consumes only 1.7% energy and increases power efficiency by 58 times. The results pave a new way toward efficient and accurate hardware memristive computing for advanced ANNs.
忆阻计算是指通过使用忆阻器件对人工神经网络(ANN)进行硬件实现。它以紧凑且高度并行的方式支持模拟乘法累加(MAC)运算,这可显著提高计算效率。然而,在深度神经网络和多模态网络等先进网络结构中应用忆阻计算效率低下,因为部分模拟计算需要在模拟域和数字域之间频繁交换数据。在此,报道了一种通过全气相沉积具有灵活可重构性和所需非线性的钙钛矿忆阻计算单元。它能够在模拟域中完全执行Transformer人工神经网络所需的所有数学运算。通过组合配置有动态MAC、激活和softmax函数等不同算子的单元来实现一个典型的注意力模块。通过在多层Transformer网络中级联这些模块,制造并测试了一个神经形态引擎用于RGB-T跟踪和视觉问答任务,充分考虑了器件的非理想性。结果发现,该网络的性能接近在图形处理单元(GPU)加速的工作站上运行的性能,但它仅消耗1.7%的能量,功率效率提高了58倍。这些结果为先进人工神经网络的高效且准确的硬件忆阻计算铺平了一条新道路。