Cheng Yuan, Zhang Jianing, Zhou Tiankuang, Wang Yuyan, Xu Zhihao, Yuan Xiaoyun, Fang Lu
Sigma Laboratory, Department of Electronic Engineering, Tsinghua University, Beijing, 100084, China.
Beijing National Research Center for Information Science and Technology (BNRist), Beijing, 100084, China.
Light Sci Appl. 2024 Feb 26;13(1):56. doi: 10.1038/s41377-024-01395-4.
Scalable, high-capacity, and low-power computing architecture is the primary assurance for increasingly manifold and large-scale machine learning tasks. Traditional electronic artificial agents by conventional power-hungry processors have faced the issues of energy and scaling walls, hindering them from the sustainable performance improvement and iterative multi-task learning. Referring to another modality of light, photonic computing has been progressively applied in high-efficient neuromorphic systems. Here, we innovate a reconfigurable lifelong-learning optical neural network (LONN), for highly-integrated tens-of-task machine intelligence with elaborated algorithm-hardware co-design. Benefiting from the inherent sparsity and parallelism in massive photonic connections, LONN learns each single task by adaptively activating sparse photonic neuron connections in the coherent light field, while incrementally acquiring expertise on various tasks by gradually enlarging the activation. The multi-task optical features are parallelly processed by multi-spectrum representations allocated with different wavelengths. Extensive evaluations on free-space and on-chip architectures confirm that for the first time, LONN avoided the catastrophic forgetting issue of photonic computing, owning versatile skills on challenging tens-of-tasks (vision classification, voice recognition, medical diagnosis, etc.) with a single model. Particularly, LONN achieves more than an order of magnitude higher efficiency than the representative electronic artificial neural networks, and 14× larger capacity than existing optical neural networks while maintaining competitive performance on each individual task. The proposed photonic neuromorphic architecture points out a new form of lifelong learning scheme, permitting terminal/edge AI systems with light-speed efficiency and unprecedented scalability.
可扩展、高容量和低功耗的计算架构是日益多样化和大规模机器学习任务的主要保障。传统的由耗电处理器驱动的电子智能体面临着能源和规模瓶颈问题,阻碍了它们持续提升性能和进行迭代多任务学习。参照光的另一种模态,光子计算已逐渐应用于高效的神经形态系统。在此,我们创新了一种可重构的终身学习光学神经网络(LONN),通过精心设计的算法 - 硬件协同设计实现高度集成的数十任务机器智能。受益于大量光子连接中固有的稀疏性和并行性,LONN通过在相干光场中自适应激活稀疏的光子神经元连接来学习每个单一任务,同时通过逐渐扩大激活范围逐步获取各种任务的专业知识。多任务光学特征由分配不同波长的多光谱表示并行处理。在自由空间和片上架构上的广泛评估证实,LONN首次避免了光子计算中的灾难性遗忘问题,在单个模型上具备处理具有挑战性的数十任务(视觉分类、语音识别、医学诊断等)的通用技能。特别地,LONN的效率比代表性的电子人工神经网络高出一个数量级以上,容量比现有光学神经网络大14倍,同时在每个单独任务上保持有竞争力的性能。所提出的光子神经形态架构指出了一种新的终身学习方案形式,使终端/边缘人工智能系统具有光速效率和前所未有的可扩展性。