Fang Xinyuan, Hu Xiaonan, Li Baoli, Su Hang, Cheng Ke, Luan Haitao, Gu Min
Institute of Photonic Chips, University of Shanghai for Science and Technology, Shanghai, 200093, China.
Centre for Artificial-Intelligence Nanophotonics, School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai, 200093, China.
Light Sci Appl. 2024 Feb 14;13(1):49. doi: 10.1038/s41377-024-01386-5.
Machine learning with optical neural networks has featured unique advantages of the information processing including high speed, ultrawide bandwidths and low energy consumption because the optical dimensions (time, space, wavelength, and polarization) could be utilized to increase the degree of freedom. However, due to the lack of the capability to extract the information features in the orbital angular momentum (OAM) domain, the theoretically unlimited OAM states have never been exploited to represent the signal of the input/output nodes in the neural network model. Here, we demonstrate OAM-mediated machine learning with an all-optical convolutional neural network (CNN) based on Laguerre-Gaussian (LG) beam modes with diverse diffraction losses. The proposed CNN architecture is composed of a trainable OAM mode-dispersion impulse as a convolutional kernel for feature extraction, and deep-learning diffractive layers as a classifier. The resultant OAM mode-dispersion selectivity can be applied in information mode-feature encoding, leading to an accuracy as high as 97.2% for MNIST database through detecting the energy weighting coefficients of the encoded OAM modes, as well as a resistance to eavesdropping in point-to-point free-space transmission. Moreover, through extending the target encoded modes into multiplexed OAM states, we realize all-optical dimension reduction for anomaly detection with an accuracy of 85%. Our work provides a deep insight to the mechanism of machine learning with spatial modes basis, which can be further utilized to improve the performances of various machine-vision tasks by constructing the unsupervised learning-based auto-encoder.
基于光学神经网络的机器学习具有信息处理的独特优势,包括高速、超宽带宽和低能耗,这是因为光学维度(时间、空间、波长和偏振)可用于增加自由度。然而,由于缺乏在轨道角动量(OAM)域中提取信息特征的能力,理论上无限的OAM态从未被用于表示神经网络模型中输入/输出节点的信号。在此,我们展示了基于具有不同衍射损耗的拉盖尔 - 高斯(LG)光束模式的全光卷积神经网络(CNN)的OAM介导的机器学习。所提出的CNN架构由作为特征提取卷积核的可训练OAM模式色散脉冲和作为分类器的深度学习衍射层组成。由此产生的OAM模式色散选择性可应用于信息模式特征编码,通过检测编码OAM模式的能量加权系数,在MNIST数据库上实现高达97.2%的准确率,以及在点对点自由空间传输中的抗窃听能力。此外,通过将目标编码模式扩展到复用OAM态,我们实现了用于异常检测的全光降维,准确率为85%。我们的工作为基于空间模式的机器学习机制提供了深入见解,可进一步用于通过构建基于无监督学习的自动编码器来提高各种机器视觉任务的性能。