• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

光学生成模型。

Optical generative models.

作者信息

Chen Shiqi, Li Yuhang, Wang Yuntian, Chen Hanlong, Ozcan Aydogan

机构信息

Electrical and Computer Engineering Department, University of California Los Angeles, Los Angeles, CA, USA.

Bioengineering Department, University of California Los Angeles, Los Angeles, CA, USA.

出版信息

Nature. 2025 Aug;644(8078):903-911. doi: 10.1038/s41586-025-09446-5. Epub 2025 Aug 27.

DOI:10.1038/s41586-025-09446-5
PMID:40866675
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12390839/
Abstract

Generative models cover various application areas, including image and video synthesis, natural language processing and molecular design, among many others. As digital generative models become larger, scalable inference in a fast and energy-efficient manner becomes a challenge. Here we present optical generative models inspired by diffusion models, where a shallow and fast digital encoder first maps random noise into phase patterns that serve as optical generative seeds for a desired data distribution; a jointly trained free-space-based reconfigurable decoder all-optically processes these generative seeds to create images never seen before following the target data distribution. Except for the illumination power and the random seed generation through a shallow encoder, these optical generative models do not consume computing power during the synthesis of the images. We report the optical generation of monochrome and multicolour images of handwritten digits, fashion products, butterflies, human faces and artworks, following the data distributions of MNIST, Fashion-MNIST, Butterflies-100, Celeb-A datasets, and Van Gogh's paintings and drawings, respectively, achieving an overall performance comparable to digital neural-network-based generative models. To experimentally demonstrate optical generative models, we used visible light to generate images of handwritten digits and fashion products. In addition, we generated Van Gogh-style artworks using both monochrome and multiwavelength illumination. These optical generative models might pave the way for energy-efficient and scalable inference tasks, further exploiting the potentials of optics and photonics for artificial-intelligence-generated content.

摘要

生成模型涵盖了各种应用领域,包括图像和视频合成、自然语言处理以及分子设计等等。随着数字生成模型规模的增大,以快速且节能的方式进行可扩展推理成为一项挑战。在此,我们展示受扩散模型启发的光学生成模型,其中一个浅层且快速的数字编码器首先将随机噪声映射为相位图案,这些图案作为期望数据分布的光学生成种子;一个联合训练的基于自由空间的可重构解码器全光处理这些生成种子,以按照目标数据分布创建前所未见的图像。除了照明功率以及通过浅层编码器生成随机种子外,这些光学生成模型在图像合成过程中不消耗计算能力。我们分别报告了按照MNIST、Fashion-MNIST、Butterflies-100、Celeb-A数据集以及梵高的绘画和素描的数据分布,对手写数字、时尚产品、蝴蝶、人脸和艺术品进行单色和多色图像的光学生成,总体性能与基于数字神经网络的生成模型相当。为了通过实验展示光学生成模型,我们使用可见光生成手写数字和时尚产品的图像。此外,我们使用单色和多波长照明生成了梵高风格的艺术品。这些光学生成模型可能为节能和可扩展推理任务铺平道路,进一步挖掘光学和光子学在人工智能生成内容方面的潜力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/78ee07ff45f1/41586_2025_9446_Fig15_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/f5c38ee22337/41586_2025_9446_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/882156d8f21f/41586_2025_9446_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/53c8e97334d6/41586_2025_9446_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/732462949a36/41586_2025_9446_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/b6f9cdbba92a/41586_2025_9446_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/2083f1bbc4cf/41586_2025_9446_Fig6_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/e3f2d0d302db/41586_2025_9446_Fig7_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/0139c0469af2/41586_2025_9446_Fig8_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/0ec2b3426037/41586_2025_9446_Fig9_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/74e154c9416b/41586_2025_9446_Fig10_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/3d43d2840b36/41586_2025_9446_Fig11_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/56130d32c80d/41586_2025_9446_Fig12_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/8d90c335f988/41586_2025_9446_Fig13_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/580fbcc9bf11/41586_2025_9446_Fig14_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/78ee07ff45f1/41586_2025_9446_Fig15_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/f5c38ee22337/41586_2025_9446_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/882156d8f21f/41586_2025_9446_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/53c8e97334d6/41586_2025_9446_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/732462949a36/41586_2025_9446_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/b6f9cdbba92a/41586_2025_9446_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/2083f1bbc4cf/41586_2025_9446_Fig6_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/e3f2d0d302db/41586_2025_9446_Fig7_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/0139c0469af2/41586_2025_9446_Fig8_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/0ec2b3426037/41586_2025_9446_Fig9_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/74e154c9416b/41586_2025_9446_Fig10_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/3d43d2840b36/41586_2025_9446_Fig11_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/56130d32c80d/41586_2025_9446_Fig12_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/8d90c335f988/41586_2025_9446_Fig13_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/580fbcc9bf11/41586_2025_9446_Fig14_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2616/12390839/78ee07ff45f1/41586_2025_9446_Fig15_ESM.jpg

相似文献

1
Optical generative models.光学生成模型。
Nature. 2025 Aug;644(8078):903-911. doi: 10.1038/s41586-025-09446-5. Epub 2025 Aug 27.
2
Improving reconstruction of patient-specific abnormalities in AI-driven fast MRI with an individually adapted diffusion model.利用个体适配的扩散模型改进人工智能驱动的快速磁共振成像中患者特异性异常的重建。
Med Phys. 2025 Jul;52(7):e17955. doi: 10.1002/mp.17955.
3
Exploring the Potential of Electroencephalography Signal-Based Image Generation Using Diffusion Models: Integrative Framework Combining Mixed Methods and Multimodal Analysis.利用扩散模型探索基于脑电图信号的图像生成潜力:结合混合方法和多模态分析的综合框架
JMIR Med Inform. 2025 Jun 25;13:e72027. doi: 10.2196/72027.
4
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
5
Artificial intelligence for diagnosing exudative age-related macular degeneration.人工智能在渗出性年龄相关性黄斑变性诊断中的应用。
Cochrane Database Syst Rev. 2024 Oct 17;10(10):CD015522. doi: 10.1002/14651858.CD015522.pub2.
6
Noise-aware system generative model (NASGM): positron emission tomography (PET) image simulation framework with observer validation studies.噪声感知系统生成模型(NASGM):用于正电子发射断层扫描(PET)图像模拟框架及观察者验证研究。
Med Phys. 2025 Jul;52(7):e17962. doi: 10.1002/mp.17962.
7
Short-Term Memory Impairment短期记忆障碍
8
A comprehensive comparative study of generative adversarial network architectures for synthetic computed tomography generation in the abdomen.用于腹部合成计算机断层扫描生成的生成对抗网络架构的全面比较研究。
Med Phys. 2025 Aug;52(8):e18038. doi: 10.1002/mp.18038.
9
Leveraging a foundation model zoo for cell similarity search in oncological microscopy across devices.利用基础模型库进行跨设备肿瘤显微镜检查中的细胞相似性搜索。
Front Oncol. 2025 Jun 18;15:1480384. doi: 10.3389/fonc.2025.1480384. eCollection 2025.
10
Exploring Generative Pre-Trained Transformer-4-Vision for Nystagmus Classification: Development and Validation of a Pupil-Tracking Process.探索用于眼球震颤分类的生成式预训练变换器-4视觉模型:瞳孔追踪过程的开发与验证
JMIR Form Res. 2025 Jun 6;9:e70070. doi: 10.2196/70070.

本文引用的文献

1
Broadband unidirectional visible imaging using wafer-scale nano-fabrication of multi-layer diffractive optical processors.利用多层衍射光学处理器的晶圆级纳米制造实现宽带单向可见光成像。
Light Sci Appl. 2025 Aug 11;14(1):267. doi: 10.1038/s41377-025-01971-2.
2
Combining one and two photon polymerization for accelerated high performance (3 + 1)D photonic integration.结合单光子和双光子聚合实现加速高性能(3 + 1)D光子集成。
Nanophotonics. 2022 Mar 29;11(8):1591-1601. doi: 10.1515/nanoph-2021-0733. eCollection 2022 Mar.
3
Nonlinear processing with linear optics.
基于线性光学的非线性处理
Nat Photonics. 2024;18(10):1076-1082. doi: 10.1038/s41566-024-01494-z. Epub 2024 Jul 31.
4
Nonlinear optical encoding enabled by recurrent linear scattering.由循环线性散射实现的非线性光学编码
Nat Photonics. 2024;18(10):1067-1075. doi: 10.1038/s41566-024-01493-0. Epub 2024 Jul 31.
5
Fully forward mode training for optical neural networks.全前向模式训练的光神经网络。
Nature. 2024 Aug;632(8024):280-286. doi: 10.1038/s41586-024-07687-4. Epub 2024 Aug 7.
6
Nonlinear encoding in diffractive information processing using linear optical materials.使用线性光学材料的衍射信息处理中的非线性编码
Light Sci Appl. 2024 Jul 23;13(1):173. doi: 10.1038/s41377-024-01529-8.
7
Large-scale photonic chiplet Taichi empowers 160-TOPS/W artificial general intelligence.大规模光子小芯片“太元”赋能160万亿次/瓦的通用人工智能。
Science. 2024 Apr 12;384(6692):202-209. doi: 10.1126/science.adl1203. Epub 2024 Apr 11.
8
Diffractive optical computing in free space.自由空间中的衍射光学计算。
Nat Commun. 2024 Feb 20;15(1):1525. doi: 10.1038/s41467-024-45982-w.
9
All-optical image denoising using a diffractive visual processor.使用衍射视觉处理器的全光图像去噪
Light Sci Appl. 2024 Feb 4;13(1):43. doi: 10.1038/s41377-024-01385-6.
10
Spatial-photonic Ising machine by space-division multiplexing with physically tunable coefficients of a multi-component model.通过多分量模型的物理可调系数进行空分复用的空间光子伊辛机。
Opt Express. 2023 Dec 18;31(26):44127-44138. doi: 10.1364/OE.508069.