• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

利用辅助任务初始化光子前馈神经网络。

Initializing photonic feed-forward neural networks using auxiliary tasks.

机构信息

Artificial Intelligence and Information Analysis Laboratory, Aristotle University of Thessaloniki, Greece.

Photonic Systems and Networks Research Group, Department of Informatics, Aristotle University of Thessaloniki, Greece.

出版信息

Neural Netw. 2020 Sep;129:103-108. doi: 10.1016/j.neunet.2020.05.024. Epub 2020 Jun 1.

DOI:10.1016/j.neunet.2020.05.024
PMID:32504819
Abstract

Photonics is among the most promising emerging technologies for providing fast and energy-efficient Deep Learning (DL) implementations. Despite their advantages, these photonic DL accelerators also come with certain important limitations. For example, the majority of existing photonic accelerators do not currently support many of the activation functions that are commonly used in DL, such as the ReLU activation function. Instead, sinusoidal and sigmoidal nonlinearities are usually employed, rendering the training process unstable and difficult to tune, mainly due to vanishing gradient phenomena. Thus, photonic DL models usually require carefully fine-tuning all their training hyper-parameters in order to ensure that the training process will proceed smoothly. Despite the recent advances in initialization schemes, as well as in optimization algorithms, training photonic DL models is still especially challenging. To overcome these limitations, we propose a novel adaptive initialization method that employs auxiliary tasks to estimate the optimal initialization variance for each layer of a network. The effectiveness of the proposed approach is demonstrated using two different datasets, as well as two recently proposed photonic activation functions and three different initialization methods. Apart from significantly increasing the stability of the training process, the proposed method can be directly used with any photonic activation function, without further requiring any other kind of fine-tuning, as also demonstrated through the conducted experiments.

摘要

光子学是最有前途的新兴技术之一,可用于提供快速且节能的深度学习 (DL) 实现。尽管这些光子学 DL 加速器具有优势,但它们也存在一些重要的局限性。例如,大多数现有的光子学加速器目前不支持 DL 中常用的许多激活函数,例如 ReLU 激活函数。相反,通常使用正弦和双曲非线性函数,这使得训练过程不稳定且难以调整,主要是由于梯度消失现象。因此,光子学 DL 模型通常需要仔细调整所有训练超参数,以确保训练过程能够顺利进行。尽管最近在初始化方案以及优化算法方面取得了进展,但训练光子学 DL 模型仍然极具挑战性。为了克服这些限制,我们提出了一种新颖的自适应初始化方法,该方法使用辅助任务来估计网络中每一层的最佳初始化方差。使用两个不同的数据集以及两个最近提出的光子激活函数和三种不同的初始化方法来证明所提出方法的有效性。除了显著提高训练过程的稳定性外,所提出的方法还可以直接与任何光子激活函数一起使用,而无需进一步进行任何其他类型的微调,通过进行的实验也证明了这一点。

相似文献

1
Initializing photonic feed-forward neural networks using auxiliary tasks.利用辅助任务初始化光子前馈神经网络。
Neural Netw. 2020 Sep;129:103-108. doi: 10.1016/j.neunet.2020.05.024. Epub 2020 Jun 1.
2
Quantization-aware training for low precision photonic neural networks.低精度光神经网络的量化感知训练。
Neural Netw. 2022 Nov;155:561-573. doi: 10.1016/j.neunet.2022.09.015. Epub 2022 Sep 19.
3
Improved weight initialization for deep and narrow feedforward neural network.深度窄前馈神经网络的改进权重初始化。
Neural Netw. 2024 Aug;176:106362. doi: 10.1016/j.neunet.2024.106362. Epub 2024 May 3.
4
Channel response-aware photonic neural network accelerators for high-speed inference through bandwidth-limited optics.通过带宽受限的光学实现高速推理的信道响应感知光神经网路加速器。
Opt Express. 2022 Mar 28;30(7):10664-10671. doi: 10.1364/OE.452803.
5
Successfully and efficiently training deep multi-layer perceptrons with logistic activation function simply requires initializing the weights with an appropriate negative mean.成功高效地训练具有逻辑激活函数的深层多层感知机,只需要用适当的负均值初始化权重。
Neural Netw. 2022 Sep;153:87-103. doi: 10.1016/j.neunet.2022.05.030. Epub 2022 Jun 7.
6
Two-stage neural network via sensitivity learning for 2D photonic crystal bandgap maximization.基于灵敏度学习的两阶段神经网络用于二维光子晶体带隙最大化
Appl Opt. 2022 Dec 1;61(34):10250-10259. doi: 10.1364/AO.470494.
7
Optimizing neural networks for medical data sets: A case study on neonatal apnea prediction.优化神经网络在医学数据集上的应用:以新生儿呼吸暂停预测为例的研究
Artif Intell Med. 2019 Jul;98:59-76. doi: 10.1016/j.artmed.2019.07.008. Epub 2019 Jul 25.
8
Mapping the design space of photonic topological states via deep learning.通过深度学习映射光子拓扑态的设计空间。
Opt Express. 2020 Sep 14;28(19):27893-27902. doi: 10.1364/OE.398926.
9
An effective SteinGLM initialization scheme for training multi-layer feedforward sigmoidal neural networks.一种用于训练多层前馈 Sigmoidal 神经网络的有效 SteinGLM 初始化方案。
Neural Netw. 2021 Jul;139:149-157. doi: 10.1016/j.neunet.2021.02.014. Epub 2021 Feb 27.
10
Deep learning with photonic neural cellular automata.基于光子神经细胞自动机的深度学习。
Light Sci Appl. 2024 Oct 8;13(1):283. doi: 10.1038/s41377-024-01651-7.