Suppr超能文献

WI-TMLEGA:基于熵增益和学习率调整的权重初始化与训练方法

WI-TMLEGA: Weight Initialization and Training Method Based on Entropy Gain and Learning Rate Adjustment.

作者信息

Tang Hongchuan, Li Zhongguo, Wang Qi, Fan Wenbin

机构信息

School of Mechanical Engineering, Jiangsu University of Science and Technology, Zhenjiang 212100, China.

School of Automotive Engineering, Nantong Institute of Technology, Nantong 226001, China.

出版信息

Entropy (Basel). 2024 Jul 23;26(8):614. doi: 10.3390/e26080614.

Abstract

Addressing the issues of prolonged training times and low recognition rates in large model applications, this paper proposes a weight training method based on entropy gain for weight initialization and dynamic adjustment of the learning rate using the multilayer perceptron (MLP) model as an example. Initially, entropy gain was used to replace random initial values for weight initialization. Subsequently, an incremental learning rate strategy was employed for weight updates. The model was trained and validated using the MNIST handwritten digit dataset. The experimental results showed that, compared to random initialization, the proposed initialization method improves training effectiveness by 39.8% and increases the maximum recognition accuracy by 8.9%, demonstrating the feasibility of this method in large model applications.

摘要

针对大模型应用中训练时间长和识别率低的问题,本文以多层感知器(MLP)模型为例,提出一种基于熵增益的权重训练方法,用于权重初始化和学习率的动态调整。首先,使用熵增益代替随机初始值进行权重初始化。随后,采用增量学习率策略进行权重更新。使用MNIST手写数字数据集对模型进行训练和验证。实验结果表明,与随机初始化相比,所提出的初始化方法将训练效率提高了39.8%,最大识别准确率提高了8.9%,证明了该方法在大模型应用中的可行性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ce7f/11353430/19e66eaad46d/entropy-26-00614-g001.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验