Suppr超能文献

解锁L1正则化的力量:一种驯服卷积神经网络(CNN)图像分类中过拟合问题的新方法。

Unlocking the power of L1 regularization: A novel approach to taming overfitting in CNN for image classification.

作者信息

Sheikh Ramla, Wahid Fazli, Ali Sikandar, Alkhayyat Ahmed, Ma Yingling, Khan Jawad, Lee Youngmoon

机构信息

Department of Information Technology, The University of Haripur, Haripur, Pakistan.

Collage of Science and Engineering, School of Computing, University of Derby, Derby, United Kingdom.

出版信息

PLoS One. 2025 Sep 5;20(9):e0327985. doi: 10.1371/journal.pone.0327985. eCollection 2025.

Abstract

Convolutional Neural Networks (CNNs) stand as indispensable tools in deep learning, capable of autonomously extracting crucial features from diverse data types. However, the intricacies of CNN architectures can present challenges such as overfitting and underfitting, necessitating thoughtful strategies to optimize their performance. In this work, these issues have been resolved by introducing L1 regularization in the basic architecture of CNN when it is applied for image classification. The proposed model has been applied to three different datasets. It has been observed that incorporating L1 regularization with different coefficient values has distinct effects on the working mechanism of CNN architecture resulting in improving its performance. In MNIST digit classification, L1 regularization (coefficient: 0.01) simplifies feature representation and prevents overfitting, leading to enhanced accuracy. In the Mango Tree Leaves dataset, dual L1 regularization (coefficient: 0.001 for convolutional and 0.01 for dense layers) improves model interpretability and generalization, facilitating effective leaf classification. Additionally, for hand-drawn sketches like those in the Quick, Draw! Dataset, L1 regularization (coefficient: 0.001) refines feature representation, resulting in improved recognition accuracy and generalization across diverse sketch categories. These findings underscore the significance of regularization techniques like L1 regularization in fine-tuning CNNs, optimizing their performance, and ensuring their adaptability to new data while maintaining high accuracy. Such strategies play a pivotal role in advancing the utility of CNNs across various domains, further solidifying their position as a cornerstone of deep learning.

摘要

卷积神经网络(CNNs)是深度学习中不可或缺的工具,能够从各种数据类型中自动提取关键特征。然而,CNN架构的复杂性可能带来诸如过拟合和欠拟合等挑战,因此需要精心设计策略来优化其性能。在这项工作中,通过在将CNN应用于图像分类时在其基本架构中引入L1正则化,这些问题得到了解决。所提出的模型已应用于三个不同的数据集。据观察,结合不同系数值的L1正则化对CNN架构的工作机制有不同影响,从而提高其性能。在MNIST数字分类中,L1正则化(系数:0.01)简化了特征表示并防止过拟合,从而提高了准确性。在芒果树叶数据集里,双重L1正则化(卷积层系数:0.001,全连接层系数:0.01)提高了模型的可解释性和泛化能力,便于进行有效的树叶分类。此外,对于像《快画!》数据集中的手绘草图,L1正则化(系数:0.001)优化了特征表示,提高了识别准确率并在不同草图类别间实现了更好的泛化。这些发现强调了L1正则化等正则化技术在微调CNN、优化其性能以及确保其在保持高精度的同时适应新数据方面的重要性。此类策略在提升CNN在各个领域的效用方面发挥着关键作用,进一步巩固了它们作为深度学习基石的地位。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ef66/12413007/b83e0a850907/pone.0327985.g001.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验