Suppr超能文献

相似文献

1
The inverse variance-flatness relation in stochastic gradient descent is critical for finding flat minima.
Proc Natl Acad Sci U S A. 2021 Mar 2;118(9). doi: 10.1073/pnas.2015617118.
2
Stochastic Gradient Descent Introduces an Effective Landscape-Dependent Regularization Favoring Flat Solutions.
Phys Rev Lett. 2023 Jun 9;130(23):237101. doi: 10.1103/PhysRevLett.130.237101.
3
Anomalous diffusion dynamics of learning in deep neural networks.
Neural Netw. 2022 May;149:18-28. doi: 10.1016/j.neunet.2022.01.019. Epub 2022 Feb 3.
4
A mean field view of the landscape of two-layer neural networks.
Proc Natl Acad Sci U S A. 2018 Aug 14;115(33):E7665-E7671. doi: 10.1073/pnas.1806579115. Epub 2018 Jul 27.
5
Shaping the learning landscape in neural networks around wide flat minima.
Proc Natl Acad Sci U S A. 2020 Jan 7;117(1):161-170. doi: 10.1073/pnas.1908636117. Epub 2019 Dec 23.
6
Accelerating Minibatch Stochastic Gradient Descent Using Typicality Sampling.
IEEE Trans Neural Netw Learn Syst. 2020 Nov;31(11):4649-4659. doi: 10.1109/TNNLS.2019.2957003. Epub 2020 Oct 29.
7
Understanding Short-Range Memory Effects in Deep Neural Networks.
IEEE Trans Neural Netw Learn Syst. 2024 Aug;35(8):10576-10590. doi: 10.1109/TNNLS.2023.3242969. Epub 2024 Aug 5.
8
Unveiling the Structure of Wide Flat Minima in Neural Networks.
Phys Rev Lett. 2021 Dec 31;127(27):278301. doi: 10.1103/PhysRevLett.127.278301.
9
Towards Better Generalization of Deep Neural Networks via Non-Typicality Sampling Scheme.
IEEE Trans Neural Netw Learn Syst. 2023 Oct;34(10):7910-7920. doi: 10.1109/TNNLS.2022.3147031. Epub 2023 Oct 5.
10
The Limiting Dynamics of SGD: Modified Loss, Phase-Space Oscillations, and Anomalous Diffusion.
Neural Comput. 2023 Dec 12;36(1):151-174. doi: 10.1162/neco_a_01626.

引用本文的文献

2
Temporal Contrastive Learning through implicit non-equilibrium memory.
Nat Commun. 2025 Mar 4;16(1):2163. doi: 10.1038/s41467-025-57043-x.
3
Machine learning meets physics: A two-way street.
Proc Natl Acad Sci U S A. 2024 Jul 2;121(27):e2403580121. doi: 10.1073/pnas.2403580121. Epub 2024 Jun 24.
5
Brain-inspired chaotic spiking backpropagation.
Natl Sci Rev. 2024 Jan 30;11(6):nwae037. doi: 10.1093/nsr/nwae037. eCollection 2024 Jun.
6
The training process of many deep networks explores the same low-dimensional manifold.
Proc Natl Acad Sci U S A. 2024 Mar 19;121(12):e2310002121. doi: 10.1073/pnas.2310002121. Epub 2024 Mar 12.
7
On the different regimes of stochastic gradient descent.
Proc Natl Acad Sci U S A. 2024 Feb 27;121(9):e2316301121. doi: 10.1073/pnas.2316301121. Epub 2024 Feb 20.
8
Thermodynamics of the Ising Model Encoded in Restricted Boltzmann Machines.
Entropy (Basel). 2022 Nov 22;24(12):1701. doi: 10.3390/e24121701.
9
Topology, vorticity, and limit cycle in a stabilized Kuramoto-Sivashinsky equation.
Proc Natl Acad Sci U S A. 2022 Dec 6;119(49):e2211359119. doi: 10.1073/pnas.2211359119. Epub 2022 Dec 2.
10
Let the robotic games begin.
Proc Natl Acad Sci U S A. 2022 Apr 26;119(17):e2204152119. doi: 10.1073/pnas.2204152119. Epub 2022 Apr 19.

本文引用的文献

1
High-dimensional dynamics of generalization error in neural networks.
Neural Netw. 2020 Dec;132:428-446. doi: 10.1016/j.neunet.2020.08.022. Epub 2020 Sep 5.
2
Shaping the learning landscape in neural networks around wide flat minima.
Proc Natl Acad Sci U S A. 2020 Jan 7;117(1):161-170. doi: 10.1073/pnas.1908636117. Epub 2019 Dec 23.
3
Reconciling modern machine-learning practice and the classical bias-variance trade-off.
Proc Natl Acad Sci U S A. 2019 Aug 6;116(32):15849-15854. doi: 10.1073/pnas.1903070116. Epub 2019 Jul 24.
4
A mean field view of the landscape of two-layer neural networks.
Proc Natl Acad Sci U S A. 2018 Aug 14;115(33):E7665-E7671. doi: 10.1073/pnas.1806579115. Epub 2018 Jul 27.
5
Overcoming catastrophic forgetting in neural networks.
Proc Natl Acad Sci U S A. 2017 Mar 28;114(13):3521-3526. doi: 10.1073/pnas.1611835114. Epub 2017 Mar 14.
6
Unreasonable effectiveness of learning neural networks: From accessible states and robust ensembles to basic algorithmic schemes.
Proc Natl Acad Sci U S A. 2016 Nov 29;113(48):E7655-E7662. doi: 10.1073/pnas.1608103113. Epub 2016 Nov 15.
7
Deep learning.
Nature. 2015 May 28;521(7553):436-44. doi: 10.1038/nature14539.
8
Optimization by simulated annealing.
Science. 1983 May 13;220(4598):671-80. doi: 10.1126/science.220.4598.671.
9
Structure of stochastic dynamics near fixed points.
Proc Natl Acad Sci U S A. 2005 Sep 13;102(37):13029-33. doi: 10.1073/pnas.0506347102. Epub 2005 Sep 1.
10
Flat minima.
Neural Comput. 1997 Jan 1;9(1):1-42. doi: 10.1162/neco.1997.9.1.1.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验