• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

批量大小:要么做大,要么回家?小批量大小的医学自动编码器有违反直觉的改进。

Batch size: go big or go home? Counterintuitive improvement in medical autoencoders with smaller batch size.

作者信息

Kerley Cailey I, Cai Leon Y, Tang Yucheng, Beason-Held Lori L, Resnick Susan M, Cutting Laurie E, Landman Bennett A

机构信息

Department of Electrical and Computer Engineering, Vanderbilt University, Nashville, TN, USA.

Department of Biomedical Engineering, Vanderbilt University, Nashville, TN, USA.

出版信息

Proc SPIE Int Soc Opt Eng. 2023 Feb;12464. doi: 10.1117/12.2653643. Epub 2023 Apr 3.

DOI:10.1117/12.2653643
PMID:37465095
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10353832/
Abstract

Batch size is a key hyperparameter in training deep learning models. Conventional wisdom suggests larger batches produce improved model performance. Here we present evidence to the contrary, particularly when using autoencoders to derive meaningful latent spaces from data with spatially global similarities and local differences, such as electronic health records (EHR) and medical imaging. We investigate batch size effects in both EHR data from the Baltimore Longitudinal Study of Aging and medical imaging data from the multimodal brain tumor segmentation (BraTS) challenge. We train fully connected and convolutional autoencoders to compress the EHR and imaging input spaces, respectively, into 32-dimensional latent spaces via reconstruction losses for various batch sizes between 1 and 100. Under the same hyperparameter configurations, smaller batches improve loss performance for both datasets. Additionally, latent spaces derived by autoencoders with smaller batches capture more biologically meaningful information. Qualitatively, we visualize 2-dimensional projections of the latent spaces and find that with smaller batches the EHR network better separates the sex of the individuals, and the imaging network better captures the right-left laterality of tumors. Quantitatively, the analogous sex classification and laterality regressions using the latent spaces demonstrate statistically significant improvements in performance at smaller batch sizes. Finally, we find improved individual variation locally in visualizations of representative data reconstructions at lower batch sizes. Taken together, these results suggest that smaller batch sizes should be considered when designing autoencoders to extract meaningful latent spaces among EHR and medical imaging data driven by global similarities and local variation.

摘要

批量大小是深度学习模型训练中的一个关键超参数。传统观点认为,更大的批量会提升模型性能。然而,我们在此展示了相反的证据,特别是在使用自动编码器从具有空间全局相似性和局部差异的数据(如电子健康记录(EHR)和医学影像)中提取有意义的潜在空间时。我们研究了来自巴尔的摩纵向衰老研究的EHR数据和多模态脑肿瘤分割(BraTS)挑战中的医学影像数据的批量大小效应。我们训练全连接和卷积自动编码器,通过针对1到100之间的各种批量大小的重建损失,分别将EHR和影像输入空间压缩到32维潜在空间。在相同的超参数配置下,较小的批量能提高两个数据集的损失性能。此外,由较小批量的自动编码器导出的潜在空间捕获了更多具有生物学意义的信息。定性地,我们可视化了潜在空间的二维投影,发现使用较小批量时,EHR网络能更好地分离个体的性别,而影像网络能更好地捕获肿瘤的左右侧性。定量地,使用潜在空间进行的类似性别分类和侧性回归表明,在较小批量大小时性能有统计学上的显著提升。最后,我们发现在较低批量大小下,代表性数据重建的可视化中局部个体差异得到了改善。综上所述,这些结果表明,在设计自动编码器以在由全局相似性和局部变化驱动的EHR和医学影像数据中提取有意义的潜在空间时,应考虑较小的批量大小。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/f92eeaeef4d0/nihms-1858227-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/c0321e3eea68/nihms-1858227-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/a4d7005b4400/nihms-1858227-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/885cf05e238e/nihms-1858227-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/c4f83a1c86b1/nihms-1858227-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/f92eeaeef4d0/nihms-1858227-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/c0321e3eea68/nihms-1858227-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/a4d7005b4400/nihms-1858227-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/885cf05e238e/nihms-1858227-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/c4f83a1c86b1/nihms-1858227-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bba1/10353832/f92eeaeef4d0/nihms-1858227-f0005.jpg

相似文献

1
Batch size: go big or go home? Counterintuitive improvement in medical autoencoders with smaller batch size.批量大小:要么做大,要么回家?小批量大小的医学自动编码器有违反直觉的改进。
Proc SPIE Int Soc Opt Eng. 2023 Feb;12464. doi: 10.1117/12.2653643. Epub 2023 Apr 3.
2
Analysis of Autoencoders for Network Intrusion Detection.自动编码器在网络入侵检测中的分析。
Sensors (Basel). 2021 Jun 23;21(13):4294. doi: 10.3390/s21134294.
3
Small molecule autoencoders: architecture engineering to optimize latent space utility and sustainability.小分子自动编码器:用于优化潜在空间效用和可持续性的架构工程
J Cheminform. 2024 Mar 5;16(1):26. doi: 10.1186/s13321-024-00817-0.
4
Locally Embedding Autoencoders: A Semi-Supervised Manifold Learning Approach of Document Representation.局部嵌入自动编码器:一种文档表示的半监督流形学习方法。
PLoS One. 2016 Jan 19;11(1):e0146672. doi: 10.1371/journal.pone.0146672. eCollection 2016.
5
A dual autoencoder and singular value decomposition based feature optimization for the segmentation of brain tumor from MRI images.基于双自动编码器和奇异值分解的特征优化在 MRI 图像脑部肿瘤分割中的应用。
BMC Med Imaging. 2021 May 13;21(1):82. doi: 10.1186/s12880-021-00614-3.
6
Sequential autoencoders for feature engineering and pretraining in major depressive disorder risk prediction.用于重度抑郁症风险预测中特征工程和预训练的序列自动编码器
JAMIA Open. 2023 Oct 9;6(4):ooad086. doi: 10.1093/jamiaopen/ooad086. eCollection 2023 Dec.
7
Benchmarking variational AutoEncoders on cancer transcriptomics data.基于癌症转录组学数据的变分自编码器基准测试。
PLoS One. 2023 Oct 5;18(10):e0292126. doi: 10.1371/journal.pone.0292126. eCollection 2023.
8
Parameter tuning is a key part of dimensionality reduction via deep variational autoencoders for single cell RNA transcriptomics.参数调整是通过深度变分自编码器进行单细胞RNA转录组学降维的关键部分。
Pac Symp Biocomput. 2019;24:362-373.
9
Latent space unsupervised semantic segmentation.潜在空间无监督语义分割
Front Physiol. 2023 Apr 25;14:1151312. doi: 10.3389/fphys.2023.1151312. eCollection 2023.
10
Compressing gene expression data using multiple latent space dimensionalities learns complementary biological representations.使用多个潜在空间维度压缩基因表达数据可学习互补的生物学表现形式。
Genome Biol. 2020 May 11;21(1):109. doi: 10.1186/s13059-020-02021-3.

引用本文的文献

1
Enhancing multiclass COVID-19 prediction with ESN-MDFS: Extreme smart network using mean dropout feature selection technique.利用 ESN-MDFS 提高多类别 COVID-19 预测的准确率:使用均值随机失活特征选择技术的极端智能网络。
PLoS One. 2024 Nov 12;19(11):e0310011. doi: 10.1371/journal.pone.0310011. eCollection 2024.

本文引用的文献

1
pyPheWAS: A Phenome-Disease Association Tool for Electronic Medical Record Analysis.pyPheWAS:用于电子病历分析的表型-疾病关联工具。
Neuroinformatics. 2022 Apr;20(2):483-505. doi: 10.1007/s12021-021-09553-4. Epub 2022 Jan 3.
2
The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS).多模态脑肿瘤图像分割基准(BRATS)。
IEEE Trans Med Imaging. 2015 Oct;34(10):1993-2024. doi: 10.1109/TMI.2014.2377694. Epub 2014 Dec 4.
3
Systematic comparison of phenome-wide association study of electronic medical record data and genome-wide association study data.
系统比较电子病历数据的表型全基因组关联研究和全基因组关联研究数据。
Nat Biotechnol. 2013 Dec;31(12):1102-10. doi: 10.1038/nbt.2749.
4
Symmetric atlasing and model based segmentation: an application to the hippocampus in older adults.对称图谱与基于模型的分割:在老年人海马体中的应用
Med Image Comput Comput Assist Interv. 2006;9(Pt 2):58-66. doi: 10.1007/11866763_8.
5
The general inefficiency of batch training for gradient descent learning.梯度下降学习中批量训练的总体低效性。
Neural Netw. 2003 Dec;16(10):1429-51. doi: 10.1016/S0893-6080(03)00138-2.
6
One-year age changes in MRI brain volumes in older adults.老年人脑磁共振成像(MRI)体积的一年期年龄变化
Cereb Cortex. 2000 May;10(5):464-72. doi: 10.1093/cercor/10.5.464.