• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

我们在测试数据上进行训练吗?清除CIFAR中的近似重复数据。

Do We Train on Test Data? Purging CIFAR of Near-Duplicates.

作者信息

Barz Björn, Denzler Joachim

机构信息

Computer Vision Group, Friedrich Schiller University Jena, Ernst-Abbe-Platz 2, 07743 Jena, Germany.

出版信息

J Imaging. 2020 Jun 2;6(6):41. doi: 10.3390/jimaging6060041.

DOI:10.3390/jimaging6060041
PMID:34460587
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8321059/
Abstract

The CIFAR-10 and CIFAR-100 datasets are two of the most heavily benchmarked datasets in computer vision and are often used to evaluate novel methods and model architectures in the field of deep learning. However, we find that 3.3% and 10% of the images from the test sets of these datasets have duplicates in the training set. These duplicates are easily recognizable by memorization and may, hence, bias the comparison of image recognition techniques regarding their generalization capability. To eliminate this bias, we provide the "fair CIFAR" (ciFAIR) dataset, where we replaced all duplicates in the test sets with new images sampled from the same domain. The training set remains unchanged, in order not to invalidate pre-trained models. We then re-evaluate the classification performance of various popular state-of-the-art CNN architectures on these new test sets to investigate whether recent research has overfitted to memorizing data instead of learning abstract concepts. We find a significant drop in classification accuracy of between 9% and 14% relative to the original performance on the duplicate-free test set. We make both the ciFAIR dataset and pre-trained models publicly available and furthermore maintain a leaderboard for tracking the state of the art.

摘要

CIFAR-10和CIFAR-100数据集是计算机视觉领域中基准测试使用最为频繁的两个数据集,常用于评估深度学习领域的新方法和模型架构。然而,我们发现这些数据集测试集中分别有3.3%和10%的图像在训练集中存在重复。这些重复图像很容易通过记忆识别出来,因此可能会在图像识别技术泛化能力的比较中产生偏差。为了消除这种偏差,我们提供了“公平CIFAR”(ciFAIR)数据集,我们用从同一域中采样的新图像替换了测试集中所有的重复图像。训练集保持不变,以免使预训练模型无效。然后,我们在这些新的测试集上重新评估各种流行的先进卷积神经网络架构的分类性能,以研究近期的研究是否过度拟合于记忆数据而非学习抽象概念。我们发现,相对于在无重复测试集上的原始性能,分类准确率显著下降了9%至14%。我们将ciFAIR数据集和预训练模型都公开提供,并且还维护了一个排行榜来跟踪当前的技术水平。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f76d/8321059/587951b21531/jimaging-06-00041-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f76d/8321059/a3f6cee4b547/jimaging-06-00041-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f76d/8321059/a363f052ee56/jimaging-06-00041-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f76d/8321059/a80b29cd8276/jimaging-06-00041-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f76d/8321059/587951b21531/jimaging-06-00041-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f76d/8321059/a3f6cee4b547/jimaging-06-00041-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f76d/8321059/a363f052ee56/jimaging-06-00041-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f76d/8321059/a80b29cd8276/jimaging-06-00041-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f76d/8321059/587951b21531/jimaging-06-00041-g004.jpg

相似文献

1
Do We Train on Test Data? Purging CIFAR of Near-Duplicates.我们在测试数据上进行训练吗?清除CIFAR中的近似重复数据。
J Imaging. 2020 Jun 2;6(6):41. doi: 10.3390/jimaging6060041.
2
Promoting the Shift From Pixel-Level Correlations to Object Semantics Learning by Rethinking Computer Vision Benchmark Data Sets.通过重新思考计算机视觉基准数据集来推动从像素级相关性到对象语义学习的转变。
Neural Comput. 2024 Jul 19;36(8):1626-1642. doi: 10.1162/neco_a_01677.
3
Musculoskeletal Images Classification for Detection of Fractures Using Transfer Learning.使用迁移学习进行骨折检测的肌肉骨骼图像分类
J Imaging. 2020 Nov 23;6(11):127. doi: 10.3390/jimaging6110127.
4
A novel end-to-end classifier using domain transferred deep convolutional neural networks for biomedical images.一种使用域转移深度卷积神经网络的新型端到端生物医学图像分类器。
Comput Methods Programs Biomed. 2017 Mar;140:283-293. doi: 10.1016/j.cmpb.2016.12.019. Epub 2017 Jan 6.
5
Targeted transfer learning to improve performance in small medical physics datasets.靶向迁移学习以提高小型医学物理数据集的性能。
Med Phys. 2020 Dec;47(12):6246-6256. doi: 10.1002/mp.14507. Epub 2020 Oct 25.
6
Evolution of Deep Convolutional Neural Networks Using Cartesian Genetic Programming.基于笛卡尔遗传编程的深度卷积神经网络进化。
Evol Comput. 2020 Spring;28(1):141-163. doi: 10.1162/evco_a_00253. Epub 2019 Mar 22.
7
Plant Disease Classification: A Comparative Evaluation of Convolutional Neural Networks and Deep Learning Optimizers.植物病害分类:卷积神经网络与深度学习优化器的比较评估
Plants (Basel). 2020 Oct 6;9(10):1319. doi: 10.3390/plants9101319.
8
Discriminative Unsupervised Feature Learning with Exemplar Convolutional Neural Networks.基于示例卷积神经网络的判别式无监督特征学习。
IEEE Trans Pattern Anal Mach Intell. 2016 Sep;38(9):1734-47. doi: 10.1109/TPAMI.2015.2496141. Epub 2015 Oct 29.
9
Deep neural networks for human microRNA precursor detection.用于人类 microRNA 前体检测的深度神经网络。
BMC Bioinformatics. 2020 Jan 13;21(1):17. doi: 10.1186/s12859-020-3339-7.
10
Diagnosing acute promyelocytic leukemia by using convolutional neural network.使用卷积神经网络诊断急性早幼粒细胞白血病。
Clin Chim Acta. 2021 Jan;512:1-6. doi: 10.1016/j.cca.2020.10.039. Epub 2020 Nov 4.

引用本文的文献

1
GUANinE v1.0: Benchmark Datasets for Genomic AI Sequence-to-Function Models.鸟嘌呤v1.0:用于基因组人工智能序列到功能模型的基准数据集。
bioRxiv. 2024 Mar 7:2023.10.12.562113. doi: 10.1101/2023.10.12.562113.

本文引用的文献

1
Improving Large-Scale Image Retrieval Through Robust Aggregation of Local Descriptors.通过稳健的局部描述符聚合来改进大规模图像检索。
IEEE Trans Pattern Anal Mach Intell. 2017 Sep;39(9):1783-1796. doi: 10.1109/TPAMI.2016.2613873. Epub 2016 Sep 27.
2
80 million tiny images: a large data set for nonparametric object and scene recognition.八千万张小图片:用于非参数化物体与场景识别的大型数据集。
IEEE Trans Pattern Anal Mach Intell. 2008 Nov;30(11):1958-70. doi: 10.1109/TPAMI.2008.128.