• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Sample Complexity Bounds for Differentially Private Learning.差分隐私学习的样本复杂度界
JMLR Workshop Conf Proc. 2011;2011:155-186.
2
Differentially Private Empirical Risk Minimization.差分隐私经验风险最小化
J Mach Learn Res. 2011 Mar;12:1069-1109.
3
Multi-Source Selection Transfer Learning with Privacy-Preserving.具有隐私保护的多源选择迁移学习
Neural Process Lett. 2022;54(6):4921-4950. doi: 10.1007/s11063-022-10841-6. Epub 2022 May 7.
4
Differentially private multivariate time series forecasting of aggregated human mobility with deep learning: Input or gradient perturbation?基于深度学习的聚合人类移动性的差分隐私多变量时间序列预测:输入扰动还是梯度扰动?
Neural Comput Appl. 2022;34(16):13355-13369. doi: 10.1007/s00521-022-07393-0. Epub 2022 Jun 3.
5
Convergence Rates for Differentially Private Statistical Estimation.差分隐私统计估计的收敛速率
Proc Int Conf Mach Learn. 2012 Jul;2012:1327-1334.
6
Privacy-preserving heterogeneous health data sharing.隐私保护的异构健康数据共享。
J Am Med Inform Assoc. 2013 May 1;20(3):462-9. doi: 10.1136/amiajnl-2012-001027. Epub 2012 Dec 13.
7
Quantum machine learning with differential privacy.带差分隐私的量子机器学习。
Sci Rep. 2023 Feb 11;13(1):2453. doi: 10.1038/s41598-022-24082-z.
8
Efficient differentially private learning improves drug sensitivity prediction.高效差分隐私学习提高药物敏感性预测。
Biol Direct. 2018 Feb 6;13(1):1. doi: 10.1186/s13062-017-0203-4.
9
A Privacy-Preserving Multi-Task Learning Framework for Face Detection, Landmark Localization, Pose Estimation, and Gender Recognition.一种用于面部检测、地标定位、姿态估计和性别识别的隐私保护多任务学习框架。
Front Neurorobot. 2020 Jan 14;13:112. doi: 10.3389/fnbot.2019.00112. eCollection 2019.
10
DPSynthesizer: Differentially Private Data Synthesizer for Privacy Preserving Data Sharing.DPSynthesizer:用于隐私保护数据共享的差分隐私数据合成器。
Proceedings VLDB Endowment. 2014 Aug;7(13):1677-1680. doi: 10.14778/2733004.2733059.

引用本文的文献

1
COINSTAC: A Privacy Enabled Model and Prototype for Leveraging and Processing Decentralized Brain Imaging Data.COINSTAC:一种用于利用和处理去中心化脑成像数据的隐私保护模型及原型。
Front Neurosci. 2016 Aug 19;10:365. doi: 10.3389/fnins.2016.00365. eCollection 2016.
2
Convergence Rates for Differentially Private Statistical Estimation.差分隐私统计估计的收敛速率
Proc Int Conf Mach Learn. 2012 Jul;2012:1327-1334.
3
iDASH: integrating data for analysis, anonymization, and sharing.iDASH:用于分析、匿名化和共享的数据集成。
J Am Med Inform Assoc. 2012 Mar-Apr;19(2):196-201. doi: 10.1136/amiajnl-2011-000538. Epub 2011 Nov 10.

本文引用的文献

1
Differentially Private Empirical Risk Minimization.差分隐私经验风险最小化
J Mach Learn Res. 2011 Mar;12:1069-1109.
2
On the sample complexity of PAC learning half-spaces against the uniform distribution.关于针对均匀分布的 PAC 学习半空间的样本复杂度。
IEEE Trans Neural Netw. 1995;6(6):1556-9. doi: 10.1109/72.471352.

差分隐私学习的样本复杂度界

Sample Complexity Bounds for Differentially Private Learning.

作者信息

Chaudhuri Kamalika, Hsu Daniel

机构信息

University of California, San Diego, 9500 Gilman Drive #0404, La Jolla, CA 92093-0404.

Microsoft Research New England, One Memorial Drive, Cambridge, MA 02142.

出版信息

JMLR Workshop Conf Proc. 2011;2011:155-186.

PMID:25285183
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC4183222/
Abstract

This work studies the problem of privacy-preserving classification - namely, learning a classifier from sensitive data while preserving the privacy of individuals in the training set. In particular, the learning algorithm is required in this problem to guarantee differential privacy, a very strong notion of privacy that has gained significant attention in recent years. A natural question to ask is: what is the sample requirement of a learning algorithm that guarantees a certain level of privacy and accuracy? We address this question in the context of learning with infinite hypothesis classes when the data is drawn from a continuous distribution. We first show that even for very simple hypothesis classes, any algorithm that uses a finite number of examples and guarantees differential privacy must fail to return an accurate classifier for at least some unlabeled data distributions. This result is unlike the case with either finite hypothesis classes or discrete data domains, in which distribution-free private learning is possible, as previously shown by Kasiviswanathan et al. (2008). We then consider two approaches to differentially private learning that get around this lower bound. The first approach is to use prior knowledge about the unlabeled data distribution in the form of a reference distribution chosen independently of the sensitive data. Given such a reference , we provide an upper bound on the sample requirement that depends (among other things) on a measure of closeness between and the unlabeled data distribution. Our upper bound applies to the non-realizable as well as the realizable case. The second approach is to relax the privacy requirement, by requiring only label-privacy - namely, that the only labels (and not the unlabeled parts of the examples) be considered sensitive information. An upper bound on the sample requirement of learning with label privacy was shown by Chaudhuri et al. (2006); in this work, we show a lower bound.

摘要

这项工作研究了隐私保护分类问题——即在保护训练集中个体隐私的同时,从敏感数据中学习分类器。具体而言,该问题要求学习算法保证差分隐私,这是近年来备受关注的一种非常强的隐私概念。一个自然要问的问题是:保证一定隐私水平和准确性的学习算法的样本需求是什么?我们在数据从连续分布中抽取且假设类无穷的学习背景下解决这个问题。我们首先表明,即使对于非常简单的假设类,任何使用有限数量示例并保证差分隐私的算法,对于至少一些未标记数据分布,必定无法返回准确的分类器。这个结果与有限假设类或离散数据域的情况不同,在有限假设类或离散数据域中,如Kasiviswanathan等人(2008年)先前所示,无分布的隐私学习是可能的。然后,我们考虑两种规避此下限的差分隐私学习方法。第一种方法是使用关于未标记数据分布的先验知识,形式为独立于敏感数据选择的参考分布。给定这样一个参考,我们给出样本需求的上限,该上限(除其他因素外)取决于参考分布与未标记数据分布之间的接近程度度量。我们的上限适用于不可实现以及可实现的情况。第二种方法是通过仅要求标签隐私来放宽隐私要求——即,仅将标签(而不是示例的未标记部分)视为敏感信息。Chaudhuri等人(2006年)给出了具有标签隐私的学习的样本需求上限;在这项工作中,我们给出了下限。