Suppr超能文献

基于隐私保护的人像抠图再思考。

Rethinking Portrait Matting with Privacy Preserving.

作者信息

Ma Sihan, Li Jizhizi, Zhang Jing, Zhang He, Tao Dacheng

机构信息

The University of Sydney, Sydney, NSW Australia.

Adobe Inc., San Jose, CA USA.

出版信息

Int J Comput Vis. 2023 May 20:1-26. doi: 10.1007/s11263-023-01797-8.

Abstract

Recently, there has been an increasing concern about the privacy issue raised by identifiable information in machine learning. However, previous portrait matting methods were all based on identifiable images. To fill the gap, we present P3M-10k, which is the first large-scale anonymized benchmark for Privacy-Preserving Portrait Matting (P3M). P3M-10k consists of 10,421 high resolution face-blurred portrait images along with high-quality alpha mattes, which enables us to systematically evaluate both trimap-free and trimap-based matting methods and obtain some useful findings about model generalization ability under the privacy preserving training (PPT) setting. We also present a unified matting model dubbed P3M-Net that is compatible with both CNN and transformer backbones. To further mitigate the cross-domain performance gap issue under the PPT setting, we devise a simple yet effective Copy and Paste strategy (P3M-CP), which borrows facial information from public celebrity images and directs the network to reacquire the face context at both data and feature level. Extensive experiments on P3M-10k and public benchmarks demonstrate the superiority of P3M-Net over state-of-the-art methods and the effectiveness of P3M-CP in improving the cross-domain generalization ability, implying a great significance of P3M for future research and real-world applications. The dataset, code and models are available here (https://github.com/ViTAE-Transformer/P3M-Net).

摘要

最近,机器学习中可识别信息引发的隐私问题受到了越来越多的关注。然而,以往的人像抠图方法都是基于可识别图像的。为了填补这一空白,我们提出了P3M-10k,这是第一个用于隐私保护人像抠图(P3M)的大规模匿名基准数据集。P3M-10k由10421张高分辨率面部模糊的人像图像以及高质量的alpha遮罩组成,这使我们能够系统地评估无trimap和基于trimap的抠图方法,并在隐私保护训练(PPT)设置下获得关于模型泛化能力的一些有用发现。我们还提出了一个统一的抠图模型P3M-Net,它与CNN和Transformer主干都兼容。为了进一步缓解PPT设置下的跨域性能差距问题,我们设计了一种简单而有效的复制粘贴策略(P3M-CP),该策略从公众名人图像中借用面部信息,并指导网络在数据和特征层面重新获取面部上下文。在P3M-10k和公共基准上进行的大量实验证明了P3M-Net优于现有方法,以及P3M-CP在提高跨域泛化能力方面的有效性,这意味着P3M对未来研究和实际应用具有重要意义。数据集、代码和模型可在此处获取(https://github.com/ViTAE-Transformer/P3M-Net)。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d666/10199740/eb16e330a444/11263_2023_1797_Fig1_HTML.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验