• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

PME:用于推荐系统的基于剪枝的多尺寸嵌入

PME: pruning-based multi-size embedding for recommender systems.

作者信息

Liu Zirui, Song Qingquan, Li Li, Choi Soo-Hyun, Chen Rui, Hu Xia

机构信息

Computer Science Department, Rice University, Houston, TX, United States.

Linkedin, Sunnyvale, CA, United States.

出版信息

Front Big Data. 2023 Jun 15;6:1195742. doi: 10.3389/fdata.2023.1195742. eCollection 2023.

DOI:10.3389/fdata.2023.1195742
PMID:37397622
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10311001/
Abstract

Embedding is widely used in recommendation models to learn feature representations. However, the traditional embedding technique that assigns a fixed size to all categorical features may be suboptimal due to the following reasons. In recommendation domain, the majority of categorical features' embeddings can be trained with less capacity without impacting model performance, thereby storing embeddings with equal length may incur unnecessary memory usage. Existing work that tries to allocate customized sizes for each feature usually either simply scales the embedding size with feature's popularity or formulates this size allocation problem as an architecture selection problem. Unfortunately, most of these methods either have large performance drop or incur significant extra time cost for searching proper embedding sizes. In this article, instead of formulating the size allocation problem as an architecture selection problem, we approach the problem from a pruning perspective and propose runing-based ulti-size mbedding (PME) framework. During the search phase, we prune the dimensions that have the least impact on model performance in the embedding to reduce its capacity. Then, we show that the customized size of each token can be obtained by transferring the capacity of its pruned embedding with significant less search cost. Experimental results validate that PME can efficiently find proper sizes and hence achieve strong performance while significantly reducing the number of parameters in the embedding layer.

摘要

嵌入在推荐模型中被广泛用于学习特征表示。然而,由于以下原因,为所有分类特征分配固定大小的传统嵌入技术可能不是最优的。在推荐领域,大多数分类特征的嵌入可以在不影响模型性能的情况下用较小的容量进行训练,因此存储等长的嵌入可能会产生不必要的内存使用。现有的尝试为每个特征分配定制大小的工作通常要么简单地根据特征的流行程度缩放嵌入大小,要么将这个大小分配问题表述为一个架构选择问题。不幸的是,这些方法大多要么有较大的性能下降,要么在搜索合适的嵌入大小时会产生显著的额外时间成本。在本文中,我们不是将大小分配问题表述为一个架构选择问题,而是从剪枝的角度来处理这个问题,并提出了基于运行的多大小嵌入(PME)框架。在搜索阶段,我们在嵌入中修剪对模型性能影响最小的维度以降低其容量。然后,我们表明每个令牌的定制大小可以通过转移其修剪后的嵌入的容量来获得,且搜索成本显著降低。实验结果验证了PME可以有效地找到合适的大小,从而在显著减少嵌入层参数数量的同时实现强大的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/7c299980b894/fdata-06-1195742-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/8f94a15f497f/fdata-06-1195742-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/1406c59af35a/fdata-06-1195742-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/b306af69cbb4/fdata-06-1195742-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/8ea5b10ddac3/fdata-06-1195742-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/11a9d88b56f7/fdata-06-1195742-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/5ad22fdbe7d1/fdata-06-1195742-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/28c95377c3e0/fdata-06-1195742-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/b54fd5d11423/fdata-06-1195742-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/e01d4e6bde73/fdata-06-1195742-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/6f5ba067b558/fdata-06-1195742-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/7c299980b894/fdata-06-1195742-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/8f94a15f497f/fdata-06-1195742-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/1406c59af35a/fdata-06-1195742-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/b306af69cbb4/fdata-06-1195742-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/8ea5b10ddac3/fdata-06-1195742-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/11a9d88b56f7/fdata-06-1195742-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/5ad22fdbe7d1/fdata-06-1195742-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/28c95377c3e0/fdata-06-1195742-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/b54fd5d11423/fdata-06-1195742-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/e01d4e6bde73/fdata-06-1195742-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/6f5ba067b558/fdata-06-1195742-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fa/10311001/7c299980b894/fdata-06-1195742-g0011.jpg

相似文献

1
PME: pruning-based multi-size embedding for recommender systems.PME:用于推荐系统的基于剪枝的多尺寸嵌入
Front Big Data. 2023 Jun 15;6:1195742. doi: 10.3389/fdata.2023.1195742. eCollection 2023.
2
Where to Prune: Using LSTM to Guide Data-Dependent Soft Pruning.修剪位置:使用长短期记忆网络指导数据依赖的软修剪
IEEE Trans Image Process. 2021;30:293-304. doi: 10.1109/TIP.2020.3035028. Epub 2020 Nov 24.
3
ARPruning: An automatic channel pruning based on attention map ranking.ARPruning:一种基于注意力图排序的自动通道剪枝方法。
Neural Netw. 2024 Jun;174:106220. doi: 10.1016/j.neunet.2024.106220. Epub 2024 Feb 29.
4
An Approach to Semantic-Aware Heterogeneous Network Embedding for Recommender Systems.一种用于推荐系统的语义感知异构网络嵌入方法。
IEEE Trans Cybern. 2023 Sep;53(9):6027-6040. doi: 10.1109/TCYB.2022.3233819. Epub 2023 Aug 17.
5
Subarchitecture Ensemble Pruning in Neural Architecture Search.神经架构搜索中的子架构集成剪枝
IEEE Trans Neural Netw Learn Syst. 2022 Dec;33(12):7928-7936. doi: 10.1109/TNNLS.2021.3085299. Epub 2022 Nov 30.
6
A class-incremental learning approach for learning feature-compatible embeddings.一种用于学习特征兼容嵌入的类增量学习方法。
Neural Netw. 2024 Dec;180:106685. doi: 10.1016/j.neunet.2024.106685. Epub 2024 Aug 31.
7
Dynamical Conventional Neural Network Channel Pruning by Genetic Wavelet Channel Search for Image Classification.基于遗传小波通道搜索的动态传统神经网络通道剪枝用于图像分类
Front Comput Neurosci. 2021 Oct 27;15:760554. doi: 10.3389/fncom.2021.760554. eCollection 2021.
8
DMPP: Differentiable multi-pruner and predictor for neural network pruning.DMPP:用于神经网络剪枝的可微分多修剪器和预测器。
Neural Netw. 2022 Mar;147:103-112. doi: 10.1016/j.neunet.2021.12.020. Epub 2021 Dec 30.
9
CATRO: Channel Pruning via Class-Aware Trace Ratio Optimization.CATRO:通过类别感知迹比优化进行通道剪枝
IEEE Trans Neural Netw Learn Syst. 2024 Aug;35(8):11595-11607. doi: 10.1109/TNNLS.2023.3262952. Epub 2024 Aug 5.
10
EvoPruneDeepTL: An evolutionary pruning model for transfer learning based deep neural networks.EvoPruneDeepTL:一种用于基于迁移学习的深度神经网络的进化剪枝模型。
Neural Netw. 2023 Jan;158:59-82. doi: 10.1016/j.neunet.2022.10.011. Epub 2022 Nov 4.

引用本文的文献

1
Design and Implementation of an LSTM Model with Embeddings on MCUs for Prediction of Meteorological Variables.基于微控制器的带嵌入层的长短期记忆网络模型用于气象变量预测的设计与实现
Sensors (Basel). 2025 Jun 7;25(12):3601. doi: 10.3390/s25123601.

本文引用的文献

1
A Survey on Learning to Hash.关于学习哈希的调查。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):769-790. doi: 10.1109/TPAMI.2017.2699960. Epub 2017 May 2.