• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种用于文本分类的混合重新融合模型。

A hybrid re-fusion model for text classification.

作者信息

Liu Qi, Xiao Kejing, Qian Zhaopeng

机构信息

School of Information Engineering, Beijing Institute of Graphic Communication, Beijing, China.

School of Computer and Artificial Intelligence, Beijing Technology and Business University, Beijing, China.

出版信息

Sci Rep. 2025 Mar 18;15(1):9333. doi: 10.1038/s41598-025-90864-w.

DOI:10.1038/s41598-025-90864-w
PMID:40102513
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11920417/
Abstract

Text classification is an important task in the field of natural language processing, aiming to automatically assign text data to predefined categories. The BertGCN model combines the advantages from both BERT and GCN, enabling it to effectively handle text data for classification. However, there are still some limitations when it comes to handling complex text classification tasks. BERT processes sequence information in segments and cannot directly capture long-distance dependencies across segments, which is a limitation when dealing with long sequences. GCN tends to suffer from over-smoothing problem in deep networks, leading to information loss. To overcome these limitations, we propose the XLG-Net model, which integrates XLNet and GCNII to enhance text classification performance. XLNet employs permutation language modeling and improvements of the Transformer-XL architecture, not only improving the ability to capture long-distance dependencies but also enhancing the model's understanding of complex language structures. Additionally, we introduce GCNII to overcome the over-smoothing problem in GCN. GCNII effectively retains the initial features of nodes by incorporating initial residual connections and identity mapping mechanisms, ensuring effective information transmission even in deep networks. Furthermore, to achieve excellent performance on both long and short texts, we apply the design philosophy of DoubleMix to the XLNet model, using a hybrid approach of mixing hidden states improves the model's accuracy and robustness. Experimental results demonstrate that the XLG-Net model achieves significant performance improvements on four benchmark text classification datasets, validating the model's effectiveness on complex text classification tasks.

摘要

文本分类是自然语言处理领域中的一项重要任务,旨在将文本数据自动分配到预定义的类别中。BertGCN模型结合了BERT和GCN的优点,使其能够有效地处理文本数据进行分类。然而,在处理复杂的文本分类任务时仍存在一些局限性。BERT按段处理序列信息,无法直接捕捉跨段的长距离依赖关系,这在处理长序列时是一个限制。GCN在深度网络中容易出现过平滑问题,导致信息丢失。为了克服这些局限性,我们提出了XLG-Net模型,该模型集成了XLNet和GCNII以提高文本分类性能。XLNet采用排列语言建模和对Transformer-XL架构的改进,不仅提高了捕捉长距离依赖关系的能力,还增强了模型对复杂语言结构的理解。此外,我们引入GCNII来克服GCN中的过平滑问题。GCNII通过合并初始残差连接和恒等映射机制有效地保留了节点的初始特征,确保即使在深度网络中也能进行有效的信息传输。此外,为了在长文本和短文本上都取得优异的性能,我们将DoubleMix的设计理念应用于XLNet模型,使用混合隐藏状态的混合方法提高了模型的准确性和鲁棒性。实验结果表明,XLG-Net模型在四个基准文本分类数据集上取得了显著的性能提升,验证了该模型在复杂文本分类任务上的有效性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/25c1b656700a/41598_2025_90864_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/ea8afa4834e9/41598_2025_90864_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/0a0d2c492eac/41598_2025_90864_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/ec366e0e9db5/41598_2025_90864_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/d4edf0e4b0c9/41598_2025_90864_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/743fa2cda264/41598_2025_90864_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/716adcb1b93c/41598_2025_90864_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/25c1b656700a/41598_2025_90864_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/ea8afa4834e9/41598_2025_90864_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/0a0d2c492eac/41598_2025_90864_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/ec366e0e9db5/41598_2025_90864_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/d4edf0e4b0c9/41598_2025_90864_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/743fa2cda264/41598_2025_90864_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/716adcb1b93c/41598_2025_90864_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bee9/11920417/25c1b656700a/41598_2025_90864_Fig6_HTML.jpg

相似文献

1
A hybrid re-fusion model for text classification.一种用于文本分类的混合重新融合模型。
Sci Rep. 2025 Mar 18;15(1):9333. doi: 10.1038/s41598-025-90864-w.
2
Positional embeddings and zero-shot learning using BERT for molecular-property prediction.使用BERT进行位置嵌入和零样本学习以预测分子性质
J Cheminform. 2025 Feb 5;17(1):17. doi: 10.1186/s13321-025-00959-9.
3
Relationship extraction between entities with long distance dependencies and noise based on semantic and syntactic features.基于语义和句法特征的具有长距离依存关系和噪声的实体之间的关系提取
Sci Rep. 2025 May 6;15(1):15750. doi: 10.1038/s41598-025-00915-5.
4
Enhanced Pneumonia Detection in Chest X-Rays Using Hybrid Convolutional and Vision Transformer Networks.使用混合卷积和视觉Transformer网络增强胸部X光片中的肺炎检测
Curr Med Imaging. 2025;21:e15734056326685. doi: 10.2174/0115734056326685250101113959.
5
Chinese text classification by combining Chinese-BERTology-wwm and GCN.结合中文BERTology-wwm和图卷积网络进行中文文本分类。
PeerJ Comput Sci. 2023 Aug 17;9:e1544. doi: 10.7717/peerj-cs.1544. eCollection 2023.
6
Brain tumor segmentation and detection in MRI using convolutional neural networks and VGG16.使用卷积神经网络和VGG16在磁共振成像(MRI)中进行脑肿瘤分割与检测
Cancer Biomark. 2025 Mar;42(3):18758592241311184. doi: 10.1177/18758592241311184. Epub 2025 Apr 4.
7
Research on adverse event classification algorithm of da Vinci surgical robot based on Bert-BiLSTM model.基于Bert-BiLSTM模型的达芬奇手术机器人不良事件分类算法研究
Front Comput Neurosci. 2024 Dec 16;18:1476164. doi: 10.3389/fncom.2024.1476164. eCollection 2024.
8
Chinese text classification method based on sentence information enhancement and feature fusion.基于句子信息增强与特征融合的中文文本分类方法
Heliyon. 2024 Aug 24;10(17):e36861. doi: 10.1016/j.heliyon.2024.e36861. eCollection 2024 Sep 15.
9
Dual selective fusion transformer network for hyperspectral image classification.用于高光谱图像分类的双选择性融合变压器网络
Neural Netw. 2025 Jul;187:107311. doi: 10.1016/j.neunet.2025.107311. Epub 2025 Mar 5.
10
A hybrid transformer and attention based recurrent neural network for robust and interpretable sentiment analysis of tweets.基于混合变压器和注意力的循环神经网络的 tweet 情感分析的鲁棒性和可解释性。
Sci Rep. 2024 Oct 22;14(1):24882. doi: 10.1038/s41598-024-76079-5.

本文引用的文献

1
RB-GAT: A Text Classification Model Based on RoBERTa-BiGRU with Graph ATtention Network.RB-GAT:一种基于带有图注意力网络的RoBERTa-双向门控循环单元的文本分类模型。
Sensors (Basel). 2024 May 24;24(11):3365. doi: 10.3390/s24113365.
2
Bert-Enhanced Text Graph Neural Network for Classification.用于分类的Bert增强文本图神经网络
Entropy (Basel). 2021 Nov 18;23(11):1536. doi: 10.3390/e23111536.
3
A Comprehensive Survey on Graph Neural Networks.图神经网络综述。
IEEE Trans Neural Netw Learn Syst. 2021 Jan;32(1):4-24. doi: 10.1109/TNNLS.2020.2978386. Epub 2021 Jan 4.