Churchill Rob, Singh Lisa
Department of Computer Science, Georgetown University, 3700 O Street, Washington, D.C., 20007 USA.
Knowl Inf Syst. 2023;65(5):2159-2186. doi: 10.1007/s10115-022-01805-2. Epub 2023 Jan 16.
Domain-specific document collections, such as data sets about the COVID-19 pandemic, politics, and sports, have become more common as platforms grow and develop better ways to connect people whose interests align. These data sets come from many different sources, ranging from traditional sources like open-ended surveys and newspaper articles to one of the dozens of online social media platforms. Most topic models are equipped to generate topics from one or more of these data sources, but models rarely work well across all types of documents. The main problem that many models face is the varying noise levels inherent in different types of documents. We propose topic-noise models, a new type of topic model that jointly models topic and noise distributions to produce a more accurate, flexible representation of documents regardless of their origin and varying qualities. Our topic-noise model, Topic Noise Discriminator (TND) approximates topic and noise distributions side-by-side with the help of word embedding spaces. While topic-noise models are important for the types of short, noisy documents that often originate on social media platforms, TND can also be used with more traditional data sources like newspapers. TND itself generates a noise distribution that when ensembled with other generative topic models can produce more coherent and diverse topic sets. We show the effectiveness of this approach using Latent Dirichlet Allocation (LDA), and demonstrate the ability of TND to improve the quality of LDA topics in noisy document collections. Finally, researchers are beginning to generate topics using multiple sources and finding that they need a way to identify a core set based on text from different sources. We propose using cross-source topic blending (CSTB), an approach that maps topics sets to an -partite graph and identifies core topics that blend topics from across sources by identifying subgraphs with certain linkage properties. We demonstrate the effectiveness of topic-noise models and CSTB empirically on large real-world data sets from multiple domains and data sources.
特定领域的文档集合,例如关于新冠疫情、政治和体育的数据集,随着平台的发展以及连接兴趣相投之人的方式不断完善而变得愈发常见。这些数据集来源广泛,从开放式调查和报纸文章等传统来源到众多在线社交媒体平台中的某一个。大多数主题模型能够从这些数据源中的一个或多个生成主题,但很少有模型能在所有类型的文档上都表现良好。许多模型面临的主要问题是不同类型文档中固有的噪声水平各异。我们提出了主题 - 噪声模型,这是一种新型的主题模型,它联合对主题和噪声分布进行建模,以生成更准确、灵活的文档表示,而不论其来源和质量差异如何。我们的主题 - 噪声模型,即主题噪声判别器(TND),借助词嵌入空间并行近似主题和噪声分布。虽然主题 - 噪声模型对于通常源自社交媒体平台的短且有噪声的文档类型很重要,但TND也可用于报纸等更传统的数据源。TND本身会生成一个噪声分布,当与其他生成式主题模型结合使用时,可以产生更连贯、多样的主题集。我们使用潜在狄利克雷分配(LDA)展示了这种方法的有效性,并证明了TND在有噪声文档集合中提高LDA主题质量的能力。最后,研究人员开始使用多个来源生成主题,并发现他们需要一种基于不同来源的文本识别核心主题集的方法。我们提出使用跨源主题融合(CSTB),该方法将主题集映射到一个 - 部图,并通过识别具有特定链接属性的子图来识别融合来自多个来源主题的核心主题。我们通过多个领域和数据源的大型真实世界数据集,实证证明了主题 - 噪声模型和CSTB的有效性。