School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213-3890, USA.
IEEE Trans Image Process. 2012 Mar;21(3):1339-51. doi: 10.1109/TIP.2011.2169269. Epub 2011 Sep 23.
The number of digital images rapidly increases, and it becomes an important challenge to organize these resources effectively. As a way to facilitate image categorization and retrieval, automatic image annotation has received much research attention. Considering that there are a great number of unlabeled images available, it is beneficial to develop an effective mechanism to leverage unlabeled images for large-scale image annotation. Meanwhile, a single image is usually associated with multiple labels, which are inherently correlated to each other. A straightforward method of image annotation is to decompose the problem into multiple independent single-label problems, but this ignores the underlying correlations among different labels. In this paper, we propose a new inductive algorithm for image annotation by integrating label correlation mining and visual similarity mining into a joint framework. We first construct a graph model according to image visual features. A multilabel classifier is then trained by simultaneously uncovering the shared structure common to different labels and the visual graph embedded label prediction matrix for image annotation. We show that the globally optimal solution of the proposed framework can be obtained by performing generalized eigen-decomposition. We apply the proposed framework to both web image annotation and personal album labeling using the NUS-WIDE, MSRA MM 2.0, and Kodak image data sets, and the AUC evaluation metric. Extensive experiments on large-scale image databases collected from the web and personal album show that the proposed algorithm is capable of utilizing both labeled and unlabeled data for image annotation and outperforms other algorithms.
数字图像的数量迅速增加,有效地组织这些资源成为一个重要的挑战。作为促进图像分类和检索的一种方法,自动图像标注受到了广泛关注。考虑到有大量未标记的图像可用,开发一种有效的机制来利用未标记的图像进行大规模图像标注是有益的。同时,一张图像通常与多个标签相关联,这些标签彼此之间存在内在的相关性。一种直接的图像标注方法是将问题分解为多个独立的单标签问题,但这忽略了不同标签之间的潜在相关性。在本文中,我们提出了一种新的图像标注归纳算法,通过将标签相关性挖掘和视觉相似性挖掘整合到一个联合框架中。我们首先根据图像的视觉特征构建一个图模型。然后,通过同时揭示不同标签之间的共享结构和视觉图嵌入标签预测矩阵,训练一个多标签分类器,以实现图像标注。我们表明,通过执行广义特征分解,可以得到所提出框架的全局最优解。我们将所提出的框架应用于 NUS-WIDE、MSRA MM 2.0 和 Kodak 图像数据集的网络图像标注和个人相册标注,并使用 AUC 评估指标进行评估。在从网络和个人相册收集的大规模图像数据库上进行的广泛实验表明,所提出的算法能够利用有标记和无标记的数据进行图像标注,并且优于其他算法。