Song Mofei
School of Computer Science and Engineering, Southeast University, Nanjing 211189, China.
Entropy (Basel). 2020 Nov 18;22(11):1314. doi: 10.3390/e22111314.
Currently, deep learning has shown state-of-the-art performance in image classification with pre-defined taxonomy. However, in a more real-world scenario, different users usually have different classification intents given an image collection. To satisfactorily personalize the requirement, we propose an interactive image classification system with an offline representation learning stage and an online classification stage. During the offline stage, we learn a deep model to extract the feature with higher flexibility and scalability for different users' preferences. Instead of training the model only with the inter-class discrimination, we also encode the similarity between the semantic-embedding vectors of the category labels into the model. This makes the extracted feature adapt to multiple taxonomies with different granularities. During the online session, an annotation task iteratively alternates with a high-throughput verification task. When performing the verification task, the users are only required to indicate the incorrect prediction without giving the exact category label. For each iteration, our system chooses the images to be annotated or verified based on interactive efficiency optimization. To provide a high interactive rate, a unified active learning algorithm is used to search the optimal annotation and verification set by minimizing the expected time cost. After interactive annotation and verification, the new classified images are used to train a customized classifier online, which reflects the user-adaptive intent of categorization. The learned classifier is then used for subsequent annotation and verification tasks. Experimental results under several public image datasets show that our method outperforms existing methods.
当前,深度学习在具有预定义分类法的图像分类中已展现出最先进的性能。然而,在更真实的场景中,给定一个图像集,不同用户通常有不同的分类意图。为了令人满意地满足个性化需求,我们提出了一个具有离线表示学习阶段和在线分类阶段的交互式图像分类系统。在离线阶段,我们学习一个深度模型,以针对不同用户的偏好提取具有更高灵活性和可扩展性的特征。我们不仅使用类间判别来训练模型,还将类别标签的语义嵌入向量之间的相似性编码到模型中。这使得提取的特征能够适应具有不同粒度的多个分类法。在在线会话期间,一个标注任务与一个高通量验证任务交替进行。在执行验证任务时,只要求用户指出错误的预测,而无需给出确切的类别标签。对于每次迭代,我们的系统基于交互效率优化选择要标注或验证的图像。为了提供高交互率,使用一种统一的主动学习算法,通过最小化预期时间成本来搜索最优的标注和验证集。经过交互式标注和验证后,新分类的图像用于在线训练定制的分类器,该分类器反映了用户自适应的分类意图。然后,学习到的分类器用于后续的标注和验证任务。在几个公共图像数据集上的实验结果表明,我们的方法优于现有方法。