Jin Haibo, Che Haoxuan, Chen Hao
IEEE Trans Image Process. 2024;33:4952-4965. doi: 10.1109/TIP.2024.3451937. Epub 2024 Sep 11.
Self-training is a simple yet effective method for semi-supervised learning, during which pseudo-label selection plays an important role for handling confirmation bias. Despite its popularity, applying self-training to landmark detection faces three problems: 1) The selected confident pseudo-labels often contain data bias, which may hurt model performance; 2) It is not easy to decide a proper threshold for sample selection as the localization task can be sensitive to noisy pseudo-labels; 3) coordinate regression does not output confidence, making selection-based self-training infeasible. To address the above issues, we propose Self-Training for Landmark Detection (STLD), a method that does not require explicit pseudo-label selection. Instead, STLD constructs a task curriculum to deal with confirmation bias, which progressively transitions from more confident to less confident tasks over the rounds of self-training. Pseudo pretraining and shrink regression are two essential components for such a curriculum, where the former is the first task of the curriculum for providing a better model initialization and the latter is further added in the later rounds to directly leverage the pseudo-labels in a coarse-to-fine manner. Experiments on three facial and one medical landmark detection benchmark show that STLD outperforms the existing methods consistently in both semi- and omni-supervised settings. The code is available at https://github.com/jhb86253817/STLD.
自训练是一种简单而有效的半监督学习方法,在此过程中,伪标签选择对于处理确认偏差起着重要作用。尽管它很受欢迎,但将自训练应用于地标检测面临三个问题:1)所选的置信伪标签通常包含数据偏差,这可能会损害模型性能;2)由于定位任务可能对有噪声的伪标签敏感,因此很难确定用于样本选择的合适阈值;3)坐标回归不输出置信度,使得基于选择的自训练不可行。为了解决上述问题,我们提出了用于地标检测的自训练(STLD),这是一种不需要显式伪标签选择的方法。相反,STLD构建了一个任务课程来处理确认偏差,该课程在自训练的轮次中从更有信心的任务逐渐过渡到信心较低的任务。伪预训练和收缩回归是这样一个课程的两个重要组成部分,其中前者是课程的第一个任务,用于提供更好的模型初始化,后者在后面的轮次中进一步添加,以从粗到细的方式直接利用伪标签。在三个面部和一个医学地标检测基准上的实验表明,STLD在半监督和全监督设置中均始终优于现有方法。代码可在https://github.com/jhb86253817/STLD获取。