Tarzanagh Davoud Ataee, Hou Bojian, Tong Boning, Long Qi, Shen Li
University of Pennsylvania.
Proc Mach Learn Res. 2023 Aug;216:2123-2133.
We present a novel Bayesian-based optimization framework that addresses the challenge of generalization in overparameterized models when dealing with imbalanced subgroups and limited samples per subgroup. Our proposed tri-level optimization framework utilizes predictors, which are trained on a small amount of data, as well as a fair and class-balanced predictor at the middle and lower levels. To effectively overcome saddle points for minority classes, our lower-level formulation incorporates sharpness-aware minimization. Meanwhile, at the upper level, the framework dynamically adjusts the loss function based on validation loss, ensuring a close alignment between the predictor and local predictors. Theoretical analysis demonstrates the framework's ability to enhance classification and fairness generalization, potentially resulting in improvements in the generalization bound. Empirical results validate the superior performance of our tri-level framework compared to existing state-of-the-art approaches. The source code can be found at https://github.com/PennShenLab/FACIMS.
我们提出了一种基于贝叶斯的新型优化框架,该框架在处理不平衡子组且每个子组样本有限的情况下,应对过参数化模型中的泛化挑战。我们提出的三层优化框架利用在少量数据上训练的预测器,以及在中层和下层的公平且类平衡的预测器。为了有效克服少数类的鞍点,我们的下层公式纳入了锐度感知最小化。同时,在上层,该框架根据验证损失动态调整损失函数,确保预测器与局部预测器紧密对齐。理论分析表明该框架具有增强分类和公平泛化的能力,可能会改进泛化界。实证结果验证了我们的三层框架相对于现有最先进方法的卓越性能。源代码可在https://github.com/PennShenLab/FACIMS上找到。