Liu Wei, Liu Baisong, Qin Jiangcheng, Zhang Xueyuan, Huang Weiming, Wang Yangyang
Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo, 315211, China.
Sci Rep. 2025 Feb 14;15(1):5516. doi: 10.1038/s41598-025-89965-3.
Fairness in recommendation systems is crucial for ensuring equitable treatment of all users. Inspired by research on human-like behavior in large language models (LLMs), we investigate whether LLMs can serve as "fairness recognizers" in recommendation systems and explore harnessing the inherent fairness awareness in LLMs to construct fair recommendations. Using the MovieLens and LastFM datasets, we compare recommendations produced by Variational Autoencoders (VAE) with and without fairness strategies, and use ChatGLM3-6B and Llama2-13B to identify the fairness of VAE-generated results. Evaluation reveals that LLMs can indeed recognize fair recommendations by recognizing the correlation between users' sensitive attributes and their recommendation results. We then propose a method for incorporating LLMs into the recommendation process by replacing unfair recommendations identified as unfair by LLMs with those generated by a fair VAE. Our evaluation demonstrates that this approach improves fairness significantly with minimal loss in utility. For instance, the fairness-to-utility ratio for gender-based groups shows that VAEgan's results are 6.0159 and 5.0658, while ChatGLM's results achieve 30.9289 and 50.4312, respectively. These findings demonstrate that integrating LLMs' fairness recognition capabilities leads to a more favorable trade-off between fairness and utility.
推荐系统中的公平性对于确保所有用户得到公平对待至关重要。受大语言模型(LLMs)中类人行为研究的启发,我们研究LLMs是否可以在推荐系统中充当“公平性识别器”,并探索利用LLMs中固有的公平性意识来构建公平推荐。使用MovieLens和LastFM数据集,我们比较了有无公平性策略的变分自编码器(VAE)生成的推荐,并使用ChatGLM3 - 6B和Llama2 - 13B来识别VAE生成结果的公平性。评估表明,LLMs确实可以通过识别用户敏感属性与其推荐结果之间的相关性来识别公平推荐。然后,我们提出了一种将LLMs纳入推荐过程的方法,即将被LLMs识别为不公平的推荐替换为公平VAE生成的推荐。我们的评估表明,这种方法在效用损失最小的情况下显著提高了公平性。例如,基于性别的群体的公平性与效用比显示,VAEgan的结果分别为6.0159和5.0658,而ChatGLM的结果分别达到30.9289和50.4312。这些发现表明,整合LLMs的公平性识别能力会在公平性和效用之间带来更有利的权衡。