Lightweight Fairness for LLM-Based Recommendations via Kernelized Projection and Gated Adapters

arXiv cs.LG / 3/26/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses social bias in LLM-based recommender systems, focusing on reducing attribute leakage when demographic cues are present in the input or representations.
  • It introduces a lightweight, scalable fairness method that uses a closed-form kernelized Iterative Null-space Projection (INLP) to remove sensitive attributes from LLM representations without adding trainable parameters.
  • To avoid sacrificing recommendation quality, the method adds a two-level gated Mixture-of-Experts (MoE) adapter that selectively restores task-relevant signals while aiming not to reintroduce bias.
  • Experiments on two public datasets show improved fairness (reduced leakage across multiple protected variables) alongside competitive recommendation accuracy.

Abstract

Large Language Models (LLMs) have introduced new capabilities to recommender systems, enabling dynamic, context-aware, and conversational recommendations. However, LLM-based recommender systems inherit and may amplify social biases embedded in their pre-training data, especially when demographic cues are present. Existing fairness solutions either require extra parameters fine-tuning, or suffer from optimization instability. We propose a lightweight and scalable bias mitigation method that combines a kernelized Iterative Null-space Projection (INLP) with a gated Mixture-of-Experts (MoE) adapter. Our approach estimates a closed-form projection that removes single or multiple sensitive attributes from LLM representations with no additional trainable parameters. To preserve task utility, we introduce a two-level MoE adapter that selectively restores useful signals without reintroducing bias. Experiments on two public datasets show that our method reduces attribute leakage across multiple protected variables while maintaining competitive recommendation accuracy.