Exploring LLM biases to manipulate AI search overview

arXiv cs.AI / 5/4/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how biases in large language models affect LLM Overview systems used in web search overviews, particularly in the source-selection stage.
  • It introduces a reinforcement-learning approach that rewrites search snippets so they become more likely to be selected and featured by an LLM Overview.
  • The experiments constrain the rewriting policy to snippet text only and limit reward-hacking, aiming to reflect realistic constraints in web search environments.
  • Results show that LLM Overview systems exhibit bias and that reinforcement learning can often optimize snippet content to manipulate the resulting overviews.
  • The study also finds that LLM Overview selections depend on relative (comparative) advantages among candidate sources and demonstrates safety risks such as context-poisoning leading to inaccurate or harmful results.

Abstract

Modern large language models (LLMs) are used in many business applications in general, and specifically in web search systems and applications that generate overviews of search results - LLM Overview systems. Such systems are using an LLM to select most relevant sources from search results and generate an answer to the user's query. It is known from many studies that LLMs have different biases, in LLM Overview application both the source selection and answer generation stages may be affected by the biases of LLMs (here we are focusing mainly on the selection stage). This research is focused on investigating the presence of the biases in LLM Overview systems and on biases exploitation to manipulate LLM Overview results. Here we train a small language model using reinforcement learning to rewrite search snippets to increase their likelihood of being preferred by an LLM Overview. Our experimental setup intentionally restricts the policy to operate only on snippets and limits reward-hacking strategies, reflecting realistic constraints of web search environments. The results prove that LLM Overview systems have biases and that reinforcement learning in most of the cases can optimize snippet's content to manipulate LLM Overview results. We also prove that LLM Overview selections are driven by comparative rather than absolute advantages among candidate sources. In addition, we examine safety aspects of LLM Overview manipulation possibilities and show that context poisoning attacks can lead to inaccurate or harmful results.