Impact of Task Phrasing on Presumptions in Large Language Models

arXiv cs.AI / 5/4/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The study investigates how the wording of tasks (task phrasing) can induce hidden assumptions (“presumptions”) in large language models (LLMs), which then limit adaptation when real-world tasks differ.
  • Using the iterated prisoner’s dilemma as a case study, the researchers show that LLM decision-making can be strongly affected by these presumptions even when the model is prompted to reason.
  • The experiments find that more neutral task phrasing reduces the emergence of presumptions, enabling the models to perform more logically under the same scenario.
  • The results emphasize that careful prompt/task design is important for improving the safety and reliability of LLMs in unpredictable applications.

Abstract

Concerns with the safety and reliability of applying large-language models (LLMs) in unpredictable real-world applications motivate this study, which examines how task phrasing can lead to presumptions in LLMs, making it difficult for them to adapt when the task deviates from these assumptions. We investigated the impact of these presumptions on the performance of LLMs using the iterated prisoner's dilemma as a case study. Our experiments reveal that LLMs are susceptible to presumptions when making decisions even with reasoning steps. However, when the task phrasing was neutral, the models demonstrated logical reasoning without much presumptions. These findings highlight the importance of proper task phrasing to reduce the risk of presumptions in LLMs.