LLM-guided headline rewriting for clickability enhancement without clickbait

arXiv cs.CL / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses how to improve news headline clickability while preserving semantic faithfulness, framing clickbait as an extreme case of disproportionate amplification of engagement cues.
  • It formulates headline rewriting as a controllable text generation problem using an inference-time guidance method based on the FUDGE paradigm.
  • The proposed LLM is steered by two auxiliary guide models: a clickbait scoring model for negative guidance to prevent excessive stylistic amplification, and an engagement-attribute model for positive guidance toward target clickability.
  • Training uses neutral headlines from a real-world news corpus, while clickbait examples are generated synthetically via LLM-based rewrites with controlled activation of engagement tactics.
  • By tuning guidance weights at inference time, the system can produce headlines spanning a continuum from neutral paraphrases to more engaging but editorially acceptable rewrites, enabling study of the trade-off between attractiveness, fidelity, and clickbait avoidance.

Abstract

Enhancing reader engagement while preserving informational fidelity is a central challenge in controllable text generation for news media. Optimizing news headlines for reader engagement is often conflated with clickbait, resulting in exaggerated or misleading phrasing that undermines editorial trust. We frame clickbait not as a separate stylistic category, but as an extreme outcome of disproportionate amplification of otherwise legitimate engagement cues. Based on this view, we formulate headline rewriting as a controllable generation problem, where specific engagement-oriented linguistic attributes are selectively strengthened under explicit constraints on semantic faithfulness and proportional emphasis. We present a guided headline rewriting framework built on a large language model (LLM) that uses the Future Discriminators for Generation (FUDGE) paradigm for inference-time control. The LLM is steered by two auxiliary guide models: (1) a clickbait scoring model that provides negative guidance to suppress excessive stylistic amplification, and (2) an engagement-attribute model that provides positive guidance aligned with target clickability objectives. Both guides are trained on neutral headlines drawn from a curated real-world news corpus. At the same time, clickbait variants are generated synthetically by rewriting these original headlines using an LLM under controlled activation of predefined engagement tactics. By adjusting guidance weights at inference time, the system generates headlines along a continuum from neutral paraphrases to more engaging yet editorially acceptable formulations. The proposed framework provides a principled approach for studying the trade-off between attractiveness, semantic preservation, and clickbait avoidance, and supports responsible LLM-based headline optimization in journalistic settings.