UniCreative: Unifying Long-form Logic and Short-form Sparkle via Reference-Free Reinforcement Learning

arXiv cs.AI / 4/8/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents UniCreative, a reference-free reinforcement learning framework aimed at unifying long-form narrative coherence with short-form textual expressiveness in creative writing.
  • It introduces AC-GenRM, an adaptive constraint-aware reward model that generates query-specific criteria to produce fine-grained, preference-style judgments without requiring static rewards or ground-truth references.
  • It proposes ACPO, a policy optimization method that aligns model outputs with human preferences on both content quality and structural paradigms while avoiding supervised fine-tuning and reference data.
  • Experiments report that AC-GenRM correlates closely with expert evaluations and that ACPO improves performance across a range of writing tasks.
  • The authors claim an emergent capability where the model autonomously decides when a task needs rigorous planning versus when direct generation is sufficient, supporting the effectiveness of the proposed direct alignment approach.

Abstract

A fundamental challenge in creative writing lies in reconciling the inherent tension between maintaining global coherence in long-form narratives and preserving local expressiveness in short-form texts. While long-context generation necessitates explicit macroscopic planning, short-form creativity often demands spontaneous, constraint-free expression. Existing alignment paradigms, however, typically employ static reward signals and rely heavily on high-quality supervised data, which is costly and difficult to scale. To address this, we propose \textbf{UniCreative}, a unified reference-free reinforcement learning framework. We first introduce \textbf{AC-GenRM}, an adaptive constraint-aware reward model that dynamically synthesizes query-specific criteria to provide fine-grained preference judgments. Leveraging these signals, we propose \textbf{ACPO}, a policy optimization algorithm that aligns models with human preferences across both content quality and structural paradigms without supervised fine-tuning and ground-truth references. Empirical results demonstrate that AC-GenRM aligns closely with expert evaluations, while ACPO significantly enhances performance across diverse writing tasks. Crucially, our analysis reveals an emergent meta-cognitive ability: the model learns to autonomously differentiate between tasks requiring rigorous planning and those favoring direct generation, validating the effectiveness of our direct alignment approach.