Self Knowledge Re-expression: A Fully Local Method for Adapting LLMs to Tasks Using Intrinsic Knowledge

arXiv cs.CL / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that LLM performance on specialized, non-generative tasks is limited by the way the model expresses its intrinsic knowledge under the next-token prediction paradigm.
  • It proposes Self-Knowledge Re-expression (SKR), a task-agnostic adaptation method that converts generic token generation into efficient, task-specific outputs.
  • SKR is fully local and requires only unannotated data, with no human supervision and no model distillation.
  • Experiments on financial-document data report large gains across tasks, including over 40% improvement in Recall@1 for retrieval, over 76% lower object detection latency, and over 33% higher anomaly detection AUPRC.
  • On the MMDocRAG dataset, SKR achieves results that beat leading retrieval models by at least 12.6%.

Abstract

While the next-token prediction (NTP) paradigm enables large language models (LLMs) to express their intrinsic knowledge, its sequential nature constrains performance on specialized, non-generative tasks. We attribute this performance bottleneck to the LLMs' knowledge expression mechanism, rather than to deficiencies in knowledge acquisition. To address this, we propose Self-Knowledge Re-expression (SKR), a novel, task-agnostic adaptation method. SKR transforms the LLM's output from generic token generation to highly efficient, task-specific expression. SKR is a fully local method that uses only unannotated data, requiring neither human supervision nor model distillation. Experiments on a large financial document dataset demonstrate substantial improvements: over 40% in Recall@1 for information retrieval tasks, over 76% reduction in object detection latency, and over 33% increase in anomaly detection AUPRC. Our results on the MMDocRAG dataset surpass those of leading retrieval models by at least 12.6%.