RAG or Learning? Understanding the Limits of LLM Adaptation under Continuous Knowledge Drift in the Real World

arXiv cs.CL / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • LLMs are tied to the fixed knowledge learned during pretraining, and continuous real-world knowledge drift can cause both outdated outputs and temporally inconsistent reasoning.
  • The paper argues that common adaptation approaches (continual fine-tuning, knowledge editing, and RAG) are insufficiently tested in benchmarks that reflect chronological, evolving knowledge.
  • It introduces a new time-stamped benchmark using dynamic real-world events to evaluate adaptation under continuous knowledge drift.
  • Results show that many existing methods, including vanilla RAG and learning-based approaches, struggle, with issues like catastrophic forgetting and temporal inconsistency.
  • To address this without extra training, the authors propose Chronos, a time-aware retrieval baseline that builds an Event Evolution Graph from progressively organized evidence to improve temporal consistency.

Abstract

Large language models (LLMs) acquire most of their knowledge during pretraining, which ties them to a fixed snapshot of the world and makes adaptation to continuously evolving knowledge challenging. As facts, entities, and events change over time, models may experience continuous knowledge drift, resulting not only in outdated predictions but also in temporally inconsistent reasoning. Although existing approaches, such as continual finetuning, knowledge editing, and retrieval-augmented generation (RAG), aim to update or supplement model knowledge, they are rarely evaluated in settings that reflect chronological, evolving, and real-world knowledge evolution. In this work, we introduce a new benchmark of real-world dynamic events, constructed from time-stamped evidence that captures how knowledge evolves over time, which enables systematic evaluation of model adaptation under continuous knowledge drift. The benchmark reveals that most existing methods, including vanilla RAG and several learning-based approaches, struggle under this setting, exposing critical limitations such as catastrophic forgetting and temporal inconsistency. To mitigate these limitations, we propose a time-aware retrieval baseline, Chronos, which progressively organizes retrieved evidence into an Event Evolution Graph to enable more temporally consistent understanding in LLMs without additional training. Overall, this work provides a foundation for analyzing and advancing LLM adaptation to continuous knowledge drift in realistic settings.