AI Navigate

DatedGPT: Preventing Lookahead Bias in Large Language Models with Time-Aware Pretraining

arXiv cs.CL / 3/13/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • DatedGPT introduces twelve 1.3B-parameter language models trained from scratch on temporally partitioned data with strict annual cutoffs from 2013 to 2024 to prevent lookahead bias in financial backtesting.
  • The models receive instruction fine-tuning on both general-domain and finance-specific datasets aligned to the same temporal cutoffs to constrain knowledge growth by time.
  • Perplexity-based probing confirms that each model's knowledge is effectively bounded by its cutoff year, reducing leakage of future information.
  • Evaluation on standard benchmarks shows competitive performance with existing models of similar scale despite the time-aware training.
  • An interactive web demo allows users to query and compare responses from models across different cutoff years, illustrating practical time-aware forecasting workflows.

Abstract

In financial backtesting, large language models pretrained on internet-scale data risk introducing lookahead bias that undermines their forecasting validity, as they may have already seen the true outcome during training. To address this, we present DatedGPT, a family of twelve 1.3B-parameter language models, each trained from scratch on approximately 100 billion tokens of temporally partitioned data with strict annual cutoffs spanning 2013 to 2024. We further enhance each model with instruction fine-tuning on both general-domain and finance-specific datasets curated to respect the same temporal boundaries. Perplexity-based probing confirms that each model's knowledge is effectively bounded by its data cutoff year, while evaluation on standard benchmarks shows competitive performance with existing models of similar scale. We provide an interactive web demo that allows users to query and compare responses from models across different cutoff years.