Relational Probing: LM-to-Graph Adaptation for Financial Prediction

arXiv cs.CL / 4/14/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes “Relational Probing,” a method that replaces a language model’s usual output head with a relation head to induce a structured relational graph from hidden states for stock-trend prediction.
  • By training the induced graph jointly with a downstream task model, the approach aims to avoid prompting-style autoregressive decoding costs and to keep graph construction aligned with downstream optimization.
  • The method is designed to preserve strict graph structure while also learning useful semantic representations, effectively transforming language-model outputs into task-specific structured formats.
  • For reproducibility, the authors introduce an operational definition of “small language models” as models fine-tunable end-to-end on a single 24GB GPU under specified batch-size and sequence-length constraints.
  • Experiments using Qwen3 backbones (0.6B/1.7B/4B) show consistent improvements over a co-occurrence baseline at competitive inference cost.

Abstract

Language models can be used to identify relationships between financial entities in text. However, while structured output mechanisms exist, prompting-based pipelines still incur autoregressive decoding costs and decouple graph construction from downstream optimization. We propose \emph{Relational Probing}, which replaces the standard language-model head with a relation head that induces a relational graph directly from language-model hidden states and is trained jointly with the downstream task model for stock-trend prediction. This approach both learns semantic representations and preserves the strict structure of the induced relational graph. It enables language-model outputs to go beyond text, allowing them to be reshaped into task-specific formats for downstream models. To enhance reproducibility, we provide an operational definition of small language models (SLMs): models that can be fine-tuned end-to-end on a single 24GB GPU under specified batch-size and sequence-length settings. Experiments use Qwen3 backbones (0.6B/1.7B/4B) as upstream SLMs and compare against a co-occurrence baseline. Relational Probing yields consistent performance improvements at competitive inference cost.