SenseAI: A Human-in-the-Loop Dataset for RLHF-Aligned Financial Sentiment Reasoning

arXiv cs.CL / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SenseAI, a human-in-the-loop (HITL) financial sentiment dataset built to include both model outputs and the underlying reasoning process, aligning with RLHF workflows.
  • SenseAI contains 1,439 labeled examples covering 40 US-listed equities and 13 financial categories, and includes reasoning chains, confidence scores, human correction signals, and links to real-world market outcomes.
  • The analysis identifies systematic LLM behaviors in financial reasoning, including a newly described failure mode called Latent Reasoning Drift where models add ungrounded information.
  • The study also reports consistent confidence miscalibration and forward projection tendencies, arguing that financial reasoning errors are structured and therefore more correctable than random.
  • The authors propose using SenseAI for targeted model improvement, including evaluation and alignment of financial AI systems using structured HITL data.

Abstract

We introduce SenseAI, a human-in-the-loop (HITL) validated financial sentiment dataset designed to capture not only model outputs but the full reasoning process behind them. Unlike existing resources, SenseAI incorporates reasoning chains, confidence scores, human correction signals, and real-world market outcomes, providing a structure aligned with Reinforcement Learning from Human Feedback (RLHF) paradigms. The dataset consists of 1,439 labelled data points across 40 US-listed equities and 13 financial data categories, enabling direct integration into modern LLM fine-tuning pipelines. Through analysis, we identify several systematic patterns in model behavior, including a novel failure mode we term Latent Reasoning Drift, where models introduce information not grounded in the input, as well as consistent confidence miscalibration and forward projection tendencies. These findings suggest that LLM errors in financial reasoning are not random but occur within a predictable and correctable regime, supporting the use of structured HITL data for targeted model improvement. We discuss implications for financial AI systems and highlight opportunities for applying SenseAI in model evaluation and alignment.