Learning When to Act: Interval-Aware Reinforcement Learning with Predictive Temporal Structure

arXiv cs.LG / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces an interval-aware reinforcement learning framework that learns the optimal time interval between an agent’s “cognitive ticks” from experience rather than relying on fixed or biologically inspired timers.
  • It augments the control policy with a predictive hyperbolic “curvature/spread signal” computed from the mean pairwise Poincaré distance among sampled futures, using higher spread to trigger earlier action under uncertainty.
  • The authors propose an interval-aware reward that penalizes inefficiency relative to the chosen wait time, addressing credit-assignment issues common in naive outcome-only timing rewards.
  • They also present a joint spatio-temporal embedding (ATCPG-ST) that adds spatial trajectory divergence signals from hyperbolic geometry, improving timing decisions beyond a state-only variant.
  • Experimental results report substantial efficiency gains, including raising the hyperbolic spread (kappa) from 1.88 to 3.37 and achieving up to 22.8% efficiency over a fixed-interval baseline plus an additional 5.8% from adding spatial position information.

Abstract

Autonomous agents operating in continuous environments must decide not only what to do, but when to act. We introduce a lightweight adaptive temporal control system that learns the optimal interval between cognitive ticks from experience, replacing ad hoc biologically inspired timers with a principled learned policy. The policy state is augmented with a predictive hyperbolic spread signal (a "curvature signal" shorthand) derived from hyperbolic geometry: the mean pairwise Poincare distance among n sampled futures embedded in the Poincare ball. High spread indicates a branching, uncertain future and drives the agent to act sooner; low spread signals predictability and permits longer rest intervals. We further propose an interval-aware reward that explicitly penalises inefficiency relative to the chosen wait time, correcting a systematic credit-assignment failure of naive outcome-based rewards in timing problems. We additionally introduce a joint spatio-temporal embedding (ATCPG-ST) that concatenates independently normalised state and position projections in the Poincare ball; spatial trajectory divergence provides an independent timing signal unavailable to the state-only variant (ATCPG-SO). This extension raises mean hyperbolic spread (kappa) from 1.88 to 3.37 and yields a further 5.8 percent efficiency gain over the state-only baseline. Ablation experiments across five random seeds demonstrate that (i) learning is the dominant efficiency factor (54.8 percent over no-learning), (ii) hyperbolic spread provides significant complementary gain (26.2 percent over geometry-free control), (iii) the combined system achieves 22.8 percent efficiency over the fixed-interval baseline, and (iv) adding spatial position information to the spread embedding yields an additional 5.8 percent.