Can Large Language Models Adequately Perform Symbolic Reasoning Over Time Series?
arXiv cs.AI / 4/27/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that extracting interpretable, context-aligned symbolic laws from time-series data is still a largely open problem for large language models despite their strengths in structured reasoning.
- It introduces SymbolBench, a benchmark for symbolic reasoning over real-world time series spanning multivariate symbolic regression, Boolean network inference, and causal discovery, covering more diverse and complex symbolic forms than prior work.
- The authors propose a closed-loop framework combining LLMs with genetic programming, where LLMs serve both as predictors and evaluators to refine symbolic hypotheses over iterations.
- Experiments show both strengths and limitations of current approaches, emphasizing the need for domain knowledge, context alignment, and explicit reasoning structure to better support automated scientific discovery.
- A publicly available implementation is provided via the project’s GitHub repository.
Related Articles

Subagents: The Building Block of Agentic AI
Dev.to

DeepSeek-V4 Models Could Change Global AI Race
AI Business

Got OpenAI's privacy filter model running on-device via ExecuTorch
Reddit r/LocalLLaMA

The Agent-Skill Illusion: Why Prompt-Based Control Fails in Multi-Agent Business Consulting Systems
Dev.to

We Built a Voice AI Receptionist in 8 Weeks — Every Decision We Made and Why
Dev.to