Neuro-Symbolic Learning for Predictive Process Monitoring via Two-Stage Logic Tensor Networks with Rule Pruning

arXiv cs.AI / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a neuro-symbolic framework for predictive process monitoring on sequential event data by embedding domain logic as differentiable constraints using Logic Networks (LTNs).
  • It encodes control-flow, temporal, and payload knowledge with Linear Temporal Logic and first-order logic to enforce rule-based relationships relevant to domains like healthcare and fraud/compliance.
  • To address a common LTNs tradeoff where logic satisfaction can harm prediction quality, it introduces a two-stage optimization strategy that uses weighted axiom loss in pretraining followed by rule pruning based on satisfaction dynamics.
  • Experiments on four real-world event logs show that injecting domain constraints improves predictive performance, and that the two-stage optimization is crucial because naive knowledge integration can degrade results.
  • The method is reported to perform especially well in compliance-constrained settings with limited compliant training examples, outperforming purely data-driven baselines while maintaining constraint adherence.

Abstract

Predictive modeling on sequential event data is critical for fraud detection and healthcare monitoring. Existing data-driven approaches learn correlations from historical data but fail to incorporate domain-specific sequential constraints and logical rules governing event relationships, limiting accuracy and regulatory compliance. For example, healthcare procedures must follow specific sequences, and financial transactions must adhere to compliance rules. We present a neuro-symbolic approach integrating domain knowledge as differentiable logical constraints using Logic Networks (LTNs). We formalize control-flow, temporal, and payload knowledge using Linear Temporal Logic and first-order logic. Our key contribution is a two-stage optimization strategy addressing LTNs' tendency to satisfy logical formulas at the expense of predictive accuracy. The approach uses weighted axiom loss during pretraining to prioritize data learning, followed by rule pruning that retains only consistent, contributive axioms based on satisfaction dynamics. Evaluation on four real-world event logs shows that domain knowledge injection significantly improves predictive performance, with the two-stage optimization proving essential knowledge (without it, knowledge can severely degrade performance). The approach excels particularly in compliance-constrained scenarios with limited compliant training examples, achieving superior performance compared to purely data-driven baselines while ensuring adherence to domain constraints.