Position: No Retroactive Cure for Infringement during Training

arXiv cs.AI / 4/22/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that post-hoc mitigation techniques in generative AI—such as machine unlearning and inference-time guardrails—cannot retroactively eliminate legal liability for unlawful data acquisition and training.
  • It explains that unauthorized copying/ingestion can be a legally completed act, and that model weights can function like fixed copies preserving training-derived expressive value, making later filtering insufficient.
  • The authors contend that contract and tort/unfair-competition rules (including licenses, Terms of Service, and anti-free-riding principles) can independently govern access and use, sometimes bypassing copyright defenses like fair use or TDM exceptions.
  • Because protected input value may persist in model weights, the paper suggests remedies could involve stripping gains and, in some cases, reaching the model itself.
  • Overall, the paper calls for moving away from “post-hoc sanitization” toward “verifiable ex-ante process compliance” for data lineage and lawful training practices.

Abstract

As generative AI faces intensifying legal challenges, the machine learning community has increasingly relied on post-hoc mitigation -- especially machine unlearning and inference-time guardrails -- to argue for compliance. This paper argues that such post-hoc mitigation methods cannot retroactively cure liability from unlawful acquisition and training, because compliance hinges on data lineage, not the outputs. Our argument has three parts. First, unauthorized copying/ingestion can be a legally complete completed act, and model weights may operate as fixed copies that retain training-derived expressive value, making later filtering beside the point for infringement. Second, contract and tort/unfair-competition rules -- via licenses, terms of service, and anti-free-riding principles -- can independently restrict access and use, often bypassing copyright defenses (e.g., fair use or TDM exceptions). Third, since value from protected inputs can persist in weights, remedies such as unjust enrichment and disgorgement may require stripping gains and, in some cases, reaching the model itself. We therefore argue for a shift from Post-Hoc Sanitization to verifiable Ex-Ante Process Compliance.