LLMs learn backwards, and the scaling hypothesis is bounded. [D]

Reddit r/MachineLearning / 4/12/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The piece argues that large language models can be trained effectively in a “backwards” (reverse-learning) manner rather than only learning left-to-right dependencies.