Video of how my LLM's decoder blocks changed while training

Reddit r/LocalLLaMA / 4/16/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The post shares a video showing how an LLM’s decoder blocks evolve during the training process.
  • The author created the video in response to earlier interest from the community, aiming to provide a clearer visual explanation of the training dynamics.
  • The author notes that Reddit’s compression reduced the quality of the original content and points readers to an alternative X post for better viewing.
  • The submission is positioned as educational/observational rather than a release of a new model or training method.
Video of how my LLM's decoder blocks changed while training

This is in response to my popular post: https://www.reddit.com/r/LocalLLaMA/comments/1sivm24/heres_how_my_llms_decoder_block_changed_while/

It was requested that I make a video of this data, so here it is. Enjoy!

Edit: I see that reddit nuked it with compression. Let me know if my X post is any better: https://x.com/curvedinf/status/2044521120250966099

submitted by /u/1ncehost
[link] [comments]