| This is in response to my popular post: https://www.reddit.com/r/LocalLLaMA/comments/1sivm24/heres_how_my_llms_decoder_block_changed_while/ It was requested that I make a video of this data, so here it is. Enjoy! Edit: I see that reddit nuked it with compression. Let me know if my X post is any better: https://x.com/curvedinf/status/2044521120250966099 [link] [comments] |
Video of how my LLM's decoder blocks changed while training
Reddit r/LocalLLaMA / 4/16/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The post shares a video showing how an LLM’s decoder blocks evolve during the training process.
- The author created the video in response to earlier interest from the community, aiming to provide a clearer visual explanation of the training dynamics.
- The author notes that Reddit’s compression reduced the quality of the original content and points readers to an alternative X post for better viewing.
- The submission is positioned as educational/observational rather than a release of a new model or training method.
Related Articles

Black Hat Asia
AI Business
Are gamers being used as free labeling labor? The rise of "Simulators" that look like AI training grounds [D]
Reddit r/MachineLearning

I built a trading intelligence MCP server in 2 days — here's how
Dev.to

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to
Qwen3.5-35B running well on RTX4060 Ti 16GB at 60 tok/s
Reddit r/LocalLLaMA