Built a surgical weight editor for local GGUF models, edit individual weights directly, no GPU, no training loop (open source)

Reddit r/LocalLLaMA / 3/30/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • PickyTrain is an open-source “surgical” weight editor for local GGUF models that allows direct per-weight inspection and patching without GPU and without a training loop.
  • The tool provides semantic awareness of model structure (e.g., attention heads, FFN layers, embeddings), presenting tensors by layer/type and issuing impact warnings when edits affect coupled weights.
  • It supports converting GGUF to an intermediate PTXY fluid format, and editing weights by coordinate (layer, row, col) with drift guardrails plus rollback journaling/undo for multiple changes.
  • Built with a Rust core for performance and safety (with Python bindings via PyO3), it currently targets common GGUF quantization formats including Q4_K, Q6_K, Q8_0, and F16/F32.
  • The release is described as early/under development and invites contributions for future features such as live learning, a visual matrix UI, and model merging.
Built a surgical weight editor for local GGUF models, edit individual weights directly, no GPU, no training loop (open source)

For the past few months I couldn't find a tool that can help me map or view the actual model weights anywhere else something that sits between "run the model as-is" and "fully retrain it."

PickyTrain lets you open a GGUF model and edit individual weights directly. Think of it like a hex editor for AI models, but with semantic awareness it knows which tensors are attention heads, which are FFN layers, which are embeddings, and warns you when an edit affects coupled weights.

What it does:

  • Converts GGUF → PTXY (a new ORIGINAL fluid format with STATIC/FLUID zones)
  • Browse all 254 tensors of a Llama model by layer and type
  • Read, patch, or nudge any weight by coordinate (layer, row, col)
  • Impact warnings before you edit (e.g. "editing ffn_gate also affects ffn_up and ffn_down")
  • Drift guardrails to prevent model collapse
  • Full rollback journal — undo any number of edits
  • CLI tool, Python library, and a curses TUI

The core engine is written in Rust (fast, safe, no garbage collector) with Python bindings via PyO3. Supports Q4_K, Q6_K, Q8_0, F16, F32 all the common GGUF quantization formats.

This is an early release and there's a lot more I want to build (live learning engine, visual matrix UI, model merging). Would love contributors and feedback.

(NOTE: STILL UNDER DEVELOPMENT, YOU MIGHT FIND BUGS!)

GitHub: https://github.com/Ainix-dev/PickyTrain

submitted by /u/SpiritOk6612
[link] [comments]