The tinylora paper shows that we can alter model behavior with only a few parameters.
https://arxiv.org/pdf/2602.04118
I tried replicating the paper, and made a tinylora implementation for qwen3.5, and it does work, it's crazy to think about. I got the same results as the paper, for example, increasing the rank just made the optimization space too large for it to converge correctly.
What did improve it, was giving the MLP and attention layers their own shared 13 parameters to adjust. IE all mlp layers has 13 parameters together, and all attention layers has 13, so a total of 26. That was better than just increasing the number of global parameters overall or having a global 13 parameter count like in the paper.
Next I would like to try giving each individual mlp and attention layer their own parameters to optimize, maybe even 2-6 for each, to see if the individual layers can better adjust the model despite lower parameters vs. a higher number of parameters shared across more layers. To test the global vs. local optimization of the model.
My hypothesis is also that this wouldn't be well suited for memorizing facts, but it seems good at altering behavior, as I tested it on downstream tasks via lm-eval.
What this might implicate
We might be able to train models with much less memory than we initially thought, but only for changing behavior. Imagine something like the new engram from the deepseek paper,
https://github.com/deepseek-ai/Engram
But instead of an engram lookup, we could have a lookup table for behaviors made of lora adapters, much larger and more varied than Moe, which could be updated over time even, as they are very small and require very little memory to train.
[link] [comments]



