There have already been opinions that Gemma is bad because it doesn’t work well, but you probably aren’t using the transformers implementation, you’re using llama.cpp.
After a model is released, you have to wait at least a few days for all the fixes in llama.cpp, for example:
https://github.com/ggml-org/llama.cpp/pull/21418
https://github.com/ggml-org/llama.cpp/pull/21390
https://github.com/ggml-org/llama.cpp/pull/21406
https://github.com/ggml-org/llama.cpp/pull/21327
https://github.com/ggml-org/llama.cpp/pull/21343
...and maybe there will be more?
I had a looping problem in chat, but I also tried doing some stuff in OpenCode (it wasn’t even coding), and there were zero problems. So, probably just like with GLM Flash, a better prompt somehow fixes the overthinking/looping.
[link] [comments]




