Here's one less-than-helpful result from HuggingFace's takeover of ggml.
When I launched the latest build of llama-server, it automatically did this:
================================================================================ WARNING: Migrating cache to HuggingFace cache directory Old cache: /home/user/.cache/llama.cpp/ New cache: /home/user/GEN-AI/hf_cache/hub This one-time migration moves models previously downloaded with -hf from the legacy llama.cpp cache to the standard HuggingFace cache. Models downloaded with --model-url are not affected. ================================================================================
And all of my .gguf models were moved and converted into blobs. That means that my launch scripts all fail since the models are no longer where they were supposed to be...
srv load_model: failed to load model, '/home/user/GEN-AI/hf_cache/models/ggml-org_gpt-oss-20b-GGUF_gpt-oss-20b-mxfp4.gguf' It also breaks all my model management scripts for distributing ggufs around to various machines.
The change was added in commit b8498 four days ago. Who releases a breaking change like this without the ability to stop the process before making irreversible changes to user files? I knew the HuggingFace takeover would screw things up.
[link] [comments]




