Reka Edge 2603 multimodal support has been merged into llama.cpp

Reddit r/LocalLLaMA / 4/23/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • Reka Edge 2603 multimodal support has been merged upstream into llama.cpp, making the model easier to run with the standard stack.
  • Users can start by downloading the Reka Edge 2603 weights from the Hugging Face repository and converting them to GGUF using the provided llama.cpp conversion script.
  • An optional quantization script is available to quantize the text decoder (e.g., Q4/Q8 variants) to improve efficiency.
  • The model currently lacks reasoning support, so users should run llama-server with `--reasoning off` to avoid incorrect behavior.

Hi r/LocalLLaMA! I work at Reka and organized our AMA last month. Some of y'all have asked for llama.cpp support - this is a follow-up to let you know that Reka Edge 2603 is now supported upstream in llama.cpp.

To get started:

One note: the model does not currently support reasoning, so run llama-server with `--reasoning off`. Happy hacking!

submitted by /u/Available_Poet_6387
[link] [comments]