I'm having a lot of fun playing with llama-server testing various flags, models and runtimes. I'm starting to wonder what's next to build out my homelab AI stack. Do I use Open WebUI for RAG/Search? Should I take a stab at something like LangGraph? My goal is to create as something as close to Claude as I can using local hardware.
[link] [comments]




