Goldfish memory

Reddit r/LocalLLaMA / 3/26/2026

💬 OpinionDeveloper Stack & InfrastructureTools & Practical Usage

Key Points

  • A Reddit user describes a setup using Mistral Nemo with Ollama, Docker, OpenWebUI, and Tavily, but reports that new messages are treated as if they start a fresh chat.
  • The core issue is that conversational context is not being preserved between turns, resulting in answers that ignore prior messages.
  • The post implicitly points to potential misconfiguration or limitations in how the UI, middleware, or backend handles chat history/session state.
  • The request is framed as a troubleshooting question for local LLM chat continuity when integrating multiple components (runtime, UI, and web/search tooling).

I have setup Mistral-nemo with ollama, docker, OpenWebUI and Tavily, but im having an issue when i send a new message the model has no previous context and answers it as if it was a new chat

submitted by /u/Plus_House_1078
[link] [comments]