| What it does Upload documents, URLs, or YouTube videos as sources. SoyLM analyzes them with a local LLM, stores structured summaries in SQLite, and lets you chat with your sources using RAG (FTS5 + BM25) and optional web search (DuckDuckGo). Features Source ingestion — Files, web URLs (with Playwright JS rendering fallback), YouTube transcripts Local LLM — Nemotron-Nano-9B via vLLM (OpenAI-compatible API), thinking mode for inference RAG search — SQLite FTS5 full-text search with BM25 ranking Web search — DuckDuckGo integration for supplementing source data SSE streaming — Real-time streamed responses Chat history — Persistent chat logs with JSON export Deduplication — SHA-256 hash prevents duplicate sources if you want to build: https://github.com/soy-tuber/SoyLM my media: https://media.patentllm.org/en/ [link] [comments] |
Open-source local NotebookLM alternative powered by Nemotron + RAG (no cloud API needed)
Reddit r/LocalLLaMA / 3/14/2026
📰 NewsDeveloper Stack & InfrastructureTools & Practical Usage
Key Points
- Open-source local NotebookLM-style assistant that runs without cloud APIs, powered by Nemotron and RAG.
- It can ingest documents, URLs, or YouTube transcripts and uses a local Nemotron-Nano-9B LLM via vLLM with a thinking mode and SSE streaming.
- It stores structured summaries in SQLite and performs RAG search using SQLite FTS5 with BM25 ranking, with optional DuckDuckGo web search to supplement data.
- It supports deduplication via SHA-256, persistent chat history export, and provides a link to the SoyLM project for building it locally.



