Ask questions, get answers with citations — all powered by local AI. Your files never leave your machine.
macOS only for now · Windows & Linux coming soon · Requires Ollama
Finding answers buried in hundreds of pages is slow, error-prone — and sending those files to cloud AI raises serious privacy concerns.
Manually skimming through reports, contracts, and research just to find one specific fact eats up your entire day.
Uploading sensitive contracts, financial reports, or internal docs to OpenAI or Google means your data lives on someone else's servers.
ChatGPT has no context about your specific documents. Copy-pasting chunks is tedious and you miss the full picture.
DocLLM indexes your documents locally and lets you have real conversations with them — with citations so you know exactly where every answer comes from.
DocLLM runs on Ollama with models like Mistral 7B and nomic-embed-text, all on your own hardware. No API keys, no data leaving your machine — ever.
Every AI response includes exact source references — which document and which page. Click any citation to jump directly to that page in the built-in PDF viewer.
Organise documents into workspaces for different projects. Ask questions that span across multiple PDFs and DOCX files simultaneously within a workspace.
DocLLM uses Tesseract.js to automatically run OCR on scanned PDFs and image-based documents, so you can chat with any file regardless of how it was created.
Pin any AI response to your notes panel for quick reference. Export full conversation history as Markdown or TXT whenever you need to share insights with your team.
No technical skills needed. DocLLM walks you through every step the first time you open it.
Install DocLLM like any Mac app — open the .dmg, drag it to Applications, and launch it. That's it.
DocLLM needs Ollama — a free app that runs AI privately on your computer. Download it from ollama.com and install it. This is the engine that powers everything locally.
Open Terminal and run the command below. This starts the AI engine in the background — keep Terminal open while using DocLLM.
DocLLM's setup wizard automatically downloads two AI models to your Mac. This happens once and takes around 5 minutes depending on your connection. After that, everything runs fully offline.
Drag in any PDF or Word doc and ask questions in plain English. You'll get answers with exact page references in seconds. Everything stays on your computer — always.
DocLLM is a privacy-first desktop app that lets you have AI conversations with your documents. It runs entirely on your local machine using local AI models via Ollama — no data ever leaves your computer.
After the one-time model download (~4.4 GB, takes about 5 minutes), DocLLM works fully offline. All AI inference happens locally using Ollama on your machine — no internet needed.
A Mac with at least 8 GB RAM running macOS 12 or later. For best performance, 16 GB RAM is recommended. Apple Silicon Macs (M1/M2/M3/M4) run models particularly well thanks to unified memory.
DocLLM supports PDF (including scanned PDFs via OCR) and DOCX files. Support for more formats is on the roadmap.
Yes — completely. All document processing, embeddings, and AI inference happens locally on your machine. No data is ever sent to a server.
Yes — completely free, forever. DocLLM is open source under the MIT licence. Download it, use every feature, and even modify the source code. No sign-up, no payment, no limits.
Yes. From the settings panel you can switch to any model available in Ollama — Llama 3, Phi-3, Gemma, and hundreds of others. You can also configure a custom Ollama server URL.
Download DocLLM today. No sign-up, no payment, no limits.