Show HN: Offline RAG System Using Docker and Llama 3 (No Cloud APIs)
3 hours ago
- #Local AI
- #RAG
- #Docker
- BLACK FRIDAY SALE: 15% OFF all source codes with code BLACKFRIDAY.
- Production-grade, 100% offline RAG architecture for chatting with proprietary documents (PDF, TXT, Markdown) using a local LLM.
- Microservices architecture, fully containerized with Docker Compose for one-click deployment.
- Components include Ollama (Meta Llama 3 8B), mxbai-embed-large embeddings, ChromaDB vector database, Python + Streamlit backend/frontend.
- Features: 100% privacy, GPU acceleration (NVIDIA CUDA), smart ingestion, context-aware chat, one-click setup.
- Recommended hardware: Windows 10/11 (WSL2) or Linux (Ubuntu), 16GB+ RAM, NVIDIA RTX 3060 (8GB VRAM) or higher.
- Includes complete source code, Docker configuration, setup guide, and premium support.
- Available for one-time payment on Gumroad with lifetime usage.