Local LLMs versus Offline Wikipedia
10 months ago
- #LLM
- #Wikipedia
- #offline
- MIT Technology Review article discusses using offline LLMs in an apocalypse scenario.
- Comparison of local LLMs and offline Wikipedia downloads by size.
- List includes models like Qwen, Deepseek-R1, Llama, and Gemma, alongside Wikipedia bundles.
- Caveats noted: different purposes of encyclopedias and LLMs, hardware requirements, and non-rigorous selection.
- Interesting observation: Wikipedia's best 50,000 articles roughly equivalent to Llama 3.2 3B in size.
- Suggestion to download both LLMs and Wikipedia for offline use.