The local LLM ecosystem doesn’t need Ollama
Licensing, Attribution, and Ethics
- Many comments focus on Ollama’s use of llama.cpp under a permissive license but with weak or late attribution.
- Several see this as technically legal but socially harmful: taking “social upside” from open source while minimizing credit and creating lock‑in.
- Others argue MIT-style licenses explicitly allow this; if developers wanted stronger reciprocity, they should have chosen GPL.
- There’s broad agreement that at minimum, clear, prominent attribution should have been there from the start.
Lock‑in, Model Storage, and Interop
- A frequent complaint: Ollama’s hashed blob storage and proprietary manifests mean downloaded models can’t be easily reused by other tools.
- This leads to duplicated multi‑GB downloads and makes switching away costly once invested.
- Some contrast this with tools that use standard GGUF files in Hugging Face caches, which multiple runtimes can share.
Performance and Technical Behavior
- Many report llama.cpp (via
llama-serveror similar) runs the same models faster, with better memory use, and supports newer architectures sooner. - Concurrency and batching in llama.cpp’s server are highlighted as stronger, especially for multi-user or multi-bot setups.
- Others note that both Ollama and llama.cpp can lag behind brand‑new models until backends update; running very new architectures may require tracking latest builds.
UX, Ease of Use, and Model Management
- A major dividing line: some say Ollama is “1000x easier,” especially for casual or Mac users wanting a quick “just works” experience.
- Others find modern llama.cpp workflows comparably simple (e.g.,
brew install+ singlellama-server -hf …command) and note it now ships with a web UI and OpenAI-compatible API. - Several tools are cited as user-friendly alternatives: LM Studio, Jan, koboldcpp, llamafile, LlamaBarn, vLLM (for servers), various MLX-based options.
- Some users say these alternatives were slower or didn’t work on their specific hardware; others report the opposite and prefer them to Ollama.
Broader Business and Community Concerns
- Multiple comments criticize a perceived VC/YC “wrapper + lock-in + cloud monetization” playbook.
- A few defend Ollama as a company under pressure to monetize and argue that open source plus commercial wrappers is inevitable.
- Several note that with strong alternatives and rapid progress in llama.cpp and related tools, Ollama’s technical and community missteps make it easier to switch.