The local LLM ecosystem doesn’t need Ollama

Licensing, Attribution, and Ethics

  • Many comments focus on Ollama’s use of llama.cpp under a permissive license but with weak or late attribution.
  • Several see this as technically legal but socially harmful: taking “social upside” from open source while minimizing credit and creating lock‑in.
  • Others argue MIT-style licenses explicitly allow this; if developers wanted stronger reciprocity, they should have chosen GPL.
  • There’s broad agreement that at minimum, clear, prominent attribution should have been there from the start.

Lock‑in, Model Storage, and Interop

  • A frequent complaint: Ollama’s hashed blob storage and proprietary manifests mean downloaded models can’t be easily reused by other tools.
  • This leads to duplicated multi‑GB downloads and makes switching away costly once invested.
  • Some contrast this with tools that use standard GGUF files in Hugging Face caches, which multiple runtimes can share.

Performance and Technical Behavior

  • Many report llama.cpp (via llama-server or similar) runs the same models faster, with better memory use, and supports newer architectures sooner.
  • Concurrency and batching in llama.cpp’s server are highlighted as stronger, especially for multi-user or multi-bot setups.
  • Others note that both Ollama and llama.cpp can lag behind brand‑new models until backends update; running very new architectures may require tracking latest builds.

UX, Ease of Use, and Model Management

  • A major dividing line: some say Ollama is “1000x easier,” especially for casual or Mac users wanting a quick “just works” experience.
  • Others find modern llama.cpp workflows comparably simple (e.g., brew install + single llama-server -hf … command) and note it now ships with a web UI and OpenAI-compatible API.
  • Several tools are cited as user-friendly alternatives: LM Studio, Jan, koboldcpp, llamafile, LlamaBarn, vLLM (for servers), various MLX-based options.
  • Some users say these alternatives were slower or didn’t work on their specific hardware; others report the opposite and prefer them to Ollama.

Broader Business and Community Concerns

  • Multiple comments criticize a perceived VC/YC “wrapper + lock-in + cloud monetization” playbook.
  • A few defend Ollama as a company under pressure to monetize and argue that open source plus commercial wrappers is inevitable.
  • Several note that with strong alternatives and rapid progress in llama.cpp and related tools, Ollama’s technical and community missteps make it easier to switch.