Ollama Turbo

Partnership, branding, and “local” identity

  • Launch is seen as coordinated with OpenAI via the gpt-oss models; some view this as OpenAI “oss-washing” using Ollama’s reputation.
  • Several commenters are surprised Ollama is an independent, VC-backed company and not part of Meta; some say learning this improves their opinion.
  • Concern that “Ollama” had become synonymous with local, offline use, and this move shifts focus toward being a conventional cloud provider.

Open source, governance, and investor influence

  • Multiple comments argue the real issue is governance, not “open source” per se: without independent foundations, companies can later relicense or restrict (Redis, Elastic, Mongo cited).
  • Ollama is praised for MIT-licensed server code but criticized for being controlled by a single VC-backed company, making long‑term direction and licensing uncertain.
  • Some say investor funding made this kind of monetization inevitable and that people should have expected it.

llama.cpp / ggml attribution and engine debate

  • Strong sentiment that the llama.cpp/ggml author “brought LLMs to the masses” and deserves far more credit and money.
  • Dispute over how much Ollama is “just a wrapper”:
    • Ollama team says they now have their own engine, using ggml for tensors and llama.cpp only for legacy models.
    • Critics reply that ggml is effectively the core of llama.cpp, that differences are small, and accuse Ollama of minimizing this dependence and “gaslighting.”
  • Some users are leaving Ollama for llama.cpp + llama-server, saying it now matches or exceeds Ollama’s usability.

Value proposition and pricing of Turbo

  • $20/month flat fee is compared to ChatGPT/Claude; many want cheaper or purely usage-based options and dislike unspecified “hourly and daily limits.”
  • Supporters see value in:
    • Easy way to test big open models without buying GPUs.
    • A simple, unified local/cloud dev story.
  • Skeptics question why pay $20 for quantized open models when state-of-the-art proprietary models cost the same or less via usage-based APIs.

Privacy, jurisdiction, and data handling

  • “Privacy first” marketing is viewed as under-specified; lack of detailed policies and closed-source desktop app reduce trust.
  • Some see no privacy advantage over other US-based providers; others would pay more for EU/Swiss hosting.
  • Debate over whether US jurisdiction is safer or more dangerous than EU/China; consensus only that local remains best for sensitive data.

Local vs cloud; production vs hobby use

  • Many still see Ollama as an excellent on-ramp: install, download models, and go—especially for less technical users.
  • Some argue it’s mainly a “toy” for individuals, with vLLM/SGLang/Bedrock/Vertex preferred for serious deployments; others say Ollama has benchmarked competitively and can be used in production in constrained environments.
  • Frustration that features like sharded GGUF and Vulkan support lag, with an old Vulkan PR cited as evidence of neglected community contributions.

Community reaction and “enshittification” fears

  • A noticeable split:
    • One camp is angry or wary, seeing a familiar pattern of VC-backed OSS turning into a locked-in, monetized platform (Docker Desktop cited).
    • Another camp defends Ollama: Turbo is optional, core remains open, and projects need revenue to survive; paying for GPUs is framed as fair.
  • Several expect more open, purely local alternatives (llama.cpp, sglang, ramalama, etc.) to benefit if Ollama drifts toward a conventional SaaS model.