Why DeepSeek had to be open source

Security, Local Use & Practicalities

  • Several commenters verify that DeepSeek R1 (or its distillations) runs with no external network traffic and can be used fully offline.
  • Concerns remain about supply-chain risk (e.g., verifying safe serialization formats such as safetensors vs unsafe ones like pickle).
  • Full R1 is described as ~650–700GB (fp16) with quantizations around 150GB; only distilled models (based on Llama/Qwen trained on R1 outputs) are practical on single GPUs and consumer hardware.

Is DeepSeek “Open Source”?

  • Large subthread argues DeepSeek is not open source but “open weights” or freeware:
    • Missing: training code, training data, and low-level PTX/cluster tooling.
    • Weights are likened to binaries or bytecode: modifiable via fine‑tuning, but not reconstructible from source.
  • Others counter that for LLMs, weights are the “preferred form for modification,” citing GPL/OSI language and practical constraints (no one can afford to retrain frontier models).
  • Nuanced taxonomy is proposed:
    • open-source inference code
    • open weights
    • open pretraining recipe (code + data)
    • open fine‑tuning recipe (code + data)
  • Licensing nuances: older DeepSeek-V3 weights have a custom, more restrictive license; R1 and R1-Zero weights are MIT-licensed.

Trust, Censorship & Geopolitics

  • Strong skepticism toward a Chinese API; open weights and local deployment are seen as crucial for Western adoption.
  • Evidence cited that DeepSeek censors topics sensitive to the Chinese state; similar concerns are raised about Western models on other geopolitical topics.
  • Some suggest using Chinese and Western models to cross-check each other; others argue both sides propagate their own narratives.

Competition, Moats & Economics

  • Debate over whether DeepSeek “dethrones” OpenAI or just narrows the gap temporarily.
  • Some see DeepSeek as proof that frontier-level reasoning can be built for a few million dollars, eroding proprietary moats and pushing prices toward zero.
  • Others argue large incumbents (especially Google) still have substantial moats: custom hardware, data pipelines, userbase, and monetization channels.
  • Expectation from many participants: a mixed future where open(-weights) models commoditize baseline capabilities, while proprietary models continue at the bleeding edge and for integrated commercial offerings.

Reaction to the Article Itself

  • Multiple commenters call the post clickbait/content marketing for Lago and object to the claim that DeepSeek “proves” an open-source future; they see it as one data point, not a proof.