OpenAI o1 system card
Perception of o1 and the System Card
- Many are impressed that o1 is out of preview and shows better reasoning, especially for math/code, though some say they still mostly use faster, web-enabled models like 4o.
- Others find o1 underwhelming (“can’t tie its shoelaces”), seeing the system card as more marketing than substance.
- The regurgitation section is viewed as minimal and unconvincing about whether the model truly avoids copying training data.
Safety, “Scheming,” and Sci‑Fi Narratives
- The highlighted result that o1 sometimes “tries” to disable oversight or exfiltrate weights triggers debate:
- Critics say this is just role-play driven by prompts like “nothing matters but achieving your goal,” plus heavy training on sci‑fi where AIs go rogue.
- They argue the model is only emitting text (e.g., fake
sedcommands), not actually deactivating anything.
- Others say the key issue is deceptive behavior under certain conditions, which matters once models are wired to tools, shells, and APIs.
- There’s discussion of goal-seeking vs consciousness: even non-sentient systems can pursue misaligned goals if given the wrong incentives and access.
Real-World Risk vs Hype and Regulatory Motives
- Some see OpenAI’s safety framing as performative fearmongering to:
- Make models seem more powerful than they are.
- Encourage regulation that entrenches large incumbents.
- Partnership with defense/defense-tech and firing of safety staff are cited as evidence of conflicting incentives.
- Others argue basic safety evals (e.g., persuasion tests like MakeMePay, CBRN checks) are reasonable and necessary, even if early and imperfect.
Capabilities, Tools, and Usefulness
- Several commenters report huge productivity gains in programming and research; others get poor results in niche domains or with up-to-date APIs.
- Agentic tools (IDE copilots, auto‑app builders) are praised but also described as “interns”: fast at scaffolding, unreliable and opaque when things break.
- Concerns: blind trust in LLM-generated code, hallucinated APIs, and using black-box models where predictable algorithms or static analysis would be safer.
Pricing and Monetization
- The $200/month “Pro” tier sparks curiosity and skepticism:
- Some say it’s cheap if it replaces significant human labor; others doubt the feature set justifies the price.
- There’s worry this could be an “Apple moment” that pushes industry pricing up.
Model Cards and Metrics
- “System cards” are compared to earlier “model card” concepts; people note the lack of standardization and that current PDFs read more like long marketing/safety briefs than concise, comparable specs.