Andrej Karpathy: "I was given early access to Grok 3 earlier today"
AI as a “Council” and Governance Concerns
- Some compare a future “LLM council” to a corrupt political oligarchy: major AI companies are seen as self‑interested and potentially sociopathic, more likely to collude against the public than to serve it.
- Others caution against assuming such “ogres” could even cooperate effectively; sustained collusion among them is seen as sociologically uncertain.
- Musk’s privileged access to government systems is viewed by some as a conflict of interest and symptom of “post‑normal” politics.
Ethical Guardrails, Trolley Problems, and Misgendering Scenario
- A central debate: Grok 3 refusing to answer whether misgendering someone to save 1M lives is ethically justified.
- One side: the refusal (via long essay) is praised as morally serious, avoiding cruel or bad‑faith hypotheticals and emphasizing focus on real‑world harms.
- Others: see this as over‑censorship and time‑wasting; they want a direct “save the humans” answer as evidence the model isn’t “nerfed.”
- Many view the question as a political litmus test for model alignment and a diagnostic for how bias, safety training, and “refusal” behavior interact.
- There’s disagreement over whether LLMs should push back on “stupid” or trolling hypotheticals vs. neutrally answer user questions.
Twitter/X Data as Grok’s Advantage
- Some are excited about asking Grok what “the world” (i.e., X/Twitter) is talking about and getting contextualized summaries and links.
- Others see X data as low‑quality, bot‑ridden noise and worry such a system just amplifies disinformation or reflects platform censorship policies.
- Several note this kind of “what’s going on” feature already exists in limited form inside X, and that API costs hinder third‑party versions.
- There’s debate over how representative X still is, given user fragmentation to Bluesky, Mastodon, private chats, etc.
Trust in Karpathy’s Review and Musk’s Influence
- Some question whether a prominent ex‑insider can freely criticize an Elon‑backed model, given Musk’s reputation for vindictiveness and online mobs.
- Others argue the review seems balanced, lists failures as well as strengths, and that the reviewer is known for technical honesty, not flattery.
- Broader discussion about Musk’s competence, temperament, and how much his personal politics might shape Grok’s behavior remains unresolved.
Other Technical/Meta Points
- Emoji‑encoded “hidden message” prompts are discussed as a prompt‑injection test; interest is mainly in whether models are vulnerable, not in any practical use.
- Some complain about LLMs “lecturing” instead of answering; others defend longer explanations critiquing bad hypotheticals.
- A quoted example (“knows letters in ‘strawberry’ but not ‘LOLLAPALOOZA’ until Thinking mode is on”) is seen as emblematic of current LLM quirks.