GPT-5 leaked system prompt?
Formatting, emphasis, and prompt structure
- People notice the prompt’s use of markdown bold instead of ALL CAPS; some speculate caps might be treated as “yelling” or be tokenized differently, possibly changing model behavior.
- The length and redundancy of instructions (e.g., “never write JSON” for
to=bio) are seen as evidence that OpenAI also struggles with prompt adherence and has to layer on “hacky patches.”
Repetition, negation, and control over behavior
- Several commenters report that LLMs routinely ignore “don’t do X” instructions (e.g., no dashes, no trailing whitespace, no emojis), especially over longer sessions.
- Some have more success phrasing constraints positively, others argue “affirmative prompting” is overrated and negation is fundamentally hard for autocomplete-style models.
- A recurring observation: instructions like “don’t output JSON” or “don’t think of an elephant” may actually increase the salience of the forbidden thing.
Tools, code, and UX biases
- The detailed sections on Python and React are read as configuration for internal tools: Python for analysis/plots, React + Tailwind + shadcn for live previews in the UI.
- This is seen as both practical (optimize common use cases) and slightly dystopian: LLM defaults could further entrench specific stacks (React/Tailwind) in the ecosystem.
Authenticity and prompt-leak skepticism
- Many doubt the leak: missing safety sections (e.g., porn, CSAM), obvious mistakes (Japanese labeled as Korean), and generic tone.
- Others argue repeatable extraction patterns, behavioral matches (e.g., song-lyric refusal), and tool-specific snippets like
guardian_tool.get_policy(election_voting)suggest at least partial authenticity. - There’s discussion of deliberate “fake” or decoy system prompts and the difficulty of ever verifying truth when the only witness is the model itself.
Safety, copyright, memory, and censorship
- Song lyrics get special treatment; some infer legal pressure and note that the model even refuses public-domain anthems.
- The
bio/memory tool raises mild privacy concerns, but reported stored facts tend to be banal rather than deeply personal. - Several users feel GPT‑5 is more censored, blander, and less willing to generate stylized violent or edgy fiction, which some see as necessary safety and others as artistic degradation.
Meta: system prompts vs training
- Commenters are struck that “programming” the model is done via huge natural-language prompts instead of deeper training or prompt-tuned embeddings.
- There’s debate over whether long, static prompts are a crude stopgap or a pragmatic, easily updatable control layer atop expensive base models.