GPT‑5.5 Bio Bug Bounty

Prize structure & incentives

  • Top prize is $25k for the first “true universal jailbreak” that answers five hidden bio questions without triggering moderation.
  • Many see this as a lottery: only one main payout regardless of how many people or distinct bugs succeed; partial successes may get nothing or only discretionary rewards.
  • Several posters call the reward insultingly low relative to OpenAI’s resources and the claimed existential stakes, comparing it unfavorably to six‑figure security bounties and OpenAI’s own prior $500k Kaggle contest.
  • Others respond that “first past the post” and discretionary partials are standard in bug bounties and contests.

NDA, access control, and secrecy

  • Participation is gated: applicants must already be ChatGPT users, be “vetted” bio red‑teamers, and sign an NDA.
  • Critics argue this turns it into unpaid or underpaid spec work where almost everyone gets nothing and also cannot publish results or even the questions.
  • Some worry the NDA allows OpenAI to reject payouts while still silencing participants. A few say this level of confidentiality is normal; others strongly disagree.
  • There is confusion/criticism around being asked to propose a jailbreak approach before even seeing the five questions.

Perceived goals: safety vs. marketing

  • Many commenters describe the program as a PR or “theatre” move:
    • To signal that models are extremely powerful and potentially dangerous.
    • To reassure regulators that OpenAI is responsibly self‑policing.
    • To contrast with and potentially stigmatize open‑source models.
  • Some think the real aim is to collect jailbreak attempts as training data for future safety systems and marketing claims (“safest model”).
  • A minority see value in a narrowly scoped, concrete biosafety red‑teaming effort.

Biorisk framing and model behavior

  • “Bio‑bugs” are described as ways to get the model to provide actionable guidance on harmful biological activities (e.g., weaponization steps), as opposed to high‑level or benign information.
  • Other AI companies’ CBRN/biorisk filters are mentioned as precedent.
  • Some users report current models already over‑blocking benign biology‑related tasks (e.g., sequence analysis, educational illustrations), calling it frustrating false positives.
  • One commenter notes that, despite flaws, over‑blocking is preferable to dangerous false negatives.

Trust in bug bounties

  • Several recount experiences of companies (including OpenAI) declaring impactful issues out‑of‑scope to avoid paying.
  • This fuels broader skepticism that corporate bug bounty programs are fair or trustworthy.