We hacked Gemini's Python sandbox and leaked its source code (at least some)
Scope of the “Hack” and Title Controversy
- Many commenters argue the title (“hacked Gemini and leaked its source”) is misleading or clickbait.
- They stress this was about the Python sandbox infrastructure, not the Gemini model or its training data.
- Some say running
stringson a binary and exploring a container is routine reverse‑engineering, not a major “hack.”
What Was Actually Exposed
- The main “leak” was internal protobuf definitions bundled into the sandbox binary by an automated build step.
- Debate on sensitivity:
- Some say proto definitions are like a schema and not inherently secret, with similar files already leaked years ago.
- Others note these particular protos touch internal authn/authz and data-classification systems, so their structure could aid attackers or reveal architecture.
- No model weights, training corpus, or broader internal systems were accessed.
Sandbox Architecture and Creation
- The sandbox runs in gVisor; Google engineer confirms they use checkpoint/restore plus a CoW overlay filesystem for very fast startup.
- Commenters compare this to alternative approaches (ZFS or LVM snapshots, unikernels), discussing copy‑on‑write performance and caching benefits.
- The same engineer says the sandbox is general-purpose for running untrusted code (data analysis, extensions), not just a one-off feature.
Security Posture and Significance
- Several people view this as a minor but valid issue that mainly exposes a gap in security review and build automation.
- Others argue the incident shows Google’s overall robustness: the sandbox largely did what it should, and the work was done in collaboration with Google’s security team.
Prompt Injection and Agent Security
- One subthread uses this as a springboard to discuss how local/agentic AIs will face prompt-injection risks when browsing the web.
- Comparison is made to humans getting “mind‑viruses” from internet content; concern that future personal agents could be subverted the same way.
Gemini, Assistant, and Product Perception
- Long side discussion about Gemini replacing Assistant:
- Some users report Gemini can’t reliably set timers, play music, or integrate with device apps; others say it works fine for them.
- Complaints about declining Google UX, “overhyped” AI, and underwhelming product execution despite strong research.
- A Googler describes internal mood as a mix of frustration over slow launches, excitement about strong models, and indifference from those who see LLMs as overhyped.
- Several commenters claim Gemini models (e.g., Flash, 2.5 Pro, Gemma) are highly capable and cost-effective for developers, despite weaker consumer perception.
Documentation, Transparency, and Developer Experience
- Parallel is drawn to scraping ChatGPT Code Interpreter’s environment to discover available packages; people lament that such basic capability lists aren’t officially documented.
- One Googler says they’ll raise the idea internally, reinforcing that missing documentation is more likely neglect than deliberate secrecy.