Show HN: A MitM proxy to see what your LLM tools are sending

Need for LLM observability & governance

  • Strong interest in seeing exactly what coding agents and CLIs send to providers, especially Claude Code, Codex, Gemini, etc.
  • Several commenters note a surprising lack of enterprise-grade tools for this, given past norms around strict data governance.
  • People expect a “pendulum swing back” toward better tracking, auditing, and governance of agentic AI use.

Use cases and perceived benefits

  • Debugging token waste: identifying excessive tool calls, verbose responses, large file reads, and inflated context windows.
  • Improving prompts and system instructions for specific projects or repositories.
  • Storing full traces (markdown/JSON) for later querying, long‑term memory, and postmortems on hallucination-induced bugs.
  • Potentially tying traces to code commits for forensic debugging.

Implementation approaches and alternatives

  • Original tool is essentially a wrapper around mitmproxy with a convenience CLI; later refactored toward an HTTP relay.
  • Some prefer direct instrumentation or using LLM clients’ configurable BASE_URL/HTTP proxy to avoid full MitM.
  • Others mention existing or custom solutions: Envoy-based proxies, LiteLLM + Langfuse, mac apps, OpenTelemetry pipelines, and direct patching of open-source CLIs like Gemini.

Security and “vibe-coded” software concerns

  • A serious issue is highlighted: the initial version disabled TLS verification (ssl_insecure=true), creating a large attack surface (e.g., DNS-based MitM, potential RCE).
  • Multiple commenters warn people not to use that version and question the author’s security understanding and overall trustworthiness.
  • This triggers a broader critique of “vibe-coded” / AI-generated projects presented as production-ready, where authors don’t fully grasp the implications.
  • Some push for more honesty (“I prompted this” vs “I built this”) so users can calibrate their trust.

Ideas for extensions

  • Export to OpenTelemetry-compatible systems (e.g., Phoenix, Logfire), with auth support and simple --otel-endpoint-style configuration.
  • Using the proxy to sanitize or block sensitive data, or to inject credentials safely from outside the agent sandbox.
  • Dynamic context optimization: smarter selection of what enters the context window, possibly using the logs themselves as long-term memory.

Meta discussion about AI tools & HN

  • Mixed feelings: excitement over rapid prototyping and richer computing experiences versus concern about security fallout and rising “AI slop.”
  • Some worry HN is increasingly filled with low‑quality AI-generated projects and even AI-written comments, reducing stars/READMEs as quality signals.