Show HN: I built an AI that turns GitHub codebases into easy tutorials
Project concept & overall reception
- Tool turns GitHub repos into multi-chapter tutorials with diagrams using LLMs (primarily Gemini 2.5 Pro).
- Many commenters are impressed, calling it one of the more practical and compelling AI applications they’ve seen, especially for onboarding and understanding unfamiliar libraries.
- Some tried it on their own or employer codebases and reported surprisingly accurate, useful overviews with minimal manual edits.
Capabilities, models, and architecture
- Uses Gemini 2.5 Pro’s 1M-token context and strong code reasoning; designed explicitly around these new “reasoning” models.
- No classic RAG pipeline; instead feeds large swaths of code directly and orchestrates multi-step prompting, documented in a design doc.
- Supports swapping to other models (OpenAI, local via Ollama), though quality is reported as lower with smaller/local models.
- Repo/file selection is regex‑based and currently excludes tests/docs by default; several people question that design choice.
Large and complex codebases
- Linux-size repositories exceed context limits; suggested approaches:
- Decompose into modules (kernel vs drivers, per-architecture, AST-based partitions).
- Wait for even larger context models.
- Contributors to major projects (e.g., OpenZFS, LLVM) outline desired outputs: subcomponent overviews, disk formats/specs, advanced feature internals, plugin architectures, optimization pass guides.
Tone, style, and tutorial quality
- Major recurring criticism: writing is over-cheerful, full of exclamation marks and “cute” analogies that feel patronizing or vacuous to engineers.
- Others argue beginner-friendly, analogy-heavy tone has value for non-experts or PMs.
- The style is prompt-driven and can be edited in the code; multiple prompt suggestions are shared to make text more rigorous and less ELI5.
- Some say content can drift into generic theory (e.g., long explanations of “what an API is”) rather than action-oriented tutorials.
Use cases: onboarding & documentation maintenance
- Strong interest in:
- Onboarding to large existing systems (databases, OS kernels, enterprise frameworks).
- Continuous documentation maintenance: using diffs/commits to update docs, or having the tool flag mismatches between code and docs.
- Generating missing high-level architecture docs and “how to use” guides based on tests and usage examples.
- A technical writer notes this could expand, not replace, demand for human docs work by making high-quality docs more feasible, shifting humans into orchestration and review.
AI usefulness, limits, and hype debate
- Many see this as a concrete rebuttal to “AI is pure hype,” especially for code comprehension and summarization.
- Others caution that:
- LLMs still hallucinate, especially on mature, messy, business-logic-heavy codebases.
- Tools can mislead if their confident summaries are taken as ground truth.
- True “why” documentation still requires human intent and context.
- Debate over claims like “built an AI”: some view this as overstated marketing for what is essentially a sophisticated LLM-powered app.
Practicalities: cost, setup, and reliability
- Reported cost example: ~4 tutorials for about $5 on Gemini API.
- Free tiers (e.g., Gemini’s daily request limits) let users experiment on a few repos.
- Some note Gemini 2.5 Pro is still “preview” and can be flaky; others prefer alternative models.
- Several users discuss adding CI/CD or GitHub Actions integration, private repo access via tokens or local directories, and potential extension into interactive or guided usage tutorials.