Logging sucks

Single-purpose site and marketing angle

  • Some readers like the interactive, “modern” presentation; others dislike the standalone domain and worry it will disappear unlike a stable personal blog.
  • Several conclude it functions as content marketing / lead generation: the “personalized report” form and tie-in to an observability SaaS (and indirectly Cloudflare) are noted.
  • A few argue this is fine as long as the content is genuinely useful.

Wide events, structured logging, and observability

  • Many see the core idea as “structured logs with rich context per request,” often already practiced with correlation/request IDs and JSON.
  • Supporters say wide events make it easy to answer product and incident questions: who is impacted, which flows, which customers, which experiments, etc.
  • Others argue this is not new; similar ideas exist in observability tools, structured logging libraries, tracing systems, and event sourcing.
  • Some emphasize schema discipline (e.g., standard field names, “related.*” fields) to avoid chaos in wide logs.

System architecture and “15 services per request”

  • The claim that a single request may touch many services sparks a microservices vs monolith debate.
  • One camp calls such architectures “deeply broken” and driven by fashion or incentives rather than need; monoliths are said to be sufficient for most apps.
  • Another camp provides detailed real-world examples (e‑g., mobility/bike‑sharing) where many distinct concerns and teams justify multiple services, or at least clear internal boundaries.
  • Discussion highlights that service boundaries are a trade-off: flexibility, independent deployment and scaling vs added latency, complexity, and logging/tracing difficulty.

Logs vs metrics, traces, and audits

  • Several stress that logs, metrics, traces, and audits serve different purposes, mainly differentiated by how you handle loss of fidelity.
  • Others propose a “grand theory of observability”: treat all signals as events that can be transformed into each other, with different storage and durability policies.
  • Large subthread debates audit logging: some insist audit streams must be transactionally durable across fault domains; others say regulators only require “reasonable” durability and availability.

Performance, sampling, and logging infrastructure

  • The recommendation to log every error/slow request while sampling healthy ones is criticized: in degraded states, log volume spikes can overload systems.
  • Counter-argument: a production service should be designed so log ingestion scales, with buffering, backpressure, and best-effort dropping where acceptable.
  • There’s debate on realistic throughput: some claim properly designed systems can handle enormous event rates; others note typical end‑to‑end stacks fall over at much lower volumes.
  • Adaptive and bucketed sampling strategies, or sampling primarily at the log backend, are mentioned as practical mitigations.

Tooling and implementation practices

  • Many say correlation/request IDs plus structured logs (often JSON) and tools like Splunk, Kibana, Loki, Tempo, or ClickHouse already solve most issues.
  • Several argue OpenTelemetry + tracing plus “wide spans” effectively implement what the article describes, but contest the article’s criticism of OTLP.
  • Suggestions include: consistent schemas, user IDs on every log, separate “audit” message types, log-based events feeding metrics systems, and using logs as an explicit product for internal consumers.

Critiques of the article (tone, examples, AI)

  • Some praise the substance and interactivity but find the “logging sucks” framing overdramatic, strawman-y, or condescending.
  • Multiple commenters feel the prose is verbose and “AI‑ish,” though there’s disagreement on this; others think calling out AI assistance is becoming unhelpful.
  • The initial bad‑log example is seen as unrealistic by some (because many already use correlation IDs), while others say they still see logs that bad in practice.

Miscellaneous insights

  • Concerns raised about logging sensitive or business-critical fields (like lifetime value) vs their usefulness for prioritization.
  • Several emphasize that logging, tracing, and metrics must be intentionally designed; tools alone don’t fix poor conventions.
  • A recurring theme: monolith vs microservices and “wide events vs traditional logs” are both trade‑offs; success depends more on discipline, consistency, and clear goals than on any single pattern.