Hacker News, Distilled

AI powered summaries for selected HN discussions.

Page 347 of 535

Mermaid: Generation of diagrams like flowcharts or sequence diagrams from text

Landscape of text-to-diagram tools

  • A curated list of ~70 browser-based text-to-diagram tools is shared; readers find it surprisingly comprehensive and valuable.
  • Many specialized tools (e.g., for sequence diagrams, database diagrams, genealogical trees) are viewed as better for their niche than generic tools like Mermaid.
  • Alternatives frequently mentioned:
    • Sequence diagrams: WebSequenceDiagrams, js-sequence-diagrams.
    • DB diagrams: DrawDB, dbdiagram.io, Cacoo, sqliteviz, Graphviz-based tools.
    • General drawing/whiteboarding: Excalidraw, Miro.
    • Other text-based diagrammers: PlantUML, Graphviz/dot, D2, Kroki as a wrapper for many syntaxes.

Mermaid’s main strengths

  • Native/inline support in GitHub, GitLab, Notion, Obsidian, Hugo, Jira, Azure DevOps, etc., makes it a de facto choice for diagrams in Markdown and internal docs.
  • Diagrams-as-code fit naturally into repos: editable, diffable, and compatible with git blame and review workflows.
  • Works offline via CLI and editor plugins (JetBrains, VS Code) despite being browser-focused.
  • A near-WYSIWYG editor (mermaidchart.com) eases layout while preserving text-source.

Critiques and limitations

  • Perceived as less powerful and less polished than PlantUML, Graphviz, or specialized tools; syntax is seen as strict and somewhat immature.
  • Local rendering can be awkward (e.g., headless Chrome flows, CLI SVG text issues).
  • Layout struggles with large or complex graphs (schemas with many tables, microservices, etc.), a problem shared with Graphviz.
  • In Notion and some ecosystems, shipped Mermaid versions are outdated.

LLMs and Mermaid

  • Many report strong synergy: LLMs can generate or refine Mermaid from:
    • High-level text descriptions.
    • Codebases or logs.
    • Hand-drawn diagrams (via multimodal models).
  • Some say certain models don’t handle Mermaid well and prefer LaTeX TikZ; others report newer models (including open ones) handle Mermaid reliably.

Use cases and philosophy of diagrams

  • Common uses: system architectures, sequence diagrams, build pipelines, database schemas, story/character relationships, internal engineering docs.
  • Some participants see diagrams as high-value for shared understanding; others argue most diagrams are “write-only,” produced mainly to satisfy process requirements and rarely consulted later.
  • There is skepticism about heavy diagramming cultures (e.g., legacy UML tooling), contrasted with appreciation for lightweight, quickly generated diagrams—especially when LLMs cut creation time to minutes.

On File Formats

Streamability, indexing, and updates

  • Several comments stress making formats streamable or at least efficient over remote/seekable I/O.
  • Strong debate about where to place indexes/TOCs:
    • Index-at-end favors append, in‑place updates, concatenation, large archives, and workflows like PDFs where small edits just append data.
    • Index-at-start favors non-seekable streams and immediate discovery of contents.
    • Some suggest hybrid or linked index structures; others note “it’s just a tradeoff, not one right answer.”
  • Many real‑world workflows recreate files rather than update in place, but formats supporting cheap updates still bring UX and performance wins.

Compression and performance tradeoffs

  • Compression is “probably desired” for large data, but algorithm and level should match use: high effort only pays off for frequently copied/decompressed data.
  • General vs domain-specific compression is noted; specialized schemes may outperform generic ones in narrow domains.

Chunking, partial parsing, and versioning

  • Chunked/binary formats are praised for incremental/partial parsing and robustness, but commenters warn chunking alone doesn’t guarantee reorderability or backward/forward compatibility; explicit versioning is essential.
  • DER/ASN.1 is cited as an example of structured, partially skippable binary encoding; others find ASN.1 overkill for most custom formats.

Using existing containers (ZIP, SQLite, etc.)

  • Strong encouragement to reuse existing containers (ZIP, tar, sBOX, CBOR tags, HDF5) instead of inventing from scratch.
  • ZIP as a multipurpose container is praised; many complex formats (Office, APK, EPUB, etc.) already use it.
  • SQLite as a file format/container splits opinion:
    • Pro: great for composite/stateful data, metadata, queries, incremental updates, encryption extensions; multiple real projects use it successfully.
    • Con: overhead, complexity, blob limits, nontrivial format, possibly inferior to ZIP for simple archives or large monolithic blobs.

Human-readable vs binary, numbers and floats

  • Consensus that human-readable formats should be extremely simple; otherwise binary is safer and clearer.
  • Textual numbers, especially floats, are called tricky to parse/round-trip correctly; binary IEEE754 with fixed endianness is seen as easier and less error-prone.
  • Ideas like hex floats or editor support for visualizing binary floats appear, but trade off readability or complexity.

Directories vs single files, diffability, and tooling

  • Some advocate directory-based “formats” (structured folders, or unzipped equivalents of ZIP-based formats) for better version control, experimentation, and debugging; ZIP can then be an export format.
  • Others note that dumping runtime data (pickle, raw object graphs, SQLite snapshots) is convenient but harms portability and can enlarge attack surface; deserializers must be strictly bounded by a spec.

File extensions and type detection

  • Suggestion: long, app-specific extensions (e.g., .mustachemingle) to minimize collisions.
  • Counterpoints: Windows hides “known” extensions; Linux often relies on MIME/magic; long extensions can hurt UX (truncation, typing).
  • Agreement that clear, specific extensions like .sqlite are still useful; distinction between generic shared formats and app-specific ones is highlighted.

Design pitfalls and backward compatibility

  • Warnings against over-clever bit-packing (splitting flags across nibbles/bytes) that later prevents extension. Real examples show such schemes becoming brittle.
  • Concern that some parsers ignore documented flexibility (e.g., header growth with offset fields) and hard-code assumptions, breaking future versions.
  • One view holds that human-editable formats can tempt developers to skip proper UI support, degrading usability.
  • Emphasis on documenting formats thoroughly; good specs and tables clarify intent more than code/flowcharts alone.

The WinRAR approach

WinRAR’s Business Model and Revenue

  • Most revenue reportedly comes from corporate licenses; consumers largely use it unpaid.
  • Public filings for the German company behind WinRAR suggest on the order of ~€1M earnings in 2023.
  • Several commenters note their companies bought WinRAR licenses and still rely on it for “mission‑critical” workflows.
  • Others argue the real “WinRAR approach” is less about goodwill and more about making license‑compliance‑driven organizations pay while everyone else uses it freely.
  • The brand now leans into meme status with community management and merchandise.

Licensing, Compliance, and Everyday Violations

  • Many companies are said to violate the 30‑day trial terms, just leaving WinRAR (and other paid tools) in perpetual use.
  • Some workplaces strictly prohibit unapproved third‑party software and would treat such violations seriously; others are lax or ignorant.
  • This triggers broader discussion about people not caring about licensing unless they personally bear risk or cost.

Why Use WinRAR/RAR vs 7-Zip or Others?

  • Some are puzzled why anyone pays for WinRAR when 7‑Zip is free, open source, and uses LZMA with strong compression.
  • Defenders cite:
    • Better Windows integration and ergonomics.
    • Rich archive features: recovery records/parity, good CLI, handling of archive flags, NTFS streams, ACLs, hard links, and built‑in Blake2 hashes.
    • Stable, non-“enshittified” UI and long history.
  • Benchmarks shared: 7‑Zip can compress ~6% smaller but much slower at extreme settings; for most people, convenience beats a small compression gain.
  • Some say they rarely see .rar now; others point out large legacy archives and “scene” rules that historically standardized on RAR (multi‑part archives, floppies, unreliable connections).

Piracy, Culture, and Shareware Patterns

  • Multiple comments describe 80s–90s Eastern and Western Europe (and elsewhere) as heavily pirated ecosystems, including businesses and governments.
  • Piracy is framed as both economic necessity and a growth hack (e.g., Microsoft tolerating it early to build dominance).
  • WinRAR’s permissive trial is compared to classic shareware: get ubiquitous at home, monetize businesses later.
  • Some now consciously pay for tools (licenses, donations, books) as a reaction against that culture and against today’s subscription/DRM backlash.

Nagware vs Goodwill and Related Models

  • Disagreement over whether WinRAR “runs on goodwill”:
    • One side: it’s essentially nagware; you pay to stop the startup dialog.
    • Other side: the nag is mild (hit Escape) and functionally it’s unlimited, which feels generous.
  • Similar “soft paywall” or generous-trial models are cited:
    • Paint Shop Pro, Sublime Text, Reaper, Renoise, Forklift, ShareX, KeePassXC, and others.
    • Immich’s model (fully usable, optional license) is praised as especially user‑friendly and aligned with open source.
  • Many commenters say these approaches make them more willing to pay, especially once they’re no longer broke.

Alternatives and Cross‑Platform Notes

  • On macOS, users missing WinRAR/Total Commander mention:
    • BetterZip for archive browsing and Quick Look integration.
    • Commander One, Marta, Transmit, and Double Commander as dual‑pane/file‑manager replacements.
    • Some still run WinRAR/Total Commander under Wine on Linux/macOS.

Instagram Addiction

Perceived Harms and Compulsive Use

  • Many describe Reels/Shorts/TikTok as “instant addiction,” especially the endless-scroll, autoplay, algorithmic feed that exploits visuals and quantified status (followers, likes).
  • Some share concrete episodes of losing hours to shorts, or watching children already glued to Reels, even while drinking from a bottle.
  • Several comment that heavy social media use leaves them feeling worse afterward, not better, despite occasional amusement or dopamine from likes.

Is It Really “Addiction”?

  • One group objects to calling Instagram use an “addiction,” arguing it misuses medical terminology and risks over-medicalizing behavior.
  • Others counter with behavioral addiction concepts (e.g., gambling), and offer criteria like unwanted, frequent, hard-to-control use that displaces more important activities.
  • There’s mention of common reward pathways between substance and behavioral addictions, with GLP‑1 drugs cited as affecting both.

Capitalism, Design Ethics, and Regulation

  • Strong frustration at platforms (Meta, Google, TikTok, Reddit, etc.) for making addictive UX (shorts, reels, autoplay, pushy notifications) hard or impossible to disable.
  • Debate over whether this is “capitalism’s evil algorithm” or just human greed manifesting under any system.
  • Some want a “digital consumer bill of rights”: disabling autoplay, algorithmic feeds, fake notifications; enforcing chronological feeds; EU-style regulation is predicted by some.
  • Others defend site owners’ freedom to design as they wish; users should vote with their feet or use browser-side modifications.

Coping Strategies and Tools

  • Common tactics: deleting apps, using only desktop versions, disabling notifications, DNS blocking, using RSS instead of in-app subscriptions, or time-based locks (e.g., systemd service, MDM-based phone blocker).
  • Numerous technical tools are mentioned: browser extensions (Unhook, user scripts), privacy-friendly mobile clients (NewPipe, Freetube), patched/“distraction-free” Instagram builds, and DNS or Pi‑hole–style blocking.
  • Non-technical habits: reading books with the phone in another room, journaling, phoneless walks, physical notebooks, and daily to-do lists before “guilt-free scrolling.”

Boredom, Attention, and Long-Term Effects

  • Several emphasize reclaiming boredom (walking, commuting, quiet time) as crucial for creativity and focus, contrasting it with constant feed consumption.
  • Concern that ubiquitous short-form content is degrading attention spans, making reading difficult and potentially harming work, learning, and mental health over decades.

Individual Differences and Article Style

  • Some say they’re largely immune or even repelled by shorts, suggesting the algorithms simply haven’t “hit” their interests.
  • Others describe profound struggles and shame around compulsive scrolling.
  • Multiple commenters found the article’s all-lowercase style off-putting or harder to read, comparing it to the unstructured flow of an infinite feed.

Google AI Ultra

Pricing & Perceived Value

  • Many feel $250/month is “nuts” or “sticker shock,” especially with multiple AI subs (OpenAI, Claude, Perplexity, etc.) already in the mix.
  • Others argue it’s equivalent to under an hour of senior dev/consultant time and easily justifiable if it saves even a few hours/month, especially for businesses.
  • Some see it as an unsustainable, loss-leading price tier to rate‑limit usage rather than a real market-clearing price.

Bundling, Plan Design & Confusion

  • Strong criticism of the bundle: users want higher LLM/coding limits and Deep Think, but not 30TB storage, image/video gen, or YouTube Premium.
  • Several see YouTube Premium in a “work” plan as odd, especially since it’s not a family plan. Comparisons to “Comcast triple play” and cross‑sell lock‑in.
  • Confusion over upgrade paths from existing Google One/YouTube subs and unclear differentiation between AI Pro vs Ultra on the marketing pages.

Economics & Sustainability of AI

  • Repeated reminders that LLM inference is expensive and current prices across the industry are heavily subsidized.
  • Debate over whether competition will bring prices down vs. a “money spigot” eventually drying up, causing prices to spike and progress to slow.
  • Some compare to past specialist software becoming cheaper over time; others think SOTA AI may remain expensive due to training and GPU costs.

Competition, Model Churn & Multi‑Provider Abstractions

  • Users hesitate to commit because SOTA leadership keeps shifting (OpenAI ⇄ Gemini ⇄ Anthropic, etc.), encouraging “subscribe for a month, then switch.”
  • Suggestions for middlemen/wrappers (LiteLLM, OpenRouter, others) that unify multiple providers and abstract away subscription switching and APIs.
  • Concern that top reasoning modes (e.g., “Deep Think”) being paywalled or restricted will fragment features between UI and API.

Enterprise vs Individual Adoption

  • Split views: some say the target is corporations and self‑employed devs; others note many companies still ban sending proprietary data to public LLMs regardless of ToS.
  • Expectation that enterprises will negotiate bulk pricing and that AI “seats” will roll out first to high‑value roles.
  • Concern that every SaaS vendor is separately upselling AI add‑ons, making overall AI licensing hard to justify at scale.

Privacy, Trust & Data Use

  • Deep distrust of Google’s data practices; some refuse to use it at all, preferring OpenAI/Anthropic or smaller “privacy‑oriented” services.
  • Desire for a paid tier that guarantees no training on chats while still keeping history and integrations—current “no training” modes remove too much functionality.
  • Skepticism that ToS promises are enforceable or verifiable; references to past big‑tech privacy misrepresentations.

Access, Inequality & Lock‑In

  • Concern that $200–250/month tiers will be accessible mainly to rich countries and corporations, exacerbating global inequality in productivity.
  • Others argue tech historically trickles down and will get cheaper; critics note AI’s massive fixed costs and potential for a permanent elite tier.
  • 30TB storage is seen as a “soft lock‑in”: once filled, users are effectively stuck.

Product Quality & Missing Pieces

  • Some early testers describe components (e.g. Firebase Studio, media generation, audio) as poor or unreliable.
  • Frustration that Google’s touted models don’t deliver obvious productivity wins in core tools (e.g., Gemini in Sheets can’t reliably help with formulas or actions).
  • Strong demand for better coding tools (CLI/agent akin to Claude Code or Codex) and broader, cheaper access to Deep Think and Flow, ideally via API.

Ads vs Subscriptions & “Enshittification” Fears

  • Discussion that Google is shifting from pure ad model toward subscriptions, but many expect a future of “subscriptions plus ads” rather than one replacing the other.
  • Worry that current generous/free access is “pre‑enshittification,” after which limits will tighten and quality may degrade to upsell higher tiers.

Overall Sentiment

  • Mixed but leaning negative: excitement about Gemini 2.5’s capabilities and Flow/Deep Think is outweighed by pricing shock, bundling frustration, distrust of Google, and fatigue with rapid, paywalled AI product churn.

Red Programming Language

Heritage and Appeal

  • Seen as a spiritual successor to REBOL with a strong pedigree; this history keeps some people interested even when the language itself feels “magical” or scary from an engineering standpoint.
  • Several commenters used REBOL for small automation/web-scraping tasks and liked it, but wouldn’t choose it for large systems today.
  • Some still write most personal projects in REBOL; others are curious about Red but haven’t migrated.

Syntax, Semantics, and Mental Model

  • Many find Red/REBOL uniquely hard to internalize; code can be hard to read without knowing how arguments are grouped.
  • Others explain it as: data “blocks” like Lisp/Logo quotations; Polish-notation–style evaluation; functions with fixed arity; almost everything is an expression; no mandatory parentheses.
  • Infix operators are a special case: they’re evaluated strictly left-to-right with higher precedence than function calls, leading to surprising expressions unless you know the rules.
  • A team member emphasizes: “everything is data until evaluated,” many literal datatypes, no keywords, and “free-ranging evaluation” (no parentheses around arguments) as core design choices.

Dialects / DSLs and parse

  • Red calls embedded DSLs “dialects”; parse (for pattern matching/grammars) and VID (GUI description) are major built‑ins.
  • parse operates on values and types, not just characters, and supports typesets (groups of related types), which some see as powerful and simpler than regexes.
  • A long subthread debates DSLs vs APIs: DSLs praised for expressiveness in domains (regex, SQL, templating, shell); criticized as brittle, poorly tooled, and hard to maintain compared to conventional APIs.

Tooling, Docs, and Website

  • Site is perceived as dated and makes it hard to find substantial, idiomatic examples; some only discover example code by digging in the GitHub repo.
  • Documentation is widely considered thin and fragmented. Commenters note promises that “proper docs” would arrive at 1.0, but the project is still at 0.6.x with similar doc quality years later.

Maturity, Performance, and Platform Support

  • Version progress is seen as very slow (0.6.4 in 2018 to 0.6.6 in 2025); some conclude the ambitious roadmap was “stratospheric.”
  • One user reports a “hello world” taking ~36 seconds to compile; the team member acknowledges compile times are slow and not yet a priority.
  • Red compiles to an internal low‑level dialect (Red/System) with direct machine code generation and cross‑compilation, but is still 32‑bit only, which is a deal‑breaker for some platforms.

Funding, Crypto, and Governance Concerns

  • The project’s flirtation with crypto/ICO funding and a blockchain dialect turned several long‑time followers away.
  • There is speculation about influence from funders and a move to China; some commenters believe this derailed the project, while others label that pure speculation.
  • A phishing‑blocklist warning on the site sparks side discussion about community security tools vs “nanny state” overreach.

Overall Sentiment

  • Technically curious language with elegant examples, powerful DSL support, and an unusual evaluation model.
  • Enthusiasts praise its conceptual depth and cross‑stack ambitions; skeptics see stalled progress, poor docs, 32‑bit limitations, and odd funding choices as reasons not to invest.

Gemma 3n preview: Mobile-first AI

Claims, Benchmarks & “What’s the Catch?”

  • Gemma 3n is pitched as a small, “effective 2–4B” on-device model with performance near large cloud models (e.g., Claude Sonnet) on Chatbot Arena.
  • Several commenters are skeptical: they argue Arena rankings increasingly reward style and “authoritative-sounding” answers rather than real problem-solving.
  • Others note that leaderboard or aggregate scores rarely predict performance on specific “hard tasks”; you still need task-specific evaluation.
  • A removed Aider coding benchmark initially suggested parity with GPT‑4-class models, but it was run at full float32 (high RAM use) and later disappeared from the model page, increasing skepticism.

Architecture, Parameters & Per-Layer Embeddings (PLE)

  • E2B/E4B models have more raw parameters than the “effective” count; the trick is parameter skipping and PLE caching so only part of the weights sit in RAM.
  • There’s confusion about what PLE actually is: the blog is vague, there’s no clear paper yet, and people speculate it’s some form of low‑dimensional token embedding fed into each layer and cached off-accelerator.
  • MatFormer is called out as a separate mechanism for elastic depth/width at inference, enabling “mix‑n‑match” submodels between E2B and E4B.
  • Unclear so far whether the architecture is straightforwardly compatible with llama.cpp and similar runtimes; licensing may also matter.

On-Device Usage & Performance

  • Multiple Android users report Gemma 3n running fully local via Google’s Edge Gallery app, with no network required after download.
  • Performance varies widely by device and accelerator:
    • Older phones can take many minutes per answer.
    • Recent high-end phones (Pixels, Galaxy Fold/Z) get several tokens per second, especially when using GPU; CPU is slower but still viable.
  • Vision works and can describe images and text in photos reasonably well, though speed and accuracy depend on hardware and image quality.
  • NPUs generally aren’t used yet; inference is via CPU/GPU (TFLite/OpenGL/OpenCL).

Capabilities, Limitations & Safety

  • Users report strong instruction-following and decent multimodal understanding for the size, but noticeably less world knowledge than big cloud models.
  • Examples of obvious logical/factual failures (e.g., size comparisons, object misclassification) show it’s far from Sonnet or Gemini quality.
  • Smaller models appear easier to jailbreak around safety filters (e.g., roleplay prompts).

Intelligence, Hype & Use Cases

  • Some are excited about “iPhone moment” implications: powerful, private, offline assistants, accessibility (for visually impaired users), inference caching, and local planning agents.
  • Others argue LLMs still resemble “smart search” or sophisticated memorization, not genuine understanding or reasoning; they expect hype to cool.
  • There’s a broader hope that OS-level shared models (Android, Chrome, iOS, Windows) will prevent every app from bundling its own huge LLM and ballooning storage use.

Veo 3 and Imagen 4, and a new tool for filmmaking called Flow

Reactions to Veo 3 / Imagen 4 Demos

  • Many find the tech leap impressive, especially synced audio+video and Flow’s UX; some Reddit examples are called “jaw-dropping.”
  • Others are underwhelmed: clips still read as “AI” (glow, smoothness, odd motion, weak physics), with uncanny-valley humans and inconsistent environments.
  • Owl/moon clip, old man, and origami piece trigger mixed reactions—impressive rendering but eerie, aggressive, or off in subtle ways.
  • Several note these are cherry‑picked seconds, not evidence that the tech can sustain long, coherent scenes.

From Clips to Full Movies / Oscars Debate

  • Optimists predict fully AI‑generated feature films (possibly Oscar‑winning) within ~5 years; skeptics think Academy politics and current model weaknesses make that unlikely.
  • Skeptics argue: writing, editing, sound design, and direction still need heavy human intervention; “entirely AI” collapses as soon as you allow that intervention.
  • Some expect AI to first win in technical or music categories (score/song) rather than Best Picture or Screenplay.

Hollywood, Economics, and Personalization

  • One camp: this “fixes” Hollywood’s problems—no expensive stars, unions, or on‑set drama; endless cheap sequels and IP recycling fit current studio incentives.
  • Counter‑camp: if anyone can generate movies at home, Hollywood’s centrality erodes; future may be ultra‑personalized content streams via platforms like YouTube/TikTok.
  • Discussion of actors licensing likenesses, eventual fully synthetic celebrities, and manufactured off‑screen drama as part of the product.

Tools, Access, and Model Quality

  • Confusion around how to actually use Veo 3 / Imagen 4; Google’s product/rollout UX seen as opaque, with country and quota restrictions.
  • Comparisons to Sora, Runway, Wan/Hunyuan: general sense that big proprietary models are ahead in video, but open tools plus composable workflows (e.g., ComfyUI, ControlNet) still matter.
  • Prompt adherence remains a major weakness; community benchmarks show Imagen 4 not clearly ahead of Imagen 3 and behind some competitors on strict specification following.

Creativity, Authorship, and Gatekeeping

  • Deep, polarized debate:
    • One side: AI is just another tool; real creativity lies in ideas, prompt craft, iteration, and editing. It democratizes filmmaking, illustration, and animation for those lacking money or training.
    • Other side: outsourcing production to opaque models trained on unconsenting artists’ work hollows out craft, floods the commons with derivative “slop,” and erodes the meaningful struggle that makes art and skill development fulfilling.
  • Some argue future “directors” of AI productions will still need strong vision and taste; others say current AI “users” are more clients than creators, with the model as de facto artist.

Jobs, Society, and Non‑Creative Work

  • Broad anxiety about job loss in VFX, animation, illustration, and broader creative sectors; comparisons to Luddites, human “computers,” and earlier automation waves.
  • Several argue the real problem is economic structure, not technology: productivity gains accrue to capital without adequate safety nets.
  • Others wish AI progress targeted mundane physical work (robots doing dishes, construction, care tasks) rather than saturating screens with more media.

Misinformation, Deepfakes, and Porn

  • Concern that realistic AI video plus cheap credits will supercharge scams, fake news, and harassment (especially deepfaked sexual content and minors).
  • Some note we’re already seeing AI clips pass as real on TikTok/YouTube; future trust in video evidence and news is expected to deteriorate.
  • A minority argue that AI‑generated CSAM might displace real abuse material; others see this as ethically and legally fraught.

Naming, Branding, and Cultural Friction

  • The “Flow” name irritates some, seen as trading on the recent Oscar‑winning open‑source animated film Flow and, more broadly, on the same creatives AI may displace.
  • More generally, anger at big labs: perceived as appropriating artists’ work, firing ethics researchers, and pushing tools that benefit studios and platforms while harming working creatives.

Show HN: A Tiling Window Manager for Windows, Written in Janet

Overall reception

  • Very positive response to a tiling window manager for Windows written in Janet.
  • Praised as “cool”, “awesome”, and exciting especially for people who like WMs, Lisp, or need to use Windows reluctantly.
  • Several plan to adopt it as a daily driver or at least “take it for a spin.”

Community and ecosystem

  • The Windows tiling WM scene (e.g., this project, komorebi, GlazeWM) is described as unusually friendly and collaborative, with authors openly praising each other.
  • Some compare this favorably to, or on par with, the long-standing Linux WM ecosystem (StumpWM, i3, dwm, ratpoison, EXWM).

Janet and Lisp-specific advantages

  • Multiple comments highlight Janet as practical, small, fast, and inspired by Clojure and Lua while feeling “better” in practice.
  • The REPL and image-based, interactive development are cited as key reasons to use a Lisp for a long-running, stateful service like a window manager.
  • Structural editing plus a live REPL are framed as Lisp’s “killer features,” turning development into rapid, low-friction experimentation.
  • Some deep side discussion: Common Lisp’s condition/restart system, images vs files, undo/audit trails, and Smalltalk/Guile comparisons.
  • Tooling for Janet is seen as weaker than ideal but improving (e.g., LSP work).

Design and behavior of Jwno

  • Main differences to komorebi:
    • Manual tiling by default vs dynamic tiling.
    • Uses native Windows virtual desktops instead of its own workspaces.
    • Self-contained rather than IPC/CLI-driven.
  • Internal model: Root → Virtual Desktops → Monitors → Frames → Windows. All monitors switch together when changing Windows desktops.
  • Cookbook shows how to adapt layouts for ultrawide monitors or reserve space (e.g., for on-screen keyboards / accessibility).
  • Can call Win32 APIs from scripts for low-level control.

UI hinting and accessibility

  • UI-hint / “walk the UI tree” feature gets special praise as a powerful keyboard-accessible mechanism.
  • One user reports issues with AltGr-based keybindings and labels obscuring targets; the author investigates and suggests workarounds and configuration options.

Tiling WMs, Windows, and nostalgia

  • Many lament weak built-in window management on Windows, though some defend Win+Arrow and PowerToys’ FancyZones.
  • Nostalgic thread about alternative Windows shells (bb4win, litestep, etc.) and their role in customization rather than lasting “innovation.”
  • Some discussion of modal dialogs and floating windows; others note modern tiling WMs offer floating “groups” as a workaround.

Show HN: 90s.dev – Game maker that runs on the web

What 90s.dev Is (and Isn’t)

  • Several commenters say they don’t really understand what it is from the article alone.
  • The author describes it as: an API around a 320×180 web canvas, focused on building game-making tools and games, with sharing built-in.
  • It’s positioned as “like a more usable PICO‑8” for prototyping, but using TypeScript/TSX and IDE-like comforts, more a platform than a conventional engine.
  • Confusion arises because the article alternates between calling it a “game maker”, “not a game maker”, and “a platform”, which some find inconsistent.

Aesthetic and Nostalgia

  • Strong positive reaction to the 90s desktop GUI look: people mention Windows 95 vibes, mouse trails, and “simpler times.”
  • Some purists note 16:9 and 320×180 are not historically 90s; others don’t care and just enjoy the style.

Resolution and Constraints

  • 320×180 is chosen partly to mimic an existing modern game and to fit code + tools on screen.
  • The system can actually resize via a config file, but this is undocumented; debate ensues:
    • One side: constraints are powerful and shouldn’t have easy “escape hatches.”
    • Other side: hiding capabilities reduces flexibility and should at least be documented.
  • The compromise is to treat such features as “easter eggs” for power users.

Onboarding, Demos, and Docs

  • Multiple users say it’s hard to get started and the pitch is unclear.
  • People ask for: a very simple landing-page explanation, example games, and a “build a tiny game” walkthrough or video.
  • Currently, only built-in apps (e.g., fontmaker, paint) exist; no finished games yet, which some see as a weakness for a “game maker” launch.

Browser Support and Technical Issues

  • Works in Chrome; various breakages reported in Firefox ESR and a Firefox-based browser due to ServiceWorker and iterator quirks, which get quickly patched.
  • A bigger limitation: Firefox lacks showDirectoryPicker, so the fast “mount a folder” workflow is Chrome-only. A manual HTTP-server workaround is suggested.
  • Some users hit cross-origin issues when using embedded iframes; the fix is to open the OS in its own subdomain/tab.

Ecosystem, Collaboration, and Licensing

  • Strong emphasis on sharing apps, libraries, and assets to enable collaborative game development.
  • Plans to move from GitHub/npm-only imports toward direct HTTPS module imports and possibly bare-specifier support.
  • License clarified as MIT and added to source headers.

Launch Timing and Feedback Culture

  • The author feels they launched too early, without enough polished tools, games, and docs to show the power of the APIs.
  • Many commenters push back, encouraging early/iterative shipping and treating this as a solid “Show HN” stage rather than a finished product.

OpenAI Codex hands-on review

Everyday usefulness & limitations

  • Many see Codex as valuable for small, repetitive changes across many repos (README tweaks, link updates, minor refactors), treating it like a “junior engineer” that needs close review.
  • Reported success rates around 40–60% on small tasks are viewed as acceptable; for larger or more conceptual work, it often degrades code quality (e.g., making fields nullable, adding ts-nocheck) to “make it compile,” increasing technical debt.
  • It’s praised for generating tests and doing “API munging,” and for quickly surfacing relevant parts of an unfamiliar codebase, but multi-file patches often get stuck or go in circles.

Integrations, UX, and environment constraints

  • GitHub integration and workflow are widely criticized: awkward PR flows, flakiness in repo connection, slow setup, and poor support for iterative commits/checkpoints.
  • Lack of network access, inability to apt install or run containers/Docker is seen as a major blocker for real-world projects, especially those relying on external services or LocalStack-style setups.
  • Users want checkpointing lighter than full git commits and better support for containers and search; current “automated PR” flows are viewed as too brittle to trust.

Workflow patterns and prompt engineering

  • Effective use often involves:
    • Running many parallel instances/rollouts of the same prompt.
    • Selecting the best attempt and iteratively tightening prompts.
    • Splitting work into small, parallelizable chunks.
  • Some find this loop 5–10x more productive for certain tasks; others find prompt-tweaking overhead and “context poisoning” negate benefits.

Non-developers, low‑code, and quality concerns

  • There’s interest in letting non-devs use Codex for content/CSS fixes while devs review the resulting PRs.
  • Several commenters warn that even “small” changes can have hidden dependencies (data, PDFs, other services).
  • Accessibility, responsiveness, and cross-platform issues are flagged as areas where LLMs readily introduce regressions and can’t be reliably guarded by linters or prompts alone.

Comparisons to other tools

  • Compared to Claude Code, Codex is described as more conservative, slower per task, but able to run many tasks in parallel.
  • Some users find Claude and Gemini’s “attach a repo and chat” model, combined with large context windows and web search, more effective for debugging and complex reasoning today.
  • Cursor and other IDE agents are seen as great for one-shotting small features; when they fail mid-stream, it can be faster to write code manually.

Automation, jobs, and economics

  • The thread contains an extensive, conflicting debate about whether tools like Codex will:
    • Mostly augment engineers (doing more “P2” work, enabling more software overall).
    • Or materially displace software developers, especially juniors, with many comparing it to past waves of automation in farming and manufacturing.
  • Some argue productivity gains historically haven’t flowed primarily to workers and fear worse conditions or unemployment for many engineers.
  • Others counter that:
    • Coding has always automated others’ jobs; developers may likewise have to adapt or switch careers.
    • High-skill engineers will remain in demand to design systems, supervise agents, review code, and build/maintain agentic infrastructure.
  • There is specific concern about how new engineers will gain experience if entry-level coding work is offloaded to agents.

Security, naming, and adoption concerns

  • Cloning private repos into Codex sandboxes raises worries about exposing trade secrets, though some acknowledge this may be analogous to earlier cloud-source-control fears.
  • Confusion around model and product naming (Codex legacy model vs new Codex tool; “o3 finetune”) is noted as an industry-wide problem that hinders understanding and trust.

Overall sentiment

  • Net sentiment is cautiously positive on Codex as an assistant for small, well-scoped tasks and background agents.
  • There is broad skepticism about fully hands-off “agent does everything” workflows, current UX/integration quality, and the long-term labor implications.

Deep Learning Is Applied Topology

Topology vs. Geometry and Manifolds

  • Many argue the post is really about geometry/differential geometry (manifolds with metrics), not topology (which discards distances, angles, and scale).
  • Topology is described as what survives after “violent” deformations; deep learning in practice heavily relies on metric structure, loss landscapes, and distances.
  • Disagreement over the statement “data lives on a manifold”:
    • Pro: low‑dimensional latent structure seems required for high‑dimensional problems to be learnable.
    • Con: real data can be discrete, noisy, branching, and “thick”; manifolds are at best approximations, not literal.
  • Some note alternative meanings of “topology” (e.g., network topology, graphs) and complain the article conflates or overextends the mathematical term.

Theory vs. Empirical “Alchemy”

  • One camp: deep learning progress is mostly empirical, advanced by trial‑and‑error, heuristics, and engineering; theory (especially topology) has had little direct impact on widely used methods.
  • Counter‑camp: many ideas are adaptations or reinventions of prior theory (linear algebra, optimization, statistics, statistical physics, control, information theory); dismissing theory is short‑sighted technical debt.
  • There is no consensus on what a satisfactory “theory of deep learning” should deliver (convergence bounds, architecture guidance, hyperparameters, efficient weight computation, etc.), and some doubt such a theory will be very practical.

Representations, Circuits, and Transformers

  • Several comments favor thinking in terms of representation geometry: embeddings, loss landscapes, and trajectories of points under training; visualization via UMAP/t‑SNE; early “violent” manifold reshaping followed by refinement.
  • Work on linear representations and “circuits” (features as directions; networks of interacting features) is cited as more productive than purely topological views, including symmetry/equivariance analyses.
  • For transformers, attention is framed as a differentiable kernel smoother or distance‑measuring operation over learned embedding manifolds, with feed‑forward layers doing the geometric warping.

AGI, Reasoning, and Manifold Metaphors

  • Several object to the article’s casual claim that current methods have reached AGI, seeing it as a credibility hit.
  • Debate over whether “reasoning manifolds” or topological views meaningfully capture logical/probabilistic reasoning; some argue all reasoning is fundamentally probabilistic, others maintain discrete logical operations are essential.
  • The idea that AGI/ASI are just other points on the same manifold as current next‑token or chain‑of‑thought models is challenged as unjustified.

Usefulness and Limits of the Topology Framing

  • Some find the manifold picture a helpful intuition pump for understanding separation surfaces, embeddings, and similarity; others see it as a rebranding of long‑known “manifold learning” with little new.
  • Requests for concrete examples where topology (in the strict sense) improved models or understanding mostly go unanswered; topological data analysis is mentioned but seen as niche so far.
  • Multiple commenters conclude: deep learning certainly can be described using topology, but calling it “applied topology” adds little unless it yields new, testable design principles or performance gains.

Why does the U.S. always run a trade deficit?

Saving Gap vs. Trade Policy

  • Many comments accept the article’s “saving gap” framing: the trade deficit mirrors the gap between domestic saving and investment, not simply “bad trade deals.”
  • Others argue causality is reversed: abundant foreign demand for U.S. assets and dollars depresses U.S. interest rates, discourages saving and reshaping the economy toward consumption.
  • Some point out that accounting identities (saving = investment) are tautologies and don’t explain real-world distribution, sectoral impacts, or which investments are actually productive.

Reserve Currency, Capital Flows, and “Exporting Dollars”

  • A major thread: as issuer of the dominant reserve currency, the U.S. effectively “exports dollars” and “imports goods,” enabling Americans to consume beyond current production.
  • Capital inflows show up as foreigners buying Treasuries, stocks, real estate, and corporate assets; the flip side is persistent current-account deficits and a large negative net international investment position.
  • Debate on whether reserve-currency status forces a trade deficit: historically the U.S. had surpluses while the dollar was already central; critics say today’s problem is fiscal policy and under-taxation, not just reserve status.

Goods vs. Services and Measurement Problems

  • Several note that focusing on goods alone overstates the “deficit problem.” The U.S. runs a sizable surplus in services (software, cloud, media, finance, IP licensing).
  • Others argue services and intangibles (IP, platform control, branding) are mismeasured: where is the value of an iPhone actually “made”—China’s assembly or U.S. design and margins?

Globalization, Manufacturing, and China

  • Long discussion on offshoring to China: cheap labor plus scale beat U.S. manufacturing; many argue this hollowed out U.S. industrial jobs while enriching asset owners.
  • Counterpoint: U.S. manufacturing value is near all-time highs but highly automated and capital-intensive, so it employs far fewer people.
  • Some propose tariffs and industrial policy to force onshoring and automation; skeptics warn this raises consumer prices and can’t recreate 1990s factory jobs.

Sustainability, Debt, and Inequality

  • Concern that large, persistent deficits plus rising public debt will eventually force either higher taxes, inflation, or default-like outcomes.
  • Others note U.S. private assets still far exceed liabilities; as long as productivity and innovation stay strong, the position is manageable.
  • Many tie trade deficits, financialization, and asset inflation to domestic inequality and political instability.

National Security and Industrial Policy

  • Several argue the deeper issue is not the deficit per se but loss of critical capabilities (chips, energy, medical supplies, defense inputs).
  • This motivates support for targeted reshoring, CHIPS-style subsidies, and selective protectionism, even at some economic cost.

Politics and Economic Narratives

  • Multiple comments see “trade deficit” as a political weapon: a simple story used to justify tariffs (especially under Trump), despite the underlying macro-identity.
  • Split views on tariffs: some see them as necessary leverage against mercantilist surplus countries; others see them as self-harming, undermining the very reserve-currency system that currently benefits the U.S.

Skepticism About Economics

  • A recurring meta-thread questions macro models themselves: heavy reliance on simplified equations, unclear causality, and failure to capture power, institutions, and class impacts.
  • Some explicitly call for shifting attention from aggregate balances to who gains (owners of capital) and who loses (workers, regions, future taxpayers).

Reports of Deno's Demise Have Been Greatly Exaggerated

Overall reaction to the post

  • Many readers see the piece as defensive “spin” and PR-driven rather than a substantive response to criticisms.
  • The title and “we’re not dead” framing are viewed as risky: acknowledging the “demise” narrative makes some people more worried, not less.
  • Several commenters note that concrete adoption numbers are vague (e.g., “doubled MAUs” without a baseline), which reduces confidence.

Node/npm compatibility and loss of original vision

  • A recurring theme: Deno’s original appeal was a clean break from Node/npm—better standard library, URL imports, permissions, modern APIs, and a reset of dependency bloat.
  • The later pivot to Node/npm compatibility is seen by some as abandoning that vision to reduce friction and satisfy investors.
  • Critics say this:
    • Increases complexity and CLI/options surface.
    • Dilutes pressure to build “Deno-native” libraries.
    • Reintroduces the npm dependency tangle Deno was supposed to escape.
  • Defenders argue the compat layer is optional, pragmatically enables migration, and expands Deno’s usefulness.

Deploy, KV, Fresh, and “momentum”

  • Reduction of Deploy regions is widely read as a negative signal despite the explanation that “most apps don’t need to run everywhere.”
  • KV staying in beta while a replacement is planned makes people reluctant to adopt it; some call it effectively dead.
  • Fresh being refactored with a long-dated alpha timeline is interpreted as loss of focus; the removal of the “no build step” idea disappoints some.
  • Overall, these moves fuel a perception of strategic drift and weak momentum.

VC funding, business model, and trust

  • Multiple commenters won’t invest in Deno (or Bun) mainly because they’re VC-funded, fearing abrupt pivots or shutdowns.
  • Others counter that monetizing hosted services while keeping the runtime FOSS is reasonable, and similar to other ecosystems.

Deno vs Node vs Bun and language/tooling debates

  • Some prefer Bun for speed and Node compatibility; others distrust Bun’s crash-heavy, Zig-based implementation and praise Deno’s Rust codebase.
  • Several argue there’s little incentive to move existing Node apps to Deno; Node is “boring but stable” with a huge ecosystem.
  • Long subthreads debate TypeScript’s merits, shared types front/back, and whether one should just use Go, Java, C#, or other back-end stacks instead of chasing a better JS runtime.

AI's energy footprint

Transparency, ESG, and Corporate Claims

  • Strong sentiment that current ESG and “green AI” claims are mostly PR without standardized measurement, independent validation, and open data.
  • Several commenters highlight big tech’s refusal to disclose detailed power use as itself a red flag.
  • Some push back on specific cited studies (e.g., arXiv paper on data-center carbon intensity), arguing the methodology and state-level numbers look unreliable.

How Big and How Fast Is AI’s Energy Footprint Growing?

  • Many agree AI will significantly increase total electricity demand; some liken it to the brain using ~20% of body energy and foresee AI consuming a similar share of human output.
  • Others argue that, compared to aviation, transportation, or water systems, AI is still a modest slice and focusing on it alone is misleading.
  • Jevons paradox appears repeatedly: efficiency gains are expected to drive more usage, not less.

Carbon Intensity, Siting, and the Grid

  • Concern that US data centers cluster where electricity is cheapest and dirtiest (e.g., coal/gas-heavy regions), making their power ~50% more carbon intensive than the national average.
  • Debate over why more centers aren’t in hydro/geothermal-rich, cooler regions (Iceland, Quebec); replies cite bandwidth limits, grid capacity, construction and logistics.
  • Some see the AI boom as a catalyst to modernize grids; others note most new “clean” energy for data centers could have displaced fossil generation instead.

Policy: Carbon Pricing vs Targeting AI

  • One camp: internalize emissions into electricity prices (carbon taxes, certificates) and stop moralizing over specific uses—AI, showers, or phones.
  • Counterpoint: carbon pricing is politically hard, can be regressive, and rich users may barely change behavior.
  • Long thread on revenue‑neutral carbon taxes, fairness, bans vs markets, and how to set a realistic social cost of carbon.

Training, Inference, and Hardware Efficiency

  • Agreement that inference dominates AI energy once models are deployed, though training is still highly energy- and capital-intensive.
  • Some argue energy cost is minor relative to GPU CapEx, so operators push for 100% utilization even if that means dirtier energy.
  • Optimists: rapid cost/energy per token declines (distillation, specialized chips like TPUs, small on-device models) could flatten AI’s energy curve, akin to past data‑center efficiency gains.
  • Skeptics: don’t expect CPU‑like exponential improvements; silicon “low‑hanging fruit” is gone, and huge parameter counts still imply massive operations.

Value vs Waste: Are AI Uses Worth the Power?

  • Many criticize “AI everywhere” (search summaries, unwanted product features, novelty image/video generation) as low‑value slop that burns power for minimal benefit.
  • Others argue AI will raise productivity, automate drudgery, and perhaps decarbonization itself—so the right metric is “economic or social value per unit energy.”
  • Comparisons to crypto recur: some see AI as another hype‑driven resource misallocation; others say, unlike proof‑of‑work, AI at least has real and potential utility.

Cooling, Water Use, and Local Impacts

  • Worry about data centers consuming millions of gallons of fresh water per day for cooling, especially where they compete with municipal supplies.
  • Oil immersion cooling is discussed; practitioners mostly dismiss it as expensive and operationally painful, with limited benefit over existing techniques.
  • A few note that evaporative cooling returns water to the cycle, so it’s less comparable to pollution-heavy uses.

Measurement Gaps and Methodological Disputes

  • Several people find the article’s numbers “all over the place,” especially on image generation, and note that hardware and aspect‑ratio choices can change energy use drastically.
  • Others emphasize missing pieces: mobile networks, wireless power amplifiers, scraping load on third‑party servers, embodied carbon in hardware, etc.
  • Critiques that the piece leans on dramatic units (billions of gallons, square feet) without systematic comparison to other sectors, which some see as manipulative.

Societal Trade‑offs and Future Paths

  • Ongoing tension between focusing on demand‑side restraint (warning users per prompt, guilt framing) vs supply‑side decarbonization (“make clean energy abundant and cheap”).
  • Some foresee AI moving from “mainframe era” hyperscale to efficient, mostly on‑device models for everyday use, with giant clusters reserved for frontier training and science.
  • Others are more pessimistic, seeing AI as yet another driver of consumption, inequality, and environmental stress unless policy, pricing, and governance change direction.

The behavior of LLMs in hiring decisions: Systemic biases in candidate selection

Observed biases in LLM hiring experiments

  • Commenters focus on two main findings from the article:
    • A consistent preference for female candidates when CV quality is controlled and genders are swapped.
    • A strong positional bias toward the candidate listed first in the prompt.
  • Several note that these are statistically robust results (tens of thousands of trials), not random variation.
  • Grok and other major models reportedly show similar patterns; DeepSeek V3 is mentioned as somewhat less biased in the tests.

Debate on causes of gender bias

  • One camp argues the models may be reflecting real-world hiring trends (e.g., efforts to rebalance male-heavy fields), or underlying “left-leaning” cultural norms baked into text data.
  • Others think the effect is more likely from post-training alignment/RLHF, which aggressively avoids discrimination and may overcorrect toward favoring women and pronoun-displaying candidates.
  • There’s extended back-and-forth over empirical studies of gender bias in academia, with participants citing conflicting papers and pointing out publication bias and narrative-driven citation patterns.

Positional / context biases

  • The first-candidate preference is widely seen as the most alarming technical flaw: it suggests LLMs don’t evenly weight context and can base “decisions” on trivial ordering.
  • People link this to known “lost in the middle” issues and warn that RAG and classification systems may be quietly influenced by such artifacts.
  • Some propose simple mitigations (randomizing order, multiple runs, blind prompts), but note most real HR users won’t do this.

Suitability of LLMs for hiring

  • Many argue LLMs should not be used to make hiring decisions at all, only to summarize CVs, translate jargon, or generate pro/con notes for human reviewers.
  • Others emphasize that LLM outputs are articulate but not grounded reasoning, and numeric scores/probabilities from them are especially untrustworthy.
  • There’s concern that companies may use biased AI as a “liability shield” for discriminatory outcomes.

Training data, politics, and gaming the system

  • Multiple comments discuss systemic leftward bias due to overrepresentation of certain professions and platforms in training data, then further amplified by alignment.
  • Some suggest synthetic data and filtering might reduce extremes, but worry self-generated data could reinforce existing bias.
  • Recruiters report that LLMs tend to favor LLM-written “optimized” resumes, and people speculate about adversarial tricks (invisible text in PDFs) to manipulate AI screening.

Finland announces migration of its rail network to international gauge

Implementation options & engineering constraints

  • Commenters discuss how to practically convert 9,000+ km of track: line‑by‑line vs building parallel standard‑gauge tracks.
  • Dual‑gauge is mostly ruled out: 1524 mm vs 1435 mm are too close; triple rail is infeasible so you need four rails and very complex switches.
  • Concrete sleepers are a major constraint: unlike old wooden ties you can’t just move a rail; most sections would need full relaying and new points.
  • Some imagine specialized track‑relaying machines doing continuous gauge conversion, but others note current machines only manage a few km/day and are rare.

Historical and international precedents

  • US South’s 1886 gauge change in ~36 hours is cited as an extreme historical example, but seen as inapplicable due to modern speeds, weights, safety and concrete sleepers.
  • Spain’s long, still‑ongoing dual‑gauge era (Iberian vs standard) is mentioned; gauge‑changing high‑speed trains work but add complexity, unreliability and political friction.
  • Rail Baltica is debated: some call it a money sink with “no visible result”, others point to ongoing high‑speed standard‑gauge construction and completed Poland–Lithuania link.
  • India’s slow, multi‑decade gauge standardisation and various smaller historical gauge shifts are also referenced.

Economic value vs military rationale

  • Several think the project will never “pay for itself” economically, given Finland’s limited rail connections to mainland Europe and the ability to transfer cargo at borders.
  • Supporters argue the core driver is defence: NATO logistics from Sweden/Norway into Finland without gauge breaks and denying Russian rolling stock easy use of Finnish rails.
  • Critics counter that rails near the Russian border could simply be destroyed in war; others reply Russia has large rail engineering units and repairs are fast, while destruction also harms Finnish/NATO logistics.
  • There’s disagreement on how much gauge breaks actually slow military throughput (minor inconvenience vs serious bottleneck).

Transition period & interoperability

  • Most expect decades of mixed operation: building parallel standard‑gauge where possible while keeping broad‑gauge in service.
  • Adjustable‑gauge bogies (as in Spain or Switzerland) are discussed but said to be too complex, climate‑sensitive, and insufficient for mass freight logistics.
  • Existing border practices—lifting carriages to swap bogies or using small dual‑gauge sections—are seen as workable but too slow for the project’s goals.

EU integration, geography & scope

  • EU’s TEN‑T policy and potential EU funding are seen as major drivers; some view this as Brussels‑driven political integration, not Finnish demand.
  • Finland is described as a near “rail island”: only a single link to Sweden in the remote north, none to Norway yet, and any Helsinki–Europe high‑speed link would likely depend on speculative megaprojects (e.g. Helsinki–Tallinn tunnel).
  • Some argue the gauge change only makes real sense if bundled with broader upgrades: electrification, signalling (ETCS), speed improvements and necessary renewals of already‑ageing Finnish track.

Making video games (without an engine) in 2025

Engines vs. Custom Code

  • Many argue big engines (Unity, Unreal, Godot) are popular because they solve hard, generic problems: animation blending/IK, physics, audio, rendering, asset streaming, cross‑platform builds, etc.
  • Others counter that for small, 2D, or tightly scoped projects, “no engine” can be simpler, faster, and more enjoyable, especially if 90% of engine features are unused.
  • Several note that using an engine lets you work on the game itself instead of “yak‑shaving” infrastructure; others say building the tech is the fun part and can be central to their motivation.

How Hard Are the “Hard” Parts?

  • Some commenters claim pieces like simple IK, blending, and 2D physics are not that bad, especially with narrow scope.
  • Others emphasize that robust, production‑grade physics, rendering, spatial audio, and advanced animation pipelines are extremely deep domains where homegrown solutions often fall short or consume huge time.

Tooling, Editors, and Asset Pipelines

  • A recurring theme: the runtime “engine” is the easy part; the real workload is tooling—importers, editors, asset baking, visualization, debugging, profiling, and content pipelines.
  • Multiple experienced devs report spending far more time on tools than on the core engine, and say this is where many custom‑engine projects stall.
  • Suggested compromise: use existing DCC tools (Blender, Tiled, Maya, Excel, etc.) plus hot‑reloading, or even build tooling on top of existing engines.

Productivity, Learning, and Procrastination

  • Several see “reinventing the wheel” as procrastination from the hard creative work of designing fun, polishing content, and shipping.
  • Others say writing subsystems yourself gives deep understanding, full ownership, and reusable code across a long career, which can justify weeks or months of work.
  • Common advice: if your goal is to ship quickly, use an engine; if your goal is learning or engine programming itself, rolling your own is valuable.

When Custom Engines Shine (or Hurt)

  • Custom approaches are seen as most justified for:
    • Highly non‑standard mechanics (non‑Euclidean, 4D, unusual physics, rewind systems, huge open worlds with dynamic lighting, etc.).
    • Very focused 2D or retro‑style games where scope is tightly controlled.
  • Others warn that even in these cases, engines can often be bent to fit, and that many acclaimed indie games with custom engines required years of prior experience.

C#, .NET, and GC

  • The thread is broadly positive on modern C#/.NET: cross‑platform, NativeAOT, hot‑reload, low‑level control, source generators.
  • GC pauses are acknowledged but seen as manageable via pooling, initialization‑time allocation, or new GC work (like Satori), especially for games that load most assets up front.

AI in my plasma physics research didn’t go the way I expected

Academic incentives & publication bias

  • Commenters stress that overselling, cherry-picking, and non-publication of negative results predate AI and stem from how careers and journals reward “exciting” results and citations.
  • AI hype amplifies this: “flag-planting” papers from big labs are hard to ignore or critique, especially for under-resourced universities that can’t replicate large-scale experiments.
  • Several note that a key function of PhD training is learning to “read through” papers, understanding them as artifacts of a sociotechnical system, not neutral truth.

Benchmarks, statistics, and replication

  • A linked medical-imaging paper argues many “state-of-the-art” claims evaporate once confidence intervals are considered; competing models are statistically indistinguishable.
  • Commenters are surprised that basic statistical practice (e.g., reporting confidence intervals) is often missing in high‑stakes fields like medicine.
  • Benchmarks in AI are criticized as fragile, often relying on secret datasets, non-replicable setups, and single-number summaries that hide uncertainty.

AI for physics & numerical methods (PINNs, FEM, etc.)

  • Multiple researchers report that physics-informed neural networks and AI structural/FEM solvers work tolerably only on simple, linear regimes and break down on nonlinear or out-of-distribution problems.
  • A recurring pattern: ML models reproduce training data but generalize poorly, while papers still imply broad applicability without actually testing it.
  • Some characterize “AI for numerical simulations” as “industrial-scale p‑hacking” or a hammer in search of nails.

Universities vs industry & funding politics

  • Once a topic becomes a resource arms race with industry, some argue it no longer fits the core mission of universities (long‑term, foundational, low‑resource work).
  • Discussion of NSF funding cuts and political attacks: waste exists (e.g., “use up the budget” equipment), but commenters view research/education as extremely high ROI and compare academic waste favorably to corporate boondoggles.

What counts as AI success in science?

  • Skeptics ask where the genuine AI-driven breakthroughs are; others cite protein folding, numerical weather prediction, drug discovery hit rates, and recent algorithm‑design work (e.g., matrix multiplication, kissing‑number bounds).
  • There’s disagreement over how overfitted or fragile some of these successes might be, and whether they represent general scientific reasoning versus powerful prediction/hypothesis‑generation tools.

LLMs, productivity, and erosion of competence

  • Many report substantial gains from LLMs for coding, document drafting, search over messy corpora, and meeting transcription; others find them slow, noisy, or dangerous in high‑stakes scientific programming.
  • A tension emerges: LLMs can speed up routine work, but may also encourage shallow understanding and brittle workflows if users stop deeply engaging with code, math, or data.

Conceptual confusion around “AI” & hype dynamics

  • Several argue “AI” is an almost meaningless marketing term, lumping together classic ML, deep learning, LLMs, and domain‑specific models; serious discussion requires more precise labels.
  • Others defend “AI” as a useful umbrella for recent neural‑network advances, while acknowledging rampant buzzword abuse (from smartphone cameras to “smart toilets”).
  • Underneath, commenters converge that:
    • The current “AI will revolutionize science” narrative is ahead of robust evidence.
    • Incentives (career, funding, corporate valuation) strongly favor overstating AI’s scientific impact.
    • Nonetheless, as a tool for search, pattern-finding, and acceleration of certain workflows, AI is already meaningfully useful—and may yet yield deeper advances if used with rigorous methods and honest statistics.

What are people doing? Live-ish estimates based on global population dynamics

Design & Overall Concept

  • Many commenters praise the visual design, especially the day–night world map and the ability to deliver the whole simulation in (essentially) a single HTML file.
  • The project is seen as an elegant, engaging “live” snapshot of humanity, even if only approximate.
  • Some prefer other time‑use visualizations (e.g., US‑only survey-based ones) as more precise, but still find this one compelling.

Emotional & Philosophical Reactions

  • Watching deaths tick by (~2 per second) is described as sobering; it foregrounds how quickly whole lifetimes disappear.
  • Others counterbalance this with the idea that at the same time billions are actively living, loving, celebrating, and building memories.
  • Several find the near‑parity between “Intimacy” and “Warfare” numbers sad; others note that intimacy at least often exceeds warfare, which is seen as a hopeful signal.
  • Restroom and smoking/sex comparisons generate humor, but also reflections on how we actually spend our time.

Data Quality, Realism & Modeling

  • Some are skeptical of the underlying “population dynamics,” calling parts of it “vibe-coded” and insufficiently sourced or grounded.
  • The flickering births/deaths‑per‑second numbers are criticized as unrealistic once the author’s deliberate randomization is noticed; suggestions include modeling births as a Poisson process or varying by local time.
  • Sleep percentages are questioned (seeming too low at certain times), with speculation that the model over-relies on sunrise/sunset and doesn’t handle latitude or cultural sleep patterns well.
  • Similar doubts arise over “Intimacy” and prison counts and whether they’re anything more than fixed ratios.

Interpretation of Specific Stats

  • The share of people in paid work (often seen around mid‑teens %) surprises some as low, but others walk through life‑cycle and hours‑per‑week math and find it plausible.
  • Combined paid work + education near one‑third of activity is viewed as roughly consistent with 8‑hour days, accounting for children and retirees.
  • Population growth sparks debate: concern about net positive growth vs. concern about long‑term demographic decline and pension sustainability; population momentum and regional fertility differences are mentioned.

UI Choices & Feature Ideas

  • The phone icon for “Leisure” sparks debate; alternatives like couch, book, beach, or dancing are proposed, with recognition that phones probably reflect reality.
  • Requested additions include: regional/zoomable views, heatmaps for births/deaths, per‑country stats, more realistic time dynamics, and explicit counts for people in the air, at sea, or in space.
  • Several suggest turning the “Live Viewers” metric into a first‑class activity category for meta-fun.