Hacker News, Distilled

AI powered summaries for selected HN discussions.

Page 355 of 536

Launch HN: Miyagi (YC W25) turns YouTube videos into online, interactive courses

Overall Reception and Concept

  • Many commenters find the idea “magical,” intuitive in hindsight, and a natural fit for LLMs (auto‑generating quizzes, tutors, and structure around existing videos).
  • Several educators and edtech practitioners say this solves a real pain point: turning scattered YouTube learning into something structured and interactive.
  • Others are more skeptical, noting users can already feed transcripts into general-purpose LLMs and asking what value Miyagi adds beyond convenience and UI.

Content Quality, Pedagogy, and Trust

  • Big concern: how to ensure generated courses aren’t “teaching nonsense,” especially for learners who can’t detect subtle errors.
  • For “official” courses, creators review and edit content; this is currently treated as the main quality signal.
  • For user-generated courses, there is no systematic human review; future plans include user feedback on questions, learning paths, and possibly more formal evaluation.
  • Edtech professionals raise questions about deeper methodologies (knowledge tracing, skills taxonomies, learning outcomes, retention), which are not yet clearly implemented.
  • Some argue assessments are overemphasized; for self-directed adult learners, quizzes may be more “demo” than true value unless backed by tracking, validation, and credentials.

Experience, Features, and UX Feedback

  • Positive feedback on the general interface and idea of integrated tutor, quizzes, and flashcards, but multiple users report bugs, long generation times, and login issues.
  • Suggestions:
    • Clean up long lecture lists and better sectioning.
    • Integrate bottom tools into a single, more agent-like tutor.
    • More gamification, cohorts/social learning, and real-world artifact-based tasks, not just trivia.
    • Smarter “watch” links that jump to the relevant video segment.
  • Debate over giving the AI tutor a “persona”: some like the idea; others find overly humanized AI off-putting.

Use Cases and Scope

  • Interest in:
    • Language learning (with repetition and grammar focus).
    • Child/elementary content, though the current tutor safety is not trusted for unsupervised young kids.
    • Poker, chess, and other domains that may need custom tooling (solvers, interactive boards).
    • Niche “how-to” content (e.g., home improvement), including the potential use of comments as errata.
  • Currently relies mostly on transcripts; no video-frame understanding yet, though that’s on the roadmap. Supports PDFs, slides, and other file types as inputs.

Copyright, Ethics, and Creator Relations

  • This is the most contentious theme.
  • Miyagi says:
    • Any monetized course includes revenue sharing with creators, with some signed partnerships.
    • Embedded videos are used, and creators can request takedowns.
    • Non-partner content can still be summarized and augmented, with opt-out rather than opt-in.
  • Critics argue:
    • Opt-out with ambiguous revenue terms is ethically weak; should be explicit opt-in and generous sharing since creators supply most of the value.
    • Even if embedding is allowed, generating derivative educational products without consent feels exploitative to many.
    • “Extra views” are not sufficient compensation if a platform monetizes AI-based derivatives.
  • Comparisons are drawn to:
    • YouTube’s own AI summaries, accused of reducing watch time while monetizing summaries without adequate creator compensation.
    • Earlier controversies where companies enrolled creators into monetized systems without clear opt-in, which produced severe backlash.
  • Legal status (derivative work vs. fair use) is left unresolved in the thread; multiple commenters flag this as a major long-term risk and perception issue.

Platform Risk and Future Direction

  • Some discuss dependence on YouTube APIs and subtitles; founders state they already support direct uploads and could expand to other subtitle sources.
  • Commenters from traditional edtech note this idea has been floated before but ran into pedagogy, licensing, and “no-ads” constraints at established platforms, hinting that a startup can move faster but must still address those same issues over time.

AI Is Like a Crappy Consultant

Perceived Productivity Gains

  • Many describe a large jump in usefulness: from “idiot intern” to something like a junior/crappy consultant.
  • Strong praise for AI code completion and refactoring: described as “god mode,” faster than static typing + IDE refactors + Vim/Emacs for large, mechanical edits.
  • Helpful for learning unfamiliar APIs and replacing much Stack Overflow–style searching.
  • Some say AI is far better than Google search and expect it to displace traditional search; others prefer curated search engines (e.g., Kagi) for deterministic, non‑hallucinated results.

Code Quality, Architecture, and “Vibe Coding”

  • Common theme: AI is poor at architecture and data-structure design, tends to force new problems into existing, suboptimal patterns and assumes current code and user instructions are correct.
  • “Vibe coding” (letting the AI build systems end‑to‑end) is seen as risky; multiple anecdotes of silent failures (e.g., broken file migration scripts) and chaotic student projects.
  • Several argue good engineering practice hasn’t changed: tests (ideally TDD), specs, and understanding the code are still critical.

Search, Fact-Checking, and Hallucinations

  • Disagreement over whether modern models meaningfully “fact check” or “cite” versus just wrapping search tools.
  • Critics stress that the core network lacks source attribution, so it can’t explain where a given code fragment or fact came from, which underlies hallucinations and licensing issues.
  • Supporters counter that tool-augmented LLMs already behave like fact‑checkers for many tasks.

Ethics, Training Data, and Citation

  • Significant concern that models are built on unconsented training data, can’t track provenance, and hallucinate licenses; contrasted with human expectations around attribution.
  • Others downplay this, focusing more on practical utility than on data origin.

Roles, Metaphors, and Anthropomorphism

  • AI compared variously to: crappy consultant, junior engineer, fast intern, or dangerous tool.
  • Some warn against anthropomorphizing; others argue it does exhibit rudimentary reasoning and produces genuinely novel, useful “knowledge,” disputing the “stochastic parrot” label.
  • Heated subthread over what “knowledge” means and whether LLMs “understand” anything.

Prompting Strategies and Tools

  • Multiple workflow tips: use AI for tests/docs and repetitive edits; constrain tasks tightly; ask for alternative solutions; force it to ask clarifying questions; reset context often.
  • Tools like Aider + advanced models are reported to outperform basic IDE integrations, though they introduce complexity (diff formats, configuration).

Socioeconomic and Cultural Notes

  • Some see AI as aligning with executive incentives: cheap, confident answers, fueling interest in replacing devs.
  • Calls for unionization and worries about low-quality, AI‑driven software becoming widespread.

Internet Artifacts

Overall Reaction & Nostalgia

  • Many commenters describe the site as a “pure nostalgia hit,” likening it to flipping through an old photo album.
  • Several recall specific formative memories: early blogging, first MP3s, Netscape vs IE wars, dial‑up, early flash games, and the “helicopter game” as a proto–Flappy Bird.
  • A few mention feeling old when they suddenly understand their parents’ nostalgia for the 50s/60s aesthetic.

Specific Artifacts & Notable Omissions

  • Strong reactions to seeing Netscape Navigator’s “meteors,” Homestar Runner, Line Rider, Million Dollar Homepage, Heaven’s Gate, etc.
  • People list many “missing” artifacts: Newgrounds, AltaVista, Yahoo Answers, LiveJournal, Something Awful, xkcd, Slashdot, digg/fark ecosystems, RuneScape/Ultima Online, toolbars, Clippy, WinRAR, early IM (ICQ), goatse and other shock sites, etc.
  • Debate over whether things like Bad Apple!! or scaruffi.com are important enough historically.
  • Some praise and some dislike the modern version of Ishkur’s Guide to Electronic Music, calling it overengineered compared to the original.

Space Jam Website & Redirect Rabbit Hole

  • Appreciation that the original 1996 Space Jam site is still accessible, and that the 2021 site kept a retro style.
  • Others note the main domain is now partially broken, with conflicting reports about redirect loops and certificate problems.
  • The thread dives into detailed HTTP/HTTPS, 301/302, and Upgrade: h2,h2c behavior, with people trading full curl -vvv traces and debating whether there is truly a loop.

Old vs Modern Internet

  • Strong sentiment that pre‑social, pre‑smartphone web felt more sincere, experimental, and less commercial.
  • Several blame the iPhone/app era and today’s “hypercommercial, grifty” environment for killing that spirit, though some note pockets of genuine self‑expression still exist.
  • Alternatives like the Gemini protocol are suggested as a way to recapture a simpler, text‑centric net.

Geography & Cultural Scope

  • Multiple commenters criticize the site as very US/Western‑centric and wish for parallel “artifact timelines” for German, Russian, Chinese, and other language communities.
  • Examples from Runet (bash.org.ru, Masyanya, “padonki” slang) and early German sites are cited as parallel but largely unknown histories.

Design & Interaction

  • Widespread praise for the polish and interactivity (scrollable “live” artifacts, playful details like Zombo.com audio persisting).
  • Minor UX complaints: some swipes jump multiple items; simulated slow image loading gets annoying over time.

Conspiracy theorists can be deprogrammed

AI “Deprogramming” as Tool and Threat

  • Many see using AI to “deprogram” conspiracy theorists as inherently political: deprogramming is just re-programming to someone else’s agenda.
  • Concern that such systems could easily become government or corporate propaganda tools, worse than a minority of conspiracy believers.
  • Others counter that this “tool” already exists and is being used; better to do it transparently than leave the field to opaque actors.

Can AI Also Radicalize?

  • Multiple comments argue the reverse is not only possible but already happening: social media algorithms bombard people with “alternative facts” for engagement.
  • Several say it’s easier and more profitable to create new believers than to deprogram.
  • Some think the real asymmetry is that extremist and Nazi-style content has many promoters and few effective deprogrammers.

Conspiracies: Noise vs Signal

  • One camp: conspiracy theories mostly add noise, making genuine conspiracies harder to detect (e.g., QAnon obscuring real trafficking).
  • Another: skepticism toward authority is rational; some “theories” later proved true (Watergate, industry cover-ups, etc.), so blanket pathologizing is wrong.
  • Debate over whether elite coordination is mostly “just incentives” or effectively a conspiracy in all but name.

Trust, Authority, and Epistemology

  • Repeated theme: conspiracists don’t actually reject authority; they just relocate it—from institutions to podcasts, influencers, and anonymous accounts.
  • Split between those who see conspiracists as curious but under-informed vs. those who see them as preferring emotionally satisfying narratives over primary sources.
  • Once institutional trust is broken, some say it is nearly impossible to restore; LLMs that follow “official lines” only deepen suspicion.

Study Design and Definition Issues

  • Criticism that the underlying research redefines “conspiracy theory” as “untrue conspiracy,” excluding widely accepted real conspiracies (e.g., lobbying, corporate cover-ups) and thereby baking in ideological bias.
  • Objections to heavy reliance on GPT-4 for screening without human validation.

Social Media, Addiction, and Regulation

  • Some argue treating social media addiction and regulating recommendation systems would be more effective than AI deprogramming.
  • Proposals include legal penalties for knowingly spreading political misinformation, but others warn this quickly becomes censorship.

LLMs as Socratic Partners

  • Several see value in AI’s patience and Socratic questioning to foster self-reflection without human fatigue.
  • Anecdotes show LLMs can generate strong counter-arguments, but may need human “shepherding” and risk being dismissed as partisan or censored.

One hundred and one rules of effective living

Ambiguity and Interpretation of Specific Rules

  • Several rules are criticized as vague or poorly phrased, e.g. “uncanny congruity between thought and experience,” which readers parse both as “experiences shape thoughts” and “attitude shapes experiences,” noting the latter can fuel both misery and happiness.
  • The rule about giving no second chances to disrespectful people is seen by some as protective and empowering (e.g., never returning to a disrespectful business, leaving a toxic boss), and by others as cruel, inflexible, and life‑shrinking.
  • The line about hourly billing is disputed: some argue hourly workers still serve the client and that billing mode is orthogonal to integrity; others note hourly is appropriate when scope is unknown, while flat‑rate work has its own pathologies.
  • The primacy of “do your work” as rule #1 is criticized as hollow and work‑obsessed; others reinterpret “work” broadly as meaningful projects, not just jobs.

Too Many Rules vs. Simpler Principles

  • Multiple commenters argue 101 rules are overengineered and self‑contradictory, preferring compact frameworks: the Daoist “three treasures,” the Golden Rule, short personal lists, or a single existential challenge.
  • Some note direct contradictions (e.g., avoid cruel people vs. endorsement of being feared) and see the list as a grab‑bag of quotes rather than a coherent ethic.
  • Others say the real value is not correctness but provoking reflection: noticing which rules trigger agreement, anger, or “that’s bullshit” becomes self‑knowledge.

Comparisons to Other Moral Programs

  • Benjamin Franklin’s 13 virtues receive attention: some admire their clarity and discipline, others find them joyless or hypocritical given Franklin’s reputed lifestyle.
  • A nuanced view emerges: strict ideals are aspirational, not fully attainable—like training a dog that will always still want to bark.

Limits of Inspirational Literature

  • Skepticism is expressed toward people deeply immersed in “inspirational” texts, which can become mind‑numbing platitudes detached from real trade‑offs.
  • Several note that many lessons on the list can only be truly learned through lived experience, but that a phrase can “hit” at the right psychological moment and catalyze change.

Meta: Speech, Knowledge, and Online Culture

  • The rule to “speak only of what you know well” is seen as incompatible with most internet discourse.
  • Some lament that online spaces often punish beginner questions and reward overconfident ignorance, pushing people toward tools like ChatGPT for nonjudgmental exploration.

Kindness, Cynicism, and Paranoia

  • A minimalist counterproposal is to boil everything down to kindness or “don’t be a jerk.”
  • Some rules in the list read as paranoid (e.g., expecting betrayal), while others explicitly condemn cynicism, creating tension that readers notice and debate.

Multiple security issues in GNU Screen

Setuid-root design and security issues

  • Core concern is Screen’s multi-user mode requiring a setuid-root binary, massively increasing attack surface for complex code.
  • Several commenters were unaware this feature existed; others used it heavily for shared debugging, training, and pair programming.
  • Some argue using setuid without fully understanding the codebase is especially bad; others note Screen’s design dates from the 1980s/1990s when setuid was common.
  • Multi-user mode is praised as powerful but now widely seen as a serious security liability.

Distro configurations and affectedness

  • Impact varies by distribution and build:
    • Some distros ship Screen setuid-root (or setgid to a special group), enabling multi-user mode and related vulnerabilities.
    • Others (Debian, Slackware, some Gentoo/Arch setups) ship Screen without setuid and are unaffected by the worst bugs, though specific CVEs differ by version/build.
  • Links to an openSUSE matrix clarify which versions and builds are affected.
  • Some distros (e.g., RHEL) have effectively deprecated Screen in favor of tmux, though it may linger in extra repositories.

Alternatives: tmux, zellij, and minimal tools

  • Many commenters long ago switched to tmux, noting:
    • No need for setuid; uses Unix domain sockets.
    • Audited in some OS bases; considered more modern and maintainable.
    • Some still prefer Screen for serial-port support and specific behavior; tmux’s server/session/window/pane model confuses a few users.
  • zellij is highlighted as a modern, more discoverable Rust-based alternative, though some report past latency and missing fully keyboard-driven copy/paste.
  • Minimal tools (dtach, abduco+dvtm, mtm) are recommended by those who equate security with very small codebases, though they have terminal capability limitations.

Project health, tech debt, and migration

  • Upstream reportedly requested the security review but appears understaffed and hard to reach; Screen is seen as “bitrotting” yet still widely used.
  • Some argue Screen’s architecture is so dated/complex that it’s effectively doomed; others warn against “rewrite from scratch” thinking.
  • Suggestions include:
    • Distros replacing Screen with tmux or a “tmux-as-screen” compatibility wrapper.
    • Keeping Screen only for niche features like serial console support.

Mailing lists and tooling / usability debates

  • Side discussion on mailing lists: praised for openness, federation, and longevity; criticized for poor web UX, accessibility issues, and difficult search.
  • Several note older projects (including GNU tools) bury issues on lists instead of using modern forge-based issue trackers, hurting discoverability and maintenance.

Git Bug: Distributed, Offline-First Bug Tracker Embedded in Git, with Bridges

Role of a Bug Tracker and Intended Audience

  • Several commenters note that real-world bug trackers are used by support, QA, design, and management, not just engineers.
  • Concern that a Git-embedded tracker may be too engineering-centric and awkward for non-developers, especially if access is via Git or terminal tools.
  • Others argue this is fine: the tool seems aimed more at OSS / engineering-heavy workflows than full corporate project management.

Why Put Issues in Git at All?

  • Proponents see clear benefits:
    • Offline, local-first workflows (file bugs, read history, etc. without network access).
    • All project knowledge travels with the repo (easier backups, mirroring, migration, and forge independence).
    • Potential for cross-forge interoperability if multiple platforms read/write the same Git-based issue data.
  • Some dislike that current forges bolt centralized SQL-backed trackers, CI/CD, and AI tightly onto Git, increasing lock-in.

Technical and Design Challenges

  • Doubts about representing inherently relational issue data in Git objects, and losing SQL-style querying and integrity.
  • Counterpoint: you can index Git-stored data into SQLite locally; ACID isn’t as critical in an offline, async, single-user context.
  • Open questions: schemas and workflows standardization, multi-repo or cross-project issues, avoiding confusing extra refs, and making this usable for non-devs.

Distributed Semantics and Conflict Handling

  • Skeptics worry about merging issue conversations and “weird” ordering when multiple people edit offline.
  • Maintainers explain they use Lamport timestamps and an operation log over a DAG, so merges don’t produce conflicts; events are replayed in logical order.
  • Late-pushed comments can appear earlier in the timeline, but that’s framed as reflecting the true creation order.

Ecosystem, Extensions, and Comparisons

  • Multiple references to Fossil and other Git-based issue tools; this project is seen as part of a broader “everything in the VCS” trend.
  • Git namespaces (e.g., refs/bugs) are praised as a flexible mechanism; maintainers position git-bug as a generic framework for storing entities in Git (issues, project boards, reviews, etc.).
  • Bridges exist for GitHub/GitLab/Jira; long-term hope is direct forge integration, though some doubt big platforms will adopt it.

UX, Documentation, and Adoption Concerns

  • Many find the README and docs fragmented and missing a clear “Why?” and basic “How do I add a bug?” path; screenshots and higher-level overviews are requested.
  • There is enthusiasm for the TUI and web UI, but also calls for better frontends (including editor integration).
  • Some see strong potential as a personal or small-team TODO/bug system; others think large companies will stick with centralized tools like Jira.

What were the MS-DOS programs that the moricons.dll icons were intended for?

Nostalgia for Windows 3.x and moricons.dll

  • Many recall discovering moricons.dll as a kid and treating it as a hidden treasure trove to decorate Program Manager and PIFs.
  • Several remember misinterpreting “moricons” as a nonsense word rather than “more icons.”
  • The icons trigger strong sensory memories: Windows startup sounds, hard-drive and floppy noises, and general early-90s PC vibes.
  • Some note using moricons not for the “intended” apps but as a general-purpose icon set, often just picking anything vaguely related and making sure no two apps shared the same icon.

WordPerfect, DOS Productivity Software, and Icon History

  • WordPerfect evokes intense nostalgia, especially its DOS-era dominance, function-key cheat strips, and “Reveal Codes,” which are compared to LaTeX-like explicit markup and praised for precise formatting (notably in legal work).
  • Discussion contrasts WordPerfect’s DOS excellence with a disastrous transition to Windows (crashes, bloat, interface and macro changes) that pushed users to Word.
  • Other canonical DOS apps mentioned: Lotus 1-2-3, dBase, Harvard Graphics, Flight Simulator, Crosstalk XVI, Q&A, DesignCAD, etc.
  • Clarification that “Access for DOS” was a communications tool, not a database.
  • Several comments answer “why ship third‑party icons?”: Windows 3.1’s setup scanned for existing DOS apps (via APPS.INF) and assigned them appropriate icons so the new GUI would feel polished and familiar.

Learning to Program: QBasic, Turbo Pascal, Borland C++

  • Many share childhood misunderstandings: thinking QBasic/Turbo Pascal were text editors that “ran games,” not compilers; not realizing .BAS files could be edited in any editor; odd confusions about strings, numeric types, or error messages.
  • QBasic, NIBBLES.BAS, and GORILLAS.BAS feature heavily as first programming experiences and bug-fix memories.
  • Turbo Pascal is praised for speed and a great debugger; Borland C++ is remembered as powerful but confusing for beginners.
  • Modern fantasy consoles (PICO-8 vs TIC‑80) are mentioned as ways to recapture that exploratory feeling, with some preferring polished constraints (PICO‑8) and others favoring open-source flexibility (TIC‑80).

PIF Files, Icons, and Technical Tidbits

  • PIF files are recalled as configuration for DOS apps under Windows (memory, options), later treated as special shortcuts whose extensions are hidden and which are effectively executable, contributing to past exploits.
  • Links are shared documenting the PIF format; some lament such lore disappearing without archives.
  • Icon craft is admired: low-res icons seen as surprisingly expressive; some criticize the “icon on lined paper” style as visually busy on modern displays.
  • A CSS snippet using image-rendering: pixelated is shared to view icons crisply, with side discussion about how CRT-era appearance differed from modern pixel-perfect scaling.

Broader Reflections on Computing Then vs Now

  • Several contrast the early PC era’s sense of user control and discovery with today’s locked-down, ad-driven, or subscription-based ecosystems.
  • Comments note that past constraints were mostly hardware or skill limits, whereas now they’re often manufacturer- or policy-imposed.
  • Others argue constraints and artificial segmentation existed even then (e.g., mainframes, 486 variants), but agree modern firmware locks and SoC keys raise the stakes.
  • There are calls for stronger consumer protection, clearer distinctions between sold vs. rented devices, and pushback against ownership-eroding practices.

Changes since congestion pricing started in New York

Perceived Early Outcomes in NYC

  • Many commenters see congestion pricing as “working”: fewer cars, faster traffic, better bus reliability, less noise, especially in Manhattan’s core.
  • Graphs showing cars averaging ~9 mph are viewed as stark evidence that pre‑policy car infrastructure was highly inefficient.
  • Some on-the-ground reports describe more foot traffic in lower Manhattan and easier bus travel from New Jersey/Queens; others anecdotally report eerie midweek emptiness and more vacant storefronts.
  • Business impacts are contested: article stats show visitors and restaurant bookings up, but several people are skeptical, noting business self‑diagnoses are unreliable and early data is thin.

Measurement, Causality, and “A/B Testing”

  • Several participants want policy changes bundled with explicit evaluation plans and success metrics (e.g., commute times, bus speeds, business revenue).
  • Others stress that true randomized A/B tests at city scale are impossible; before/after comparisons are confounded by macro trends (economy, seasons, other policies).
  • Some call these “fundamentally unanswerable questions” where we can only collect suggestive evidence rather than clean causality.
  • There’s criticism that the NYT piece feels like advocacy: positives are quantified, while sections on pollution, low‑income commuters, and public opinion are “too soon to tell.”

Equity, Fairness, and Who Pays

  • Critics argue congestion fees are inherently regressive, turning Manhattan into a “playground for the rich” and burdening low‑income drivers who lack viable transit.
  • Others counter that:
    • Most commuters into lower Manhattan already use transit.
    • Car commuters into the zone skew higher income.
    • A very small share of working poor drive into the zone and some receive waivers or discounts.
  • Debate over fairness: some see a flat fee as fair (same price for same service); others see any flat access charge as exclusionary when it changes behavior mainly for poorer users.
  • Alternative ideas (lottery/plate bans, income-based pricing) are discussed but criticized as harder to tune, more intrusive, and more disruptive to people with fixed-time obligations.

Cars vs. Transit, Bikes, and Urban Form

  • Strong anti‑car contingent: cars dominate space, generate noise, pollution, injuries, and sprawl; they argue many “needs” for cars are artifacts of car‑centric planning.
  • Defenders emphasize convenience, carrying capacity, weather protection, perceived personal safety, and US low density; they resist policies that intentionally raise “at‑use” costs.
  • Congestion pricing is framed by some as a bootstrapping tool: shifting marginal trips to transit, justifying better service and eventually supporting denser, more walkable neighborhoods.
  • Bikes and e‑bikes feature heavily:
    • Advocates highlight huge capacity and space gains, plus evidence that bike and foot traffic spend more locally.
    • Others describe “lawless” delivery e‑bikes, serious crashes, and call for registration and camera enforcement; replies stress cars remain orders of magnitude more dangerous.

Families, Accessibility, and Quality of Life

  • Some parents describe car‑centric suburbs as “hell” with kids and see dense, transit‑rich neighborhoods (Manhattan, Brooklyn, European cities) as ideal.
  • Others highlight the practical difficulties: double strollers on stairs, limited elevators, crowded trains, expensive family housing, and the need for cars in outer boroughs or US regions without good transit.
  • There’s broad agreement that improved transit accessibility (elevators, better buses) is a necessary complement to congestion pricing.

Politics, Slippery Slopes, and Scaling

  • NYC is seen as a special case: uniquely dense, pre‑car transit skeleton, and massive existing demand; many doubt simple transferability to car‑dependent US metros.
  • Some fear a “slippery slope” from pricing to broader driving restrictions or de facto bans on internal combustion vehicles; others call this normal policy diffusion (successful ideas spreading) rather than escalation.
  • The governor’s earlier delay and fee reduction are criticized as short‑term political maneuvering that risked underfunding the transit improvements pricing is meant to support.

The world could run on older hardware if software optimization was a priority

Economic incentives vs. optimization

  • Many comments frame this as a resource-allocation problem: developer time vs. hardware.
  • Hardware, cloud compute, and storage are seen as “dirt cheap” compared to engineers; it’s usually rational to ship features on slower code rather than spend months optimizing.
  • Several argue this creates negative externalities: e‑waste, higher energy use, constant forced upgrades, and time wasted by end‑users staring at sluggish apps. Others push back that this is just the tradeoff users and buyers implicitly choose.
  • Some connect this to “enshittification” and “market for lemons”: buyers can’t easily judge software quality/performance up front, so vendors optimize for features and sales decks, not efficiency.

Bloat, UX, and real hardware

  • A long thread of examples (Slack, Teams, VS Code, Outlook, web apps, Windows 11 UI, Electron generally) describe multi‑second startup times and janky interactions even on high‑end machines.
  • Developers often build and test on fast Macs with fiber; typical users run mid‑range Windows laptops or old phones on slow networks, where the same apps feel painful.
  • Some note that older native apps (Win9x/XP era, early Photoshop/Office, simple IDEs) felt instant on far weaker hardware; “modern” equivalents often feel slower on 100–1000× more capable machines.

Where performance went: abstractions, features, safety

  • Several claim most of the 1000× raw CPU gain since ~1980 has been spent on layers of abstraction (high‑level languages, VMs, ORMs, web stacks, Electron) and richer UIs, not on bounds checks or security.
  • Others counter that abstractions bought enormous developer productivity and software abundance; the marginal user often values “software exists at all” more than 2×–5× speed.
  • Heated subthreads debate whether writing “fast” C++/Rust is actually that costly, or whether huge speedups (10–1000×) are often low‑hanging fruit lost to ignorance, deadlines, and cargo‑cult patterns (N+1 queries, over‑network calls, JSON‑everywhere).

Servers vs clients; I/O, energy, and scale

  • Platform/infra engineers note that in large fleets, most CPU sits idle; the true bottlenecks are I/O (disks, networks, databases), contention, and bad schema/query design.
  • Even small inefficiencies in widely deployed code or libraries multiply to large global energy and hardware costs, but those costs are diffuse and rarely drive product decisions.
  • Others point out that in data centers newer hardware is upgraded mainly for power density and efficiency, not raw speed—older gear really can be too wasteful to keep.

Old hardware already in use

  • Several remind that much of the world already runs on “old” or low‑end chips: embedded controllers, industrial systems, cars, ATMs, power plants, even space hardware with decades‑old rad‑hard CPUs.
  • At home, many report perfectly usable experiences on 10–15‑year‑old desktops and laptops, provided they avoid heavy modern web apps or switch to lighter OSes and software.

Performance vs. safety and correctness

  • A substantial side discussion: if dynamic bounds checking and memory‑safe languages cost ~5% performance, would that be a good trade for drastically fewer vulnerabilities and easier debugging?
  • Some argue we collectively chose “1000× speed with same bugs” over “950× speed with far fewer memory issues”, and that this was a mistake; others reply that even safety competes with features, time‑to‑market, and other forms of value.

Cultural and tooling questions

  • Older developers lament a lost “craft” of efficiency, replaced by “get it working” on fast machines and frameworks.
  • Others note that where users or money demand speed (HFT, exchange engines, game engines, core infra, some big web properties), optimization still happens aggressively.
  • There’s recurring speculation that future AI tooling might eventually help auto‑optimize bloated codebases—but also concern that current AI waves mostly amplify compute waste rather than reduce it.

Working on complex systems: What I learned working at Google

Scale vs. Complexity

  • Several commenters stress that most projects are not at Google scale but can still be genuinely complex (e.g., game engines, logistics, enterprise integrations).
  • Others argue “complex” is orthogonal to “large”: small systems can be conceptually hard; large systems can do relatively simple things.
  • Some push back on developers copying FAANG-style architectures for tiny workloads (“cargo culting”), noting Google’s architecture is a response to scale, not a recipe for success.

Complex vs. Complicated / Types of Complexity

  • Multiple threads debate the article’s complex/complicated distinction and note overlaps with established notions: essential vs. incidental complexity, domain vs. accidental/environmental complexity.
  • Critics say the article doesn’t clearly separate:
    • complexity you choose (architecture, tools),
    • complexity imposed by the domain (logistics, manufacturing, regulations, human behavior).
  • Examples like electronic proof-of-delivery in logistics highlight socio-technical, behavioral, and infrastructural constraints as true “complexity,” not just big codebases.

Life Inside Google: Organizational Complexity

  • Commenters describe “collaboration headwind”: many approvals, domain owners, and process layers turning trivial changes (e.g., button text) into months of work.
  • A common claim is that much effort goes into fighting incidental organizational complexity rather than pure technical difficulty.
  • One perspective defends stricter process and testing as rational in a large org with high downside risk; others see it as demoralizing and leading to workarounds instead of improvements.

Testing, Code Ownership, and Friction

  • There is an extended back-and-forth on code review expectations:
    • Some reviewers insist on adding or backfilling tests before accepting even obvious fixes.
    • Contributors find it unreasonable when asked to introduce full test suites into untested code just to land a small change.
    • Others argue this is exactly what professional engineering should require; untested changes are seen as liabilities.

Incentives and Peer Bonuses

  • Peer bonus systems are described as “tips” funded by the company, intended to reward extra help.
  • Some see them as positive and motivating (especially when paid in cash); others view them as gamified, low-value recognition that substitutes for real raises or PTO.

Google, Games, and Product Judgment

  • Discussion notes Google’s failed or shuttered gaming efforts (e.g., Stadia) and past work via Niantic.
  • Several argue Google and also mobile platform owners struggle to understand the games industry and tend to push complexity onto developers.

Meta and Miscellaneous

  • Multiple people point out the irony of a buggy, reappearing cookie banner on an article about complex systems.
  • Others reference systems thinking (Meadows, Cynefin, complex systems literature) and appreciate the topic but criticize redefining established terms or glossing over long-standing research.

Nextcloud cries foul over Google Play Store app rejection

Nextcloud’s Play Store Issue and Current Status

  • Google rejected the Nextcloud Android app for using MANAGE_EXTERNAL_STORAGE (“all files” access), a permission it had been allowed to use as an exception for ~2 years.
  • This breaks features like auto-upload / backup of arbitrary folders (photos, WhatsApp, app data, SD card trees, etc.) in the Play Store build.
  • A later comment says Google responded and agreed to re‑grant the permission on resubmission, so full functionality should return soon.

Security Lockdown vs Functionality

  • One side argues Google is rightly tightening storage access after years of abuse by shady apps exfiltrating data via broad filesystem permissions.
  • They say Nextcloud should use Android’s Storage Access Framework (SAF) / scoped storage, letting users grant per‑folder access via system dialogs.
  • Others say this is paternalistic: advanced users want to explicitly grant “full disk” access to trusted tools like sync/backup apps, and Google shouldn’t forbid that.

Can SAF Replace All-Files Access? (Technical Dispute)

  • Supporters of SAF:
    • Claim SAF has existed since Android 4.4 and can handle “pick directory and sync everything under it” scenarios.
    • Argue Nextcloud misreads docs: SAF can maintain persistent directory access and doesn’t inherently expose its data to other apps.
  • Critics of SAF:
    • Point out restrictions: no access to Downloads/ root/ Android/data/obb, and performance / complexity issues, especially with native code.
    • Say some workflows (e.g., full-device backups, arbitrary app data, SD cards) are impossible or degraded versus legacy APIs.
    • Note other projects (Syncthing, Kiwix, some editors) dropped or weakened Android support over these constraints.

Competition, Monopolies, and Policy Asymmetry

  • Many see this as anticompetitive: Google can protect its own backup/Drive ecosystem while gatekeeping competitors via Play policy.
  • Others counter that Google Drive on Android also doesn’t sync arbitrary folders and doesn’t use the same high-privilege permission; some Google system apps (Files, Android Auto) do.
  • Broader comparison with Apple:
    • Debates over DMA, notarization, “core technology fees,” and walled‑garden vs “prison cell” metaphors.
    • Some argue both platforms are converging on tightly controlled, less “general-purpose” devices.

User Control, Rooting, and Attestation

  • Some say sideloading/F-Droid/GrapheneOS are the “escape hatch.”
  • Others respond that hardware attestation and bank apps’ root checks already punish users who deviate, making Android “a shitty version of iOS.”

Developer Experience & Nextcloud Quality

  • Several devs describe Play review as opaque, copy‑pasted, and easily abused via bogus reports.
  • Opinions on Nextcloud itself are mixed: powerful and empowering for self-hosters, but often buggy, unpolished, and painful to maintain beyond simple use cases.

Ask HN: How are you acquiring your first hundred users?

Overall range of experiences

  • Founders report everything from “dumb luck and a good product” to years of grind to get first 100 users.
  • Some hit 100+ off a single HN/Reddit/App Store launch; others struggle to reach a few dozen despite lots of content and outreach.
  • Several say there is no formula—only experimentation and channel–product fit.

Finding and reaching early users

  • Repeated advice: identify a very specific ideal customer profile and go where they already congregate (subreddits, forums, LinkedIn groups, niche communities, Slack/Discord, professional associations).
  • Heavy emphasis on 1:1 outreach: cold email, cold calling, DMs, walking into offices, using existing professional networks and friends, and hand-holding early onboarding.
  • Some advocate interviewing 10–30 potential customers first (Mom Test style) and building from their real pains.

Free tiers, pricing, and conversion

  • Strong debate on free/freemium:
    • Pro: free lowers friction, especially for developers; success stories from DB/infra tools and freemium dev SaaS.
    • Con: fears it trains users not to pay; risk of competing on price and never converting.
  • Several use “free but limited” (time, data retention, branding) or generous trials; others lead with premium pricing but offer steep discounts/early adopter deals.

Channels that worked (or failed)

  • Worked for some:
    • HN Show/Ask posts, Reddit niche subs, app store listings, SEO (especially long-tail/programmatic pages), content marketing, YouTube, LinkedIn posts, TikTok demos, newsletters/Substack, marketplaces (e.g., Xero, Slack, integration directories), open source funnels, partner distributions, “engineering-as-marketing” tools.
  • Mixed/weak results: Product Hunt, generic Reddit subs (mods and hostility), broad paid ads early on, postcard campaigns, AI tool directories.
  • Many stress measuring each channel (unique links, discounts, QR codes).

Product, brand, and trust

  • Common pattern: build for yourself → test with small network (5–20 people) → expand via communities and referrals.
  • Word-of-mouth and virality (referrals, built-in sharing, embedded branding) seen as powerful but hard to engineer.
  • Debate over visual polish (AI art vs bespoke vs simple screenshots); some think early users don’t care, others argue presentation matters a lot before reputation exists.

Meta and skepticism

  • Cynicism about fake growth (bot farms, “growth hacking” bordering on fraud).
  • Suspicion of self-promotional posts and AI-generated comments/blogs; concern that marketing discourse itself is becoming performative and automated.
  • Several insist that a “great product” without a repeatable acquisition channel is a hobby, not a business.

Odin: A programming language made for me

Zero Initialization (ZII) and Uninitialized Memory

  • Big thread on Odin’s “everything is zero-initialized”: critics say it silently propagates wrong values; they’d prefer compile‑time errors or hard crashes on reads before writes.
  • Defenders argue ZII removes undefined behavior and matches common OS behavior (zeroed pages), improving repeatability and security and fixing a notable fraction of bugs/CVEs linked to uninitialized stack use.
  • Others counter that ZII prevents compilers and analyzers from warning about missing initialization, turning logic errors into “valid” zero cases that are harder to catch.
  • Several comparisons:
    • Rust/C#/Dart do definite‑assignment analysis; Go/Odin/Java zero‑init.
    • Some want “dangerous” behavior (UB, wrapping, uninitialized) to be explicitly opt‑in.
    • Example code and Rust/C UB discussions emphasize that uninitialized variables literally have no value, not just “unknown” ones.
  • Odin still allows explicit uninitialized locals via a special --- syntax, and nil dereferences are defined to panic rather than be UB.

Safety, Ergonomics, and Software Quality

  • One camp wants stricter languages and more compile‑time guarantees, arguing software quality and security are broadly poor and rigor should be default.
  • Another camp pushes back on “everything must be Rust/Ada strict”, stressing many programs are not life‑critical and ergonomics, iteration speed, and test‑based correctness matter.
  • There is disagreement about how much risk in everyday software is acceptable (from calculators asking permissions to system‑wide outages) and whether striving for “better” should always trump practicality.

Odin vs C (and Rust/Zig)

  • Several argue C’s core semantics and type system (arrays, safety, UB) are too broken to “fix” via libraries; attempts to bolt on slices, defer, or better stdlibs hit hard limits.
  • Odin is seen as “what C would be if redesigned now”: tagged unions, slices, better stdlib, allocators, SoA support, generics, distinct types, multiple returns, context system, with C‑like simplicity and fast compilation.
  • Some note you can emulate many patterns in C (custom allocators, arenas, safer wrappers), but ergonomics and defaults are poor, and there’s no agreed replacement stdlib.
  • Rust and Zig are generally respected; Odin fans emphasize Odin’s lower conceptual overhead and familiarity for C‑style workflows rather than safety guarantees.

Data as Bytes vs Higher-Level Abstractions

  • One line of argument: as long as programmers think in “flat bytes and allocators”, abstraction leaks (e.g., painful SoA refactors) are inevitable; data types should be abstract, with layout left to the system except where explicitly constrained.
  • Many push back: in systems domains (games, HPC, compilers) cache locality and layout are the problem; compilers can’t infer intent or universally “figure out the optimal layout”.
  • Data‑oriented design and custom allocators are cited as crucial for performance; skepticism remains about “sufficiently smart compilers” solving layout, SIMD, and parallelism automatically.

Experiences Using Odin

  • Multiple users describe Odin as a “sweet spot” C‑replacement: procedural, low‑level, but with first‑class slices, dynamic arrays, maps, allocators, SoA attributes, and a well‑designed stdlib.
  • Game‑oriented workflows benefit from ZII, arenas, context passing, and SoA; some treat Odin almost like a compiled scripting language (simple syntax, near‑instant builds, Raylib bindings).
  • The implicit “context” feature sparks mixed reactions: powerful for cross‑cutting concerns (e.g., time contexts in games), but some worry about hidden coupling.

Ecosystem, Tooling, and Language Competition

  • Discussion of whether C, Odin, Zig, Jai, etc. will coexist or whether one low‑level language will capture most mindshare; many expect Rust to dominate industry jobs, with others remaining niche.
  • Odin has good C FFI, but still lacks broad tools, tutorials, and AI‑assistant support; LLM autocompletion is reported as weak due to low training exposure.
  • Some worry that many niche languages fragment ecosystems and duplicate effort; others see diversity and domain‑specific trade‑offs as healthy.

Mozilla Firefox – Official GitHub repo

Move from Mercurial to Git

  • Firefox’s canonical repo has moved from Mercurial to Git; GitHub is now the source of truth, with hg repos synced from Git instead of vice versa.
  • Previously both hg and git (via git-cinnabar) were supported; this change effectively begins the phase-out of hg.
  • Some lament the “git monoculture” and see this as a final blow to Mercurial, though others note hg is still maintained and used elsewhere.
  • Reported practical issues with hg included very slow initial clones compared to the unofficial git mirror.

Why GitHub and a New Org

  • Using mozilla-firefox instead of mozilla is attributed to GitHub’s org-level scoping: SSO, permissions, visibility, and policies are all per-org, so isolating Firefox can reduce risk and complexity.
  • GitHub’s limited hierarchy (essentially just org + repo) makes multiple orgs a common workaround; some contrast this with GitLab’s namespace model or GitHub Enterprise-level features.

Centralization, GitHub vs Alternatives

  • Strong disagreement over choosing GitHub instead of self-hosted Forgejo/Codeberg/GitLab.
  • Pro-GitHub arguments:
    • Contributor familiarity and discoverability (“be where the contributors are”).
    • Free, robust hosting for a very large repo; easier than running high-availability VCS infra yourself.
  • Anti-GitHub concerns:
    • Proprietary, Microsoft-owned, used for Copilot training; seen as misaligned with Mozilla’s open-web mission.
    • No IPv6, U.S. sanctions, and phone-number requirements block some potential contributors.
    • Risk of ecosystem lock-in around issues/PRs and social graphs, even though the code itself remains portable.
  • Codeberg is mentioned but criticized for uptime; GitLab.com’s FOSS program terms are seen as legally problematic for some.

How Firefox Uses GitHub

  • GitHub is currently used only for code and PR hosting; PRs are auto-closed with instructions to use existing workflows.
  • Bugs stay in Bugzilla; code review remains in Phabricator/Phorge with Lando; CI stays on TaskCluster.
  • Branch mapping: former mozilla-centralmain; autoland remains as a staging branch merged into main when CI passes; various tree-named forks are used for large feature work.

Git, Workflow, and Tooling Debates

  • Long subthread on distributed vs centralized reality: git is still distributed technically, but the ecosystem is socially centralized on GitHub.
  • Suggestions to store issues/metadata inside git (e.g., git-bug, Radicle, git-notes) and to federate forges via ActivityPub; adoption remains low.
  • Extended arguments over pull requests vs email-based workflows vs Gerrit/Phabricator; PR UIs seen as both a huge usability win and a source of low-quality drive-by changes.
  • Many view Git’s UX as clunky compared to Mercurial or Fossil, but acknowledge GitHub significantly softened the learning curve.

Contributions, Gatekeeping, and UX

  • One camp argues GitHub lowers friction and is essential for attracting new contributors; another claims that people unwilling to learn non-GitHub workflows often produce low-value contributions.
  • Big debate over “gatekeeping”:
    • Some say raising process barriers filters spam and poorly thought-out PRs.
    • Others counter that any extra barrier discourages legitimate contributors and that maintainers should solve spam with tooling, not platform choice.
  • Contributors describe giving up on Firefox patches due to the complexity of combined GitHub + Phabricator flows before this migration.

Bug Tracking and Code Search

  • Many insist Bugzilla should remain, at least read-only, as a unique trove of historical web-compat reasoning.
  • GitHub Issues is widely viewed as feature-poor compared to Firefox’s customized Bugzilla instance.
  • For code navigation, some welcome GitHub’s search; others say Mozilla’s own Searchfox (and predecessors MXR/DXR) is significantly better for deep, cross-language navigation.

What if humanity forgot how to make CPUs?

Plausibility of “forgetting” how to make CPUs

  • Many see the premise as unrealistic: to fully lose global CPU manufacturing would require extreme events (world war, civilization-scale collapse, or a magical “no more silicon” rule).
  • Critics argue that if we’re in a state where no CPUs can be made for decades, the deeper problem is societal collapse, not AWS uptime.
  • Others treat it as pure SF / thought experiment: its value is in exploring hardware lifetimes and dependencies, not realism.

Impact on civilization and population

  • Debate on how dependent 8B people are on CPUs for food production and logistics. Some think we can’t support current population without them; others think we’d adapt at lower scale.
  • Emphasis that the most critical chips aren’t in phones or PCs but in power grids, factories, transportation, and healthcare.
  • Historical analogies: Rome, “Dark Ages,” WWII logistics, and lost techniques show that advanced knowledge can disappear while civilization continues, but restarting now would be harder due to depleted easy fossil fuels and higher complexity.

Reinventing computation and fallback tech

  • Many argue we’d quickly recreate 1970s–1990s-level CPUs because:
    • We know they’re possible and have abundant documentation, die shots, and theory.
    • Numerous non–cutting-edge fabs (e.g., 180–28 nm, microcontrollers) already exist.
  • Others stress tacit, undocumented process knowledge and the singularity of EUV / ASML-like tooling; recreating cutting edge might take decades.
  • Several note fallbacks: vacuum tubes, relays, electromechanical computers, older lithography, alternative semiconductors (e.g., non-silicon materials).
  • Even crude CPUs (6502-class) or discrete-logic machines can run robots and bootstrap more factories.

Institutional knowledge and supply-chain fragility

  • Strong theme: institutional know‑how is brittle. Examples from other domains:
    • Difficulties restarting closed manufacturing lines (appliances, weapons materials, bluing formulas, CRTs).
    • Specialized processes lost or hard to reproduce (FOGBANK, CRTs, high‑end tape/CD players).
  • Concern that heavy reliance on a few private companies (TSMC, ASML, others) makes advanced nodes a strategic single point of failure, even if older nodes persist.

Hardware longevity and decay

  • Several posts focus on failure timelines:
    • Consumer gear often starts failing after 10–15 years (capacitors, plastics, drives).
    • Large-feature chips resist electromigration, but true lifetimes for 5/3 nm are unknown.
    • Keeping complex automated equipment idle often ruins it if shutdown wasn’t designed/documented.
  • Consensus: there’s enough existing hardware to bridge at least a decade or two, likely enough to rebootstrap some semiconductor capability, but not indefinitely.

Broader reflections

  • Some think a constraint on new CPUs might finally force efficient software and less wasteful computing—though others reply that CPU frequency scaling has already mostly stalled.
  • Several highlight that digital records and infrastructure are far more fragile than we assume; analog artifacts (e.g., ancient scrolls) may outlast our data centers.

A conversation about AI for science with Jason Pruet

Role of DOE/National Labs and Context of the Interview

  • Many see the piece as partly PR but still useful context on how DOE and national labs frame “AI for science” and national security.
  • Labs are under strong top-down pressure to “do AI,” following an earlier public-cloud push; some view this as creating long-term rent streams for cloud vendors.
  • DOE already runs major GPU-based exascale supercomputers and plans to provide infrastructure to universities.
  • Some technical gripes surface about working on classified systems (no binaries, no PyTorch, awkward FIPS constraints), contributing to earlier reliance on tools like Mathematica.
  • Separate note: “1663” is explained as LANL’s science-and-tech magazine, named after its WWII PO box.
  • One comment mentions recent heavy LANL layoffs and a sense of anxiety inside the lab.

Public–Private Partnerships, Capture, and Tech Transfer

  • Central worry: if labs depend on industry for frontier models and compute, public research could be “captured” by commercial agendas—data, methods, and IP effectively controlled by a few firms.
  • Some view this as part of a broader pattern of privatizing state capacity (real estate, missile defense as a subscription service, etc.), leaving government structurally dependent on contractors.
  • Others argue this is exactly what tech transfer is for: public R&D → private commercialization, which historically enabled much of today’s tech stack.
  • Counterpoint: that logic assumes fair, transparent processes; critics doubt that’ll hold when a handful of AI firms control crucial infrastructure.
  • There is sharp distrust of defense contractors and “entrepreneurs” seen as driving cost inflation, fraud, and lock-in.
  • Some call for AI critical to national security to be kept entirely inside DoD, without private IP; others stress that labs do work with a wide range of companies, including startups.

AI vs. Climate, Energy, and National Priorities

  • One camp: fix global warming and build clean energy first; worries that AI hype diverts capital, power, and political attention.
  • Another: you can and must do both; delaying AI means trailing other economies and militaries. Historical analogies (Manhattan Project, Space Race) are invoked for large-scale tech investment.
  • Debate on whether AI meaningfully accelerates climate solutions (optimization, power efficiency, planning) or is mostly a distraction that increases energy demand.
  • Some argue a rapid build-out of renewables, storage, EVs, and heat pumps could halve emissions without AI; others emphasize that electricity is only a fraction of total emissions, with hard tradeoffs elsewhere.
  • Subthread on “degrowth”: some say reduced per-capita energy use is necessary; others insist increasing energy use is core to progress (Kardashev-style thinking).
  • Nationalism and geopolitics recur: concerns about “China will win if we slow AI” vs arguments that competitive framing itself worsens both AI risk and climate risk.

What “AI for Science” Means

  • Several comments note the article’s “AI for science” framing isn’t just about LLMs; examples like AlphaFold or geometry-proving systems are cited as more emblematic scientific advances.
  • Some readers suspect that current AI enthusiasm at LANL and DOE is being politically driven (“because they’re being told to”) and worry about administration-level capture by tech interests.
  • Others see real promise if national labs focus on scientific applications—simulation, materials, climate modeling—rather than primarily chasing commercial generative AI.

Benchmarks, Hype, and Real-World Performance

  • Interview claims about AI surpassing humans on almost all benchmarks are widely challenged.
  • Commenters note that tools like Gemini can look impressive in short “play” sessions yet fail unpredictably on simple tasks, hallucinate, or produce plausible nonsense.
  • There is frustration that current benchmarks underweight reliability, non-hallucination, and long-horizon reasoning—areas where humans still excel.
  • Some think LLM progress is already plateauing and that parameter-scaling gives diminishing returns; others argue agent capabilities are clearly improving even if raw knowledge isn’t exploding anymore.
  • A meta-critique emerges: optimists accuse skeptics of “performative cynicism” and moving goalposts; skeptics say claims of inevitable rapid improvement are marketing, not evidence.

Are LLMs “Good Coders”?

  • Strong disagreement about the statement that modern models are “very good coders.”
  • Supportive view:
    • They’re excellent at syntax, boilerplate, pattern recall, quick prototypes, and occasionally at insights that would take humans far longer.
    • They can transform workflows for developers who know enough to validate output.
  • Critical view:
    • They lack domain understanding, don’t know requirements, can’t judge ticket quality, and don’t understand cross-team impacts, so they’re autocomplete tools, not programmers.
    • Their failures are unpredictable: sometimes flounder on trivial tasks, sometimes nail advanced ones.
    • They exhibit a kind of “systematic Dunning–Kruger”: confidently wrong, always producing something instead of admitting ignorance.
  • Many see them as useful assistants when you can rapidly check their work, but not reliable enough to own end-to-end software tasks.
  • There is also skepticism about claims of being “very good legal analysts,” which some find especially implausible.

Governance, AI Futures, and General Mood

  • Some prefer commercially driven AI development over state/military-led programs; others argue that once research and infrastructure depend on private platforms, control shifts dangerously fast.
  • A playful but serious tangent speculates about AI CEOs: one commenter extrapolates from current agent time-horizon benchmarks to predict AI-led firms in the 2030s, while others treat this as premature.
  • The thread ends with a mix of awe and dread: AI feels like it could usher in a renaissance or a breakdown, and many commenters explicitly say it’s unclear which path we’re on.

Can you trust that permission pop-up on macOS?

Slack / Electron “helper tool” prompts on macOS

  • Multiple people report frequent, intrusive dialogs like “Slack is trying to install a new helper tool,” often asking for the admin password and reappearing if canceled.
  • Explanation offered: these come from macOS’s Service Management framework; apps (often Electron-based ones like Slack, Discord, VS Code) install privileged helper tools, mainly for auto-updates or system-level tasks.
  • Users question why simple apps need root-equivalent helpers and note that MDM/EDR tools (SentinelOne, CrowdStrike, etc.) can interfere, causing repeated prompts.
  • Some avoid native Electron apps entirely, using web apps or macOS web-app shortcuts instead.

Permission dialogs as “security theater” & prompt fatigue

  • Many describe severe “permission fatigue”: constant prompts for admin passwords, local-network access, removable drives, app downloads, etc., to the point they stop reading dialogs and just click Allow or Cancel by habit.
  • Corporate-managed Macs can show dozens of prompts a day, especially with security tools and frequent updates.
  • Users compare this unfavorably to earlier Apple ads mocking Windows Vista’s UAC, arguing macOS is now worse.
  • Some run as non-admin and/or install apps into ~/Applications so updates don’t need elevation, though others note this may reduce security in some cases and interacts oddly with macOS protections.

Spoofing, TCC, and trust in the UI

  • Central worry: any app or website can visually mimic macOS permission/password dialogs; users are being trained to trust and respond to random prompts.
  • People discuss ideas like security images, LEDs, “secure desktop” like Windows UAC, touch ID-only flows, or dialogs attached to specific windows/Settings, but note these are only partial defenses and can confuse users.
  • macOS’s TCC and capability model are criticized as bolted-on and inconsistent: they hinder legitimate devs, confuse users, and yet keep getting bypasses (like the CVE from the article).

Apple’s patching and platform direction

  • Several are disturbed it took Apple about a year to patch this bug and that the fix landed only in macOS Sequoia 15.5, leaving Ventura and Sonoma vulnerable by design.
  • Debate over whether Apple leans too much on App Store review/notarization instead of hardening the runtime; some see the prompts as partly a funnel into the App Store ecosystem.
  • Comparisons with Windows and Linux highlight that no platform gets this balance of security vs usability right, but macOS’s current UX is widely viewed as confusing and easily abused.

HealthBench – An evaluation for AI systems and human health

Model performance, visibility, and access

  • Commenters note Grok scores surprisingly well, and argue its lower mindshare vs Gemini/Llama is due more to lack of API access until recently than to open‑weights issues.
  • Some point out that open weights are mostly irrelevant here since only one of the ten benchmarked models is open anyway.
  • Gemini’s performance is seen as better than expected, with speculation that its tendency to refuse health topics (“censorship”) likely hurt scores. Med‑PaLM is mentioned as obsolete, superseded by Gemini.

Trust, bias, and conflict of interest

  • Many see an inherent conflict when a model vendor designs its own benchmark, especially one where its model narrowly beats competitors.
  • Others argue the benchmark is still useful, but should be read skeptically given no company would publish a study that makes its product look bad.
  • Some suggest such benchmarks should come from neutral or nonprofit entities.

Real‑world behavior: successes and failures

  • Multiple anecdotes:
    • Serious hallucinations (invented cancer on a lab report, misdiagnosed anemia vs thalassemia) and generic, outdated advice (e.g., low‑fat diets).
    • Strong positive cases where o3/o3‑deep‑research gave plausible diagnoses, timelines, and rehab plans that matched or surpassed prior human input.
  • Users highlight confusion over which model ChatGPT is using (4o vs 4o‑mini), and how “normies” can’t be expected to understand model quality differences.

Use cases, benchmarks, and system design

  • Some want a benchmark focused narrowly on diagnosis (symptoms + history → ground‑truth diagnosis).
  • Others question benchmark realism since real deployments often wrap base models with RAG, guardrails, and workflows; counterpoint: this setup accurately reflects “people chatting to ChatGPT.”

Healthcare economics, access, and substitution

  • Strong sentiment that many simple cases (e.g., cough medicine prescriptions) could be safely handled by AI, reducing unnecessary visits and costs—especially in systems with severe doctor shortages.
  • Others respond that expertise matters precisely in non‑obvious cases, and that patients can’t reliably tell “simple” from dangerous.
  • There is concern that AI will be used to justify shifting more responsibility to less‑qualified staff while maintaining prices, exacerbating profit extraction rather than lowering costs.

Safety, liability, and regulation

  • One side: LLMs are pseudo‑random text machines that hallucinate and should not be trusted for health advice; this “insanity” must be tightly regulated.
  • The other side: human clinicians are also biased, overworked, and fallible; a careful human–AI synthesis could outperform either alone if properly regulated and benchmarked.
  • Debate centers on acceptable tradeoffs: saving time and money for many vs the risk of missed serious diagnoses, and how to quantify those tradeoffs.

Doctors using AI vs replacement fears

  • Some report doctors already using ChatGPT to look up guidelines and organize thinking, seeing it as an extension of their judgment, not a replacement.
  • Others worry that institutions will treat AI outputs as authoritative, degrading human judgment and using the tech to justify staff downgrading.

Miscellaneous

  • Several nitpick the “worst‑case at k samples” chart as visually confusing due to nearly identical colors.
  • One commenter laments lack of Greek language support despite Greek roots of much medical terminology.

I hacked a dating app (and how not to treat a security researcher)

Security tools and reverse engineering

  • Several comments highlight Charles Proxy as a standard, widely used tool for intercepting and reverse‑engineering mobile app traffic (akin to IDA Pro for binaries).
  • Certificate pinning is mentioned as the main barrier to using these tools on modern apps.
  • Some readers discover Charles for the first time and share their own MITM setups for inspecting or isolating devices.

Company response and disclosure duties

  • Many see the company’s initial engagement (meeting, fixing the bug) but subsequent silence as an attempt to “push it under the rug.”
  • Strong view that users should be notified, especially given the sensitivity of leaked data (passports, sexual preferences, chats, phone numbers, location).
  • Others argue the company’s only real duty is to fix the issue; informing the researcher or “the public” is framed by some as optional, though others point out breach‑notification laws likely apply.
  • There’s concern that weak or absent penalties make this “business as usual.”

Legality and risk for security researchers

  • Multiple commenters note that what the researcher did is likely illegal in many jurisdictions once they started enumerating and accessing other users’ data.
  • The Auernheimer/AT&T case and similar prosecutions are referenced as cautionary examples; intent and what data is stored or disclosed matter a lot.
  • Some advocate 90‑day public disclosure deadlines after being ignored; others warn this is a good way to get sued or criminally charged and stress getting legal advice and minimizing data collection.

Technical failings of the app

  • Returning the OTP in the API response is widely ridiculed as “wild” and symptomatic of having no security model or treating the client as trusted.
  • Likely cause is seen as naive scaffolding: serializing DB models directly to JSON, returning created rows verbatim, or leaving in test conveniences.
  • Simple ID enumeration and lack of proper access controls are noted as extremely basic, preventable mistakes, especially egregious for a dating app holding passport images and intimate data.

Responsibility, competence, and “student app” debate

  • Some argue the developers are students/junior and shouldn’t be judged as harshly as big, well‑funded companies that do similar or worse.
  • Others vehemently reject this: if you handle high‑risk PII (passports, sexuality, intimate chats), you have no excuse not to understand basic security—or you shouldn’t build or launch the product.
  • Broader discussion emerges about “move fast and break things,” shipping POCs to production, and weak organizational prioritization of security versus features and timelines.

Regulation, penalties, and professionalization

  • Many call for stronger regulation and real financial or legal consequences for mishandling PII (GDPR is cited as a partial deterrent; US law seen as weaker).
  • Suggestions include: large fines, breach‑reporting requirements with teeth, even treating PII like “nuclear waste” with near‑existential penalties after leaks.
  • A substantial subthread debates licensing or professionalization of software engineers (analogies to civil engineering, food safety); others worry about over‑regulation, gatekeeping, and unintended harm to open‑source and small developers.

Platforms, users, and systemic issues

  • Apple’s app review is criticized as “security theater”: it doesn’t and realistically can’t vet backend security, but its walled‑garden image may give users false confidence.
  • Some argue users should be more cautious about giving such apps sensitive data; others push back that blaming users is unfair and systemic protections and enforcement are needed.
  • Anecdotes from other insecure apps (e‑commerce, dating, even government systems) reinforce that similarly egregious flaws are common and often quietly patched without user notification.