Hacker News, Distilled

AI powered summaries for selected HN discussions.

Page 333 of 535

Curtis Yarvin's Plot Against America

Perceived Authoritarian Drift & “Plot Against America”

  • Several commenters see Yarvin’s ideas reflected in current U.S. trends: normalized political violence, stronger “unitary executive,” targeting of dissidents, and ambitions to hollow out the federal state in favor of state–corporate power blocs.
  • Some argue Trump is only a “vibes coup”; the real danger is a future, more competent authoritarian who inherits the altered norms and legal tools.
  • Others think this is alarmist or analytically shallow, arguing his premises have already failed predictive tests.

Platforming vs. Ignoring Yarvin

  • There’s tension between “don’t amplify cranks” and “ignoring cranks got us here.”
  • One side says giving him New Yorker treatment legitimizes him and wastes attention; the other says once his ideas influence people in power, mainstream scrutiny is a civic duty.
  • Multiple people stress that debate itself doesn’t “legitimize” ideas; suppression by omission is seen as both unrealistic and dangerous.

Influence on Tech & Current Politics

  • Commenters highlight ties to wealthy tech figures and to a sitting vice president who has publicly praised Yarvin and borrowed his rhetoric about purging the civil service.
  • Project 2025 and related plans to remake the executive branch are repeatedly linked, by commenters, to Yarvin-ish blueprints for “technological fascism.”

Assessment of His Ideas & Ideology

  • Harsh view: his political program is neo-feudal, unoriginal (likened to old technocracy), internally inconsistent, and morally abhorrent (race science, genocide talk, “states own citizens”).
  • Milder critics say his diagnoses of how power actually works and of democratic failure are sometimes insightful, but his solutions are teenage thought experiments.
  • A few defend him as “thought-provoking” and worth reading to stress-test democratic assumptions, while explicitly rejecting his conclusions.

Urbit & Technocracy as Mirrors

  • Urbit is cited as a technical analogue of his politics: grandiose re‑invention of everything, extremely complex, few real problems solved, but impressive-looking to non-experts.
  • Broader critique of technocrats: confusing aesthetic neatness for real-world effectiveness; “Brasilia in Substack form.”

Cultural & Meta Notes

  • Cyberpunk and Stephenson’s worlds (Snow Crash, Diamond Age) are invoked as dystopias misread as aspirational—parallels drawn to Yarvin fans.
  • Some lament that this kind of far-right guru gets major-media oxygen while socialist or egalitarian alternatives remain marginal.
  • On HN itself, users are split between flagging the thread out of exhaustion and insisting that, given his proximity to power, discussion is necessary.

Apple Notes Expected to Gain Markdown Support in iOS 26

Markdown support: export vs input

  • Initial reporting implied full Markdown input; commenters noted the underlying source only mentioned Markdown export.
  • The article was later corrected to clarify it’s export-only, disappointing those who want to write in Markdown or edit notes as plain text.
  • Export to Markdown is still seen as a big win for moving content into tools like Obsidian or Joplin, especially if indentation and attachments are preserved; some are skeptical Apple will cover all metadata and media.

Editing experience: Markdown vs rich text

  • Some argue Markdown typing is slower on touch keyboards due to symbol entry; others say it’s faster than diving into the formatting UI, especially with physical keyboards.
  • Selection and formatting of already‑typed text on iOS is widely described as “fiddly,” so adding Markdown markers after the fact may be easier than using rich‑text tools.
  • There’s a clear split: some love Markdown and even mentally “render” it; others find raw Markdown ugly (especially inline URLs) and want rich text, or at least syntax that auto‑hides like Bear’s implementation.

Data portability and lock‑in

  • Strong warnings against deeply investing in Apple Notes because of opaque storage, weak export, and iCloud dependence.
  • Workarounds mentioned: IMAP‑backed notes (with limited feature support and potential corruption between clients), AppleScript tools on macOS, Shortcuts‑based exporters, and third‑party apps that bulk‑export to Markdown.
  • Some users accept lock‑in and treat Notes as a scratchpad for throwaway or temporary content; important material is moved to plain‑text/Markdown systems elsewhere.

iOS versioning change

  • Many were initially confused by “iOS 26,” assuming it was a joke or typo, then realized Apple is rumored to be moving to year‑style version numbers.
  • CalVer is seen as more understandable for non‑experts and helpful in aligning Apple’s OS releases, though some view it as marketing‑driven (and possibly influenced by competitors’ naming).

Markdown’s broader rise

  • Discussion connects Apple’s move and Microsoft Notepad’s new formatting/Markdown support to Markdown’s mainstreaming.
  • Several note Markdown’s importance as a “native language” for LLMs and its de facto dominance among lightweight markup formats, despite many dialects and alternatives (Org, AsciiDoc, LaTeX, etc.).

Perceptions of Apple Notes quality and scope

  • Users appreciate Notes’ simplicity, E2E encryption, and attachment support, but complain about bugs (cursor jumps, disappearing text, table glitches) and poor search.
  • Some want Notes to stay simple and free of LLM features, others find it too limited and migrate to Bear, Obsidian, Upnote, or Org‑based tools.
  • A recurring theme: Apple ships minimal features, leaves apps stagnating for years, then slowly backfills obvious capabilities like Markdown export.

A proposal to restrict sites from accessing a users’ local network

Security motivation and threat models

  • Many commenters support restricting public sites from talking to localhost or private IPs, calling current behavior “insane” given:
    • Web pages can already issue HTTP requests to LAN devices and localhost (including no-cors “simple” requests, <img> and <form> CSRF, WebSockets, DNS rebinding, timing attacks).
    • Examples discussed: Meta/Facebook localhost tracking IDs, Zoom’s localhost server zero‑day, routers/printers/dev servers with unauthenticated or weakly checked endpoints, Bitcoin or JSON‑RPC APIs with poor Content‑Type checks.
  • Clarifications: CORS mainly governs whether JS can read responses; many “simple” requests still reach local devices and can change state or trigger exploits even if the response is opaque.

Legitimate use cases that might break

  • Wide range of existing patterns rely on browser→local HTTP:
    • Desktop helpers behind web UIs (password managers, CAD/3D mouse support, 3D printers, Plex‑style dashboards, enterprise agents, OAuth redirect listeners on localhost).
    • IoT and NAS/router management UIs using a single cloud page that CORS‑fetches metrics directly from multiple local devices.
    • Local signing with smartcards/e‑IDs, local LLM/AI servers, Jupyter, test report viewers, LAN file‑sharing tools (e.g. PairDrop).
  • Some argue these are better implemented as:
    • Browser extensions with native messaging, custom protocol handlers, or fully local/static UIs.
    • UIs hosted on the device itself; but others describe severe pain around HTTPS, certificates, and “secure context” APIs for devices without public domains.

Permissions, UX, and granularity

  • Many welcome per‑site browser prompts (“this site wants to access devices on your local network”) similar to camera/mic.
  • Others think prompts are largely ineffective, citing macOS/iOS/UAC patterns where users reflexively click “Allow,” especially when sites instruct them to do so.
  • Concerns that the proposed permission is coarse‑grained (all local hosts/ports) and violates least‑privilege; suggestions include:
    • Firewall‑like controls (CIDRs, ports, localhost vs LAN vs VPN) and possibly mDNS‑backed, human‑readable device names in prompts.
    • API surfaces for extensions to enforce more detailed policies (some of which were weakened by Chrome’s Manifest V3).

Definitions of “local” and IPv6 / enterprise edge cases

  • Debate over using RFC1918 as “local”:
    • Enterprises often run massive 10/8 networks across many hops; some clients have public IPs but access RFC1918 services via VPN or site‑to‑site tunnels.
    • CGNAT and IPv6 (global addressing with ULAs/link‑local) blur local vs remote; some say there is no robust way to infer “site‑local” from an IPv6 address alone.
  • Several prefer admin‑configurable policies (group policy CIDR lists), or basing decisions on origin+destination categories rather than raw address classes.

Trust in Google and standards politics

  • Significant skepticism that Google’s motives are purely security‑driven; some see this as raising the bar for self‑hosted/local solutions while Google’s own tracking and ads remain untouched.
  • Others counter that the threat (local probing, tracking, and exploitation) is real and that a vendor‑neutral spec any browser can adopt is still valuable despite Google’s market power.

Alternatives and current mitigations

  • Users mention:
    • uBlock Origin’s “Block outsider intrusion into LAN” list, Firefox’s Behave extension, legacy IE security zones, Safari on iOS already asking per‑site for local access.
  • Some advocate going further and generally clamping down on cross‑site requests, pushing aggregation back to servers instead of browsers.

Mistral Code

Product clarity and availability

  • Several commenters couldn’t tell from the landing page whether Mistral Code is a CLI, IDE plugin, or standalone IDE; the VS Code extension link is buried and then leads back to “enterprise only.”
  • The official FAQ confirms: Mistral Code is currently a premium feature only available to Enterprise customers with an active agreement and allocated “seats.”
  • Many developers say they’d like to try it personally but are blocked by the enterprise gate, making the HN launch feel pointless for them.

Enterprise-only and “contact us” pricing

  • Strong pushback on “contact us” pricing, especially with a form asking for company size and revenue; some call it “dead on arrival” for individual developers and small companies.
  • Others argue this model is standard and effective for high-ticket enterprise software, where deals are in the tens or hundreds of thousands and require contracts, security reviews, and ongoing relationship management.
  • There’s debate over whether bespoke pricing is primarily about value-based, complex deployments vs. extracting maximum money and filtering out “small potatoes.”

Target market and competitive context

  • Several infer Mistral is targeting large, compliance‑sensitive orgs (e.g., EU banks) that need EU-hosted/self-hosted solutions and are wary of US providers.
  • Critics note Mistral’s models are not generally seen as SOTA and suggest that if they were, transparent pricing and self-serve trials would be easier to justify.
  • Others point out that Copilot (with free quotas), Zed+Ollama, and various open agents already provide similar or better experiences, often locally and cheaply.

Fork of Continue and open‑source monetization

  • The VS Code listing states Mistral Code Enterprise is a fork of Continue, with credit given.
  • Some say you can largely replicate Mistral Code by configuring Continue with Mistral models (e.g., Codestral/Devstral) and possibly fine‑tuning via Mistral’s API.
  • This triggers a broader discussion:
    • Monetizing Apache/MIT/BSD-licensed projects is legally fine; concerns are moral/economic, not legal.
    • Others recommend copyleft (GPL/AGPL), MPL, BSL, or dual licensing if you don’t want big companies to capture all the value.
    • There’s frustration that permissively-licensed projects get commercialized by better-resourced companies while original authors see little financial benefit.

Local vs hosted assistants and future of this niche

  • Some believe IDE assistants will trend toward “zero-cost and local” using open models and local runners; competition against that with locked-down, remote enterprise products may be difficult.
  • Others counter that many enterprises explicitly want managed, centrally controlled solutions and will accept “contact us” friction to get compliance, security, and support.

Tesla shows no sign of improvement in May sales data

Investor Focus & Valuation

  • Several comments argue current shareholders largely ignore weak sales and are pricing in speculative future cash flows from robotaxis, Optimus robots, and “largest AI project” narratives.
  • Others say this is pure faith/FOMO: market cap and P/E are seen as meme‑stock levels, disconnected from car‑company fundamentals.
  • Some suspect institutional investors are quietly exiting while retail and ETFs hold the bag; a few even flirt with “Enron 2.0”–style conspiracy theories, while others push back that conspiracies are overused explanations.

Robotaxis, FSD, and Safety

  • Many are deeply skeptical that Tesla’s robotaxi plans justify its valuation, even if technically successful; comparisons to Uber’s thin margins are common.
  • Waymo is repeatedly cited as the benchmark: millions of paid driverless rides, lidar-based sensing, few major safety scandals.
  • Tesla’s camera‑only approach is criticized as inherently riskier; some suggest Musk can’t admit lidar might be necessary without massive retrofits.
  • Owners are split: some say FSD now drives long distances with almost no interventions; others report it struggles on nonstandard roads and see it as far from the promises (e.g., coast‑to‑coast drives, 1M robotaxis by 2020).
  • Calls for independent, third‑party safety validation instead of Tesla‑supplied statistics are strong.

Build Quality, Reliability, and Design

  • Numerous references to “ridiculous” manufacturing defects, longstanding QC problems, and poor reliability surveys; Cybertruck is cited as a particularly botched launch.
  • Critics say interiors have been cheapened and UX degraded: loss of stalks, everything on a central touchscreen, awkward door handles, and minimal physical controls viewed as cost‑cutting, not elegance.
  • Some disagree, praising the 3/Y as comfortable, simple, and cheaper than German EVs with similar features.

Competition and Alternatives

  • Many point out that Tesla’s early technical lead has eroded: Hyundai/Kia, Ford (Mach‑E, F‑150 Lightning, future models), GM, Nissan, Toyota, and especially Chinese makers like BYD/NIO now offer compelling EVs.
  • The shift to NACS by other automakers is seen as removing Tesla’s Supercharger moat.
  • Some buyers are explicitly choosing hybrids (e.g., Prius) or non‑Tesla EVs for cost, reliability, or road‑trip economics.

Politics, Musk, and Brand Toxicity

  • A major thread is political: multiple commenters say they won’t buy or will sell Teslas due to Musk’s behavior, perceived authoritarian and extremist sympathies, and heavy entanglement with government contracts.
  • Others lament “giving up high‑quality EVs over politics,” but get strong pushback that supporting or boycotting a CEO is a legitimate factor once products are no longer uniquely superior.
  • Several owners report they like their cars but will never buy another; some sold solely to avoid association with the brand.

Consumer Behavior & Market Trajectory

  • Anecdotes suggest a growing cohort of ex‑or‑current owners moving to Hyundai, Toyota, etc., and friends who abandoned Tesla purchase plans.
  • Some still praise Teslas as “transformational” and hold the stock, hoping for long‑term upside.
  • Overall sentiment in the thread leans toward: once‑innovative product, now facing stronger competition, quality issues, and severe reputational damage that may not yet be fully reflected in sales or stock price.

We are no longer a serious country

Bond markets, tariffs, and Fed risk

  • Commenters agree bond markets matter more than equities and that policy whiplash can erode confidence in the dollar.
  • Some fear a future purge of Fed leadership and replacement with loyalists, seeing this as a 1930s-style risk.
  • Others note markets have already forced the administration to back down on tariffs more than once, but opponents say the repeated attempts show persistence, not restraint.
  • The “Taco Trade” is cited as an example of traders exploiting tariff-related volatility.

Is the administration “dumb” or strategically ruthless?

  • One camp sees the administration as incompetent, pointing to poorly drafted executive orders, constitutional overreach, and weak implementation capacity.
  • Another argues it is legally aggressive and coherent in its own goals: generating volatility, enriching insiders, expanding foreign-policy tools, and sending hardline messages to allies and rivals.
  • Several emphasize that incompetence and danger can coexist; even “failed” orders still enable abuses by agencies.

Symbolism, bill naming, and seriousness

  • Some dismiss focus on the “One Big Beautiful Bill Act” name as trivial; only legal effects matter.
  • Others see the name as emblematic of shallow branding and a broader lack of seriousness among enablers.
  • There is debate over whether such branding is actually quite effective at agenda-setting and distraction.

US hegemony, rules-based order, and tariffs

  • Critics argue US elites long ago embraced an exceptionalist “rules-based order” that mainly means “the US makes the rules,” with diplomacy hollowed out.
  • Defenders counter that US security spending and interventions (e.g., protecting shipping lanes) deliver global public goods and soft power.
  • Tariffs are viewed both as a dangerous form of saber-rattling that could undermine reserve-currency status and as a potent diplomatic lever.

Red vs blue politics and voter grievances

  • There is disagreement over how much current policy is “red state,” but many see strong overlap on social conservatism, deregulation, tax cuts, and hostility to federal bureaucracy.
  • Multiple comments probe why poorer red-state voters support such policies:
    • Some emphasize emotional grievance politics, scapegoating of out-groups, and delight in “punching down.”
    • Others note shared left–right grievances about elites, inequality, and loss of control, even if right-wing policy responses worsen those problems.
  • Data and anecdote alike are cited that red states generally have worse socioeconomic outcomes, yet out-migration toward them suggests their policy mix still attracts many.

Interpreting the economic signals and Krugman’s framing

  • Some see the divergence between interest rates and the dollar as historically worrying and consistent with the article’s thesis of declining seriousness.
  • Others accuse the article (and the Financial Times chart it references) of cherry-picking time windows; longer-range plots paint a more ambiguous picture.
  • A recurring frustration is the mix of valid concerns with rhetorical flourishes (e.g., title, bill name) that feel more like partisan doomsaying than sober risk assessment.

Decline, rhetoric, and public disengagement

  • Historical declines (Venezuela, Argentina, Portugal, Rome) are invoked to argue rich countries can indeed fail and the US might be on that path.
  • Another group criticizes absolutist claims like “no longer a serious country” as lazy, ahistorical, and unhelpful for finding solutions.
  • Several comments highlight widespread public detachment from markets and policy, polarized information ecosystems, and the difficulty of finding non-partisan, analytically rigorous guidance on what genuinely matters.

VC money is fueling a global boom in worker surveillance tech

AI, capitalism, and the purpose of surveillance

  • Commenters frame surveillance tech as an old trend supercharged by AI, arguing it’s being used to punish and control rather than help workers.
  • Several contrast “AI + socialism” (post-scarcity / The Culture–style utopia) with “AI + capitalism” (techno-feudalism, Manna-like micro‑management).
  • A minority suggest surveillance is a pragmatic response to eroded norms of duty and self‑accountability; others strongly counter that productivity is already high and stagnating wages, not “counterculture,” explain worker disengagement.

Global scope, labor rights, and sector differences

  • Some say strong labor protections (EU, Switzerland) make such tech largely irrelevant or illegal, but others insist “bossware is everywhere,” just with different legal constraints.
  • Academic, research, and healthcare jobs are seen by some as relatively insulated; others report growing monitoring of students, lecturers, labs, and hospitals.
  • Multiple participants say small businesses can be more invasive than large enterprises because owners face less internal oversight.

Business justifications vs critiques

  • Supporters of basic tools (GPS‑based clock-in, identity checks) argue they’re vital for small, low‑margin businesses and remote hourly work, especially outside rich countries.
  • Critics respond that constant monitoring reflects managerial failure: if performance changes can’t be seen via outcomes, the business model or management is broken.
  • There is debate over whether identity verification is “surveillance tech”; some see it as neutral infrastructure, others note it underpins blacklists and abusive profiling (e.g., in Mexico or with national ID schemes).

VCs, “Little Tech,” and regulation

  • Many see VC‑backed “worker surveillance” and “Little Tech” narratives as rent‑seeking: calls to relax regulation are read as attempts to exploit loopholes and externalize harm.
  • Others push back that US regulation is already heavy in many sectors; the real VC complaint is about IPO/SPAC rules limiting liquidity and exits.

Social consequences, resistance, and arms race

  • Surveillance is linked to broader wealth concentration, enclosure‑like loss of autonomy, and low‑trust societies.
  • Suggested responses: stronger legal limits (narrow scope, minimal retention, strict sharing rules), refusing to work for surveillance startups, and public awareness.
  • An arms race is noted: monitoring mouse activity vs mouse jigglers, with warnings that such “fraud” can increase employers’ leverage over workers.

IRS Direct File on GitHub

Political fate and lobbying concerns

  • Many commenters say Direct File is being shut down after the 2024 season due to a new funding bill, tying this to long‑standing influence from the tax‑prep lobby (especially Intuit) and anti‑tax activists.
  • Others push back on specific claims (e.g., timelines, donation patterns), but broadly agree Congress, not the IRS, blocked expansion and continuation.
  • There’s debate over partisan responsibility: several blame Republican opposition to easy filing; others argue both parties are susceptible to lobbying and neoliberal privatization.
  • Some note Direct File proved the idea is technically and operationally feasible; ending it is now framed as a political choice, not an impossibility.

Role of Direct File vs. commercial tax software

  • A recurring theme: most Americans have simple returns (W‑2, a few 1099s, standard deduction, common credits) and should not have to pay for filing.
  • Commenters argue the IRS already receives much of the needed data and could pre‑fill returns or at least show taxpayers what it knows and ask for corrections.
  • Opponents highlight missing information (marital status changes, dependents, self‑employment details, state and local taxes, ambiguous deductions) and fear that if the IRS “starts the return,” some people won’t add tax‑increasing items.
  • Some see business opportunities in building on Direct File; others respond that the whole point was to remove the need for such businesses in simple cases.

Technical design and codebase

  • Interoperability is via the existing Modernized e‑File (MeF) API used by commercial providers; several say the hard part was encoding tax law, not wiring the API.
  • The codebase is large, “boring enterprise” JS/TS/Java with Scala for the “Fact Graph” rules engine; some praise its cleanliness and extensive design docs.
  • The Fact Graph models tax rules as a DAG of facts and derived values, good for incomplete information but limited where the law is circular or interpretive.
  • A deeply nested reactive Java controller sparked criticism as hard to read and test, with some blaming misuse of reactive frameworks.

Open source, licensing, and trust

  • As federal work, the code is in the public domain; this is welcomed as a potential baseline for competitors, but it also means anyone can redistribute modified versions.
  • Some wish for commit signing and a verifiable chain of custody; others note the GitHub mirror is just a scrubbed dump and that public‑domain status weakens the value of signature semantics.

Broader tax‑system reflections

  • Multiple commenters compare to countries where pre‑filled returns take minutes, viewing the US system as deliberately complex to support private intermediaries and maintain tax “pain.”
  • Others note that instructions and paper forms are usable for many filers, but software shines at discovering which forms and edge cases apply.

Prompt engineering playbook for programmers

Confusion, Non-determinism, and Inconsistency

  • Several programmers describe LLM behavior as disorienting: small wording changes yield different answers, results vary run-to-run, and model updates break previously “good” prompts.
  • This randomness makes it hard to justify heavy upfront “prompt engineering” as a reliable process.

Utility vs Overhead for Programmers

  • Many say they’d rather just write the code: by the time a long prompt is crafted, refined, and run, the task could be done manually.
  • LLMs are seen as most useful for quick documentation lookup, simple refactors, boilerplate, and debugging when you paste actual code or stack traces.
  • Agentic tools are criticized for burning money/tokens and sometimes mangling codebases.

Are Prompt Guides Really Needed?

  • One camp: guides are mostly hype; you learn “prompt fu” quickly through use, similar to early “Google-fu” books.
  • Another camp: many users won’t systematically experiment, so curated tips and watching experts can meaningfully improve results and reveal non-obvious tricks.

Long vs Short Prompts, Context, and Iteration

  • Multiple commenters report that shorter, focused prompts plus iterative refinement outperform long, intricate ones.
  • Irrelevant detail is seen as harmful; relevant, structured context (markdown specs, project structure, style guides) can help.
  • Some feel verbose prompts just get echoed back and give a false sense of control.

Specific Prompting Techniques Discussed

  • Widely cited “real” techniques:
    • In-context examples (few-shot).
    • Chain-of-thought / step-by-step reasoning (less critical on newer “reasoning” models).
    • Structured output (JSON or schemas, sometimes via two-stage prompts).
    • Role / framing prompts (expert vs junior, critic vs collaborator).
  • Using domain-specific jargon and expert vocabulary can noticeably change answers.

Debate Over “Prompt Engineering” as a Concept

  • Strong pushback against calling this “engineering”: it’s described as trial-and-error, craft, or just clear technical writing/communication.
  • Others argue that once you systematically test, version, and benchmark prompts—especially in pipelines and API workflows—it starts to resemble engineering practice.
  • There’s concern that the focus on prompting is displacing core programming skills, though some see “native language as a programming language” as an inevitable shift.

Limits and Need for Better Tooling

  • Consensus that if a task is beyond current LLM capability, no prompt will fix it; you must decompose into subtasks.
  • Some argue that IDEs/agents should handle low-level prompt optimization, with developers focusing on clear requirements and tests rather than micromanaging prompts.

FFmpeg merges WebRTC support

What was added

  • FFmpeg’s libavformat now includes a WHIP muxer, enabling it (and apps using its libraries) to send media via WebRTC.
  • WHIP (WebRTC-HTTP Ingestion Protocol) = push media into a WHIP gateway that handles full WebRTC (ICE/STUN/SCTP, etc.) to endpoints.
  • It does not yet implement WHEP (the receiving/egress side) and does not expose the low-level SCTP datachannel parts of WebRTC.

Practical use cases and impact

  • Any FFmpeg-based tool (OBS, GStreamer, custom apps, Unreal streaming setups, etc.) can now ingest into WHIP-enabled services (e.g., some CDNs, Twitch, Cloudflare, LiveKit, Dolby).
  • Enables low-latency browser consumption of streams without users opening ports, e.g.:
    • Remote desktop / remote control, drones, robots.
    • Art installations / city “portals” with near real-time audio/video.
    • Recording or transcoding live WebRTC streams directly with FFmpeg CLI or libraries.
  • Seen as a major step toward a ubiquitous open broadcasting protocol across desktop, mobile, web, and embedded platforms; potentially lowers cost vs traditional transcoding-heavy setups, especially with simulcast and modern codecs.

Latency, protocols, and alternatives

  • WebRTC is praised for sub-second latency over the open internet, especially under loss, where UDP+FEC/NACK can outperform TCP.
  • Some point out that on clean LANs, very low latency is also possible with plain TCP streams; browsers’ HLS-centric ecosystem is the real barrier.
  • Discussion of HLS vs LL-HLS and why the industry ended up using complex WebRTC for low latency.
  • Comparisons with SRT; some users struggle to get SRT under ~1s, while WebRTC can reach ~100 ms on LAN.

Security and hardening

  • Concern that adding WebRTC increases FFmpeg’s attack surface, referencing many WebRTC-related CVEs in browsers.
  • Others argue most CVEs are implementation-specific (browsers, libvpx), not inherent to the protocol; FFmpeg’s WHIP piece is small and narrow.
  • General consensus: FFmpeg is already sensitive software; treat it as untrusted when handling user input:
    • Run in containers or VMs, use seccomp, minimize enabled muxers/decoders (e.g., --disable-muxer=whip if unused).

Implementation status and project process

  • Building requires --enable-muxer=whip and --enable-openssl; some got 500 errors with certain providers until they added an audio stream.
  • There is interest in future WHEP and possibly more complete peer-side functionality (ICE, SCTP, datachannels), but that’s not standardized or exposed yet.
  • Some FFmpeg developers criticize the merge process: a single large squashed commit, limited tests, and missing features like NACK support; concern about half-finished vendor-driven features entering tree before being production-ready.

AGI is not multimodal

Scope and meaning of AGI and “intelligence”

  • Many argue “AGI” is poorly defined or even meaningless; likewise for “intelligence” itself, since there is no agreed test that definitively requires it.
  • Some suggest humans aren’t truly “general” either, since each person only covers a subset of domains, yet the human brain architecture is general and can be specialized.
  • Others propose a practical definition: an architecture that can be cheaply copied and fine‑tuned across many tasks is “functionally AGI,” in which case current large models may already cover a big fraction of what’s needed.

Embodiment, world models, and the physical environment

  • A major thread supports the article’s claim that general intelligence requires being situated in and acting on an environment (physical or rich digital), learning through interaction and consequences.
  • Examples: self‑driving failures where “everyone just knows” the dynamics of vehicles; developmental psychology notions of enriched environments, episodic memory, and executive control.
  • Some broaden embodiment to any agent loop with perception and action, including purely digital office or simulation environments, while others insist physical reality is uniquely constraining and essential.

Multimodality and senses

  • Many commenters say AGI “must” be multimodal (at least vision, audio, etc.), but disagree on whether it needs all human senses (smell, taste) or human‑like ranges.
  • Disability analogies (blind or deaf humans) are used on both sides: to argue senses are not essential to intelligence, or to argue they shape the scope and kind of understanding.
  • There is interest in non‑human modalities (infrared, EM fields) and the idea that more modes increase the ability to correlate and generalize.

LLMs, next‑token prediction, and world models

  • One side criticizes the article’s framing of LLMs as “just next‑token predictors” and notes that this interface does not constrain internal computation; long coherent outputs imply substantial internal planning and semantic structure.
  • Others stress that current models operate over symbol statistics, not grounded experience; coherence and “reasoning” may be sophisticated mimicry of human text rather than learned causal world models.
  • Debate continues over whether semantics can ultimately “reduce” to patterns over symbols, or whether sensorimotor grounding and continuous experience of time are indispensable.

Research directions, capital, and practicality

  • Some recount attempts at embodied AI (e.g., robotics startups) that burned large sums without clear success, contrasting that with immediate commercial value of LLMs, which pulls funding away from harder embodiment work.
  • A few see the paper as mostly re‑stating that “we need to understand how to build AGI to build AGI,” with little concrete technical guidance; others find it a valuable synthesis that pushes beyond naive multimodal “model‑gluing” and scale‑only strategies.

The Right to Repair Is Law in Washington State

Scope of the Washington law & major carve‑outs

  • Many welcome the shift in regulatory “winds” but note the law excludes many high-impact categories: game consoles, cars, agricultural/construction equipment, medical devices, ISP equipment, security systems, off‑road vehicles, large energy/solar systems, and LEO broadband gear.
  • Several commenters say these carve‑outs cover “most of what I actually want to repair,” especially tractors, road vehicles, and consoles.
  • Carve‑outs are widely attributed to lobbying and corporate pressure; some point out this is especially ironic given tractors and autos helped spark the movement.

What the law does change (electronics & parts pairing)

  • For covered consumer electronics and wheelchairs, manufacturers must provide parts, tools, documentation, and can’t use parts‑pairing to:
    • Block repairs or degrade features (e.g., cameras, fingerprint sensors).
    • Show non‑dismissable, misleading “unknown part” warnings or reduce performance.
  • Debate over a secondary source’s claim that parts must be sold “at cost”: the statute actually says “at costs that are fair to both parties.”
  • Long thread on economics of spares (tooling, logistics, low volume pricing) vs the need to prevent gouging (e.g., tying parts prices to full device value).

Wheelchairs and disability impacts

  • Strong support for the wheelchair provisions from people citing extremely locked‑down, expensive chairs (e.g., batteries and simple components requiring factory service, five‑figure subframes).
  • Others worry manufacturers may exit the WA market or raise prices, reducing options for users whose devices are often funded by government programs.

Security, anti‑theft, and “genuine” parts

  • Examples: Apple serial‑paired MacBook components; TPMs and fingerprint sensors; vehicle ECUs.
  • One camp argues pairing is needed to deter theft and supply‑chain attacks; critics say users could hold the keys, and current designs mainly protect OEM repair revenue.
  • Ongoing debate over quality and safety of non‑OEM parts vs monopolistic pricing on “genuine” components.

Broader right‑to‑repair context and limits

  • Clarifications for newcomers: right‑to‑repair addresses active barriers—DRM, unavailable parts/docs, voided warranties, and bricking—rather than knowledge of how to fix things.
  • DMCA §1201 is cited as a major remaining obstacle (e.g., third‑party generator and vehicle tools becoming legally risky).
  • Some fear manufacturers will region‑limit compliance or use contract terms to claw back control; others argue once a law exists, it’s easier to tighten loopholes and attack exemptions later.

"AI Will Replace All the Jobs " Is Just Tech Execs Doing Marketing

Limits on “AI Will Do All Jobs”

  • Many argue current AI lacks reliability, embodiment (“no thumbs”), accountability, and true understanding, so can’t fully replace complex or physical work (e.g., toilets, construction, medicine).
  • Several note strong human preference for human-made things (live music, handmade food, human therapists), predicting markets and platforms will enforce “human-only” spaces.
  • Others stress AI is parasitic on human-created data; if humans stop producing, models stagnate.

Degradation of Work vs Replacement

  • A recurring fear: AI won’t remove jobs outright but will deskill them and make them worse—workers become “meat robots” monitored and directed by algorithms.
  • Existing examples cited: warehouse workers, call centers, and fast-food “management AIs,” where computers already dictate pace and behavior.
  • Senior engineers worry about becoming AI babysitters/code reviewers instead of creators or mentors; some find that dystopian, others enjoy delegating boilerplate.

Juniors, Deskilling, and Partial Job Loss

  • Many expect fewer entry-level roles: seniors + AI replace juniors; long-term this hollows out expertise pipelines.
  • Historical analogies: barcodes and scanners deskilled retail clerks; more jobs remained but became lower-paid “McJobs.”
  • Several see “some but not all” job loss as the worst outcome: widening inequality without a clear crisis to force systemic reform.

Economic and Political Context

  • Strong thread framing AI as capital vs labor: a tool to replace labor with capital, weaken bargaining power, and justify layoffs.
  • Skepticism that displaced workers will easily “reskill,” citing past automation where many never recovered comparable livelihoods.
  • Concerns that without stronger safety nets (UBI, healthcare), AI-driven displacement will fuel social unrest.

Hype, Exec Marketing, and Uncertainty

  • Many view “AI will end all jobs” as stock-pumping and layoff cover rather than current reality; examples of firms loudly touting AI while cutting staff.
  • Others argue exponential compute and recent LLM gains make human-level or superhuman AI plausible, but timelines are contested and often likened to a “religious” debate.
  • Consensus: AI is already powerful enough to affect most businesses, but claims of near-term total replacement are unproven.

AI as Tool Today: Benefits and Risks

  • Working developers describe real productivity gains (boilerplate, queries, refactors) but still needing domain expertise, design judgment, and security review.
  • Worries about security: hallucinated libraries, subtle vulnerabilities, and supply-chain attacks seeded via AI-generated code.
  • Some non-experts already ship AI-written scripts into sensitive workflows, raising hidden risk even where no jobs are formally cut.

The time bomb in the tax code that's fueling mass tech layoffs

Scope of the problem: does Section 174 “explain” mass layoffs?

  • Many argue Section 174 cannot fully explain mass tech layoffs:
    • Most laid‑off staff weren’t doing what people normally call “R&D” (e.g. games, line‑of‑business apps, ops).
    • Other major drivers cited: end of zero‑interest‑rate era (ZIRP), pandemic over‑hiring, slower growth, Twitter’s high‑profile cuts, and classic wage‑suppression tactics.
  • Others counter that, for tax purposes, a huge share of tech headcount is classified as R&D, especially since the code now explicitly treats software development as R&D. So the rule change does affect a large fraction of engineers, PMs, data scientists, etc.

What actually changed in Section 174?

  • Pre‑2022: qualified R&D (including software dev if elected) could be 100% expensed in the year incurred.
  • Post‑change: “specified research or experimental expenditures” must be amortized:
    • 5 years for US R&D, 15 years for foreign R&D.
    • Software development is explicitly deemed R&D; most dev labor now becomes capitalized, not immediately deductible.
  • Transition effect: companies went from deducting 100% to deducting only ~10–20% in year one, with the rest spread over later years.

Cash‑flow mechanics and who gets hurt

  • Several concrete examples show how a company can have zero (or negative) cash profit but be taxed as if it were strongly profitable, because only a fraction of dev salaries is deductible in the current year.
  • In steady state with flat headcount, total deductions over time are similar; the real costs are:
    • Time value of money (you’re effectively lending the IRS cash at 0%).
    • A multi‑year “tax spike” during the transition and during rapid growth.
  • Consensus in the thread:
    • Big, stable, cash‑rich firms (FAANG‑type) are affected but can absorb it; many were already capitalizing dev work.
    • Small, thin‑margin or modestly profitable software firms, and fast‑growing startups, are hit hardest; some report tax bills quadrupling, hiring freezes, or shutdowns.

Is this good policy? Competing views

  • One side: this simply closes a generous R&D loophole; software is a capital asset like a factory, so its creation costs should be amortized; subsidies distort markets and shift burden to taxpayers and non‑R&D firms.
  • Other side: historically, immediate expensing was a deliberate pro‑innovation policy; removing it:
    • Discourages R&D and early revenue, especially in software where code ages quickly and many firms never reach year 5.
    • Entrenches incumbents and weakens the startup ecosystem and open‑source work.

Politics and “time bomb” framing

  • Change came from the 2017 tax act, effective 2022, widely viewed as a budget‑scoring gimmick to offset corporate rate cuts.
  • There is broad bipartisan interest in restoring immediate expensing, but fixes have repeatedly stalled and are now bundled as a temporary rollback inside a larger, controversial tax bill.
  • Some see Section 174 as a deliberate political time bomb and a bargaining chip; others attribute the mess to routine reconciliation games and legislative dysfunction rather than targeted malice.

Go is a good fit for agents

Go’s perceived strengths for agents

  • Many see agents as orchestration-heavy: coordinating HTTP calls, streams, timeouts, retries, backpressure, and cancellation. This is described as “bread and butter” for Go’s goroutines, channels, context.Context, and strong standard library.
  • Static binaries and simple deployment are cited as advantages over Python’s virtualenv/pip complexity for production agents.
  • Some report good experience building real agentic tools/CLIs in Go, and say LLMs generate Go code reliably because the language is small, consistent, and has high‑quality training data.

Elixir/Erlang and the BEAM alternative

  • Several argue that, by the same concurrency logic, Elixir/Erlang (BEAM) are even better for agents: lightweight isolated processes, built‑in distribution, fault tolerance, hot code upgrades.
  • BEAM + durable storage (SQLite/Postgres/S3/DuckDB or mnesia) is proposed as an ideal combo for stateful agents and orchestrators, with examples of job/workflow systems and IoT‑style networks.
  • Some note BEAM’s VM and performance ceiling as drawbacks versus a compiled language like Go.

JavaScript/TypeScript and Python ecosystems

  • There’s strong defense of JS/TS for agents: native JSON handling, event-driven style, npm ecosystem, browser + Node deployment, sandboxing in the browser, and mature MCP tooling.
  • Python is defended as the de facto ML/AI language with unmatched libraries, tooling, and community; many agent frameworks, observability tools, and SDKs appear there first.
  • Others dislike JS/TS or Python on language-design grounds but concede that ecosystem gravity matters in AI more than runtime performance.

Durable execution, task queues, and state

  • Long‑running agents raise concerns about losing in‑memory state on process death. Multiple comments argue you must break work into resumable chunks and persist state in a DB or durable log.
  • Temporal, Hatchet, task queues, and similar durable-execution systems are discussed as language-agnostic solutions that replay workflows from event history.
  • Some Go users are hand‑rolling checkpointed agents using logs, context trees, and databases, but note the complexity overhead.

Performance and bottlenecks

  • Many point out agents spend most of their time waiting on LLM calls and external I/O; runtime speed rarely dominates.
  • JSON (de)serialization and conflict‑resolution (diff/patch/merge on shared state) are mentioned as the main non‑LLM hotspots; language choice only marginally affects these.

Go’s language and ecosystem criticisms

  • Critics argue Go’s type system (enums, generics ergonomics), error handling verbosity, and channel footguns hurt general productivity, not just for agents.
  • Others counter that the simplicity, interfaces, and tooling outweigh these issues and reduce long‑term maintenance risk.
  • Several note that Go’s ML and agent‑orchestration library ecosystem lags far behind Python and TS; you often end up writing your own middleware (logging, tracing, retries, DSLs, SDKs).

“Agents are just software” / language neutrality

  • A recurring view: everything described as “an agent” is standard async/distributed programming—loops, branching, queues, workflows.
  • From that perspective, “best language for agents” largely reflects developer preference; any language with decent async and libraries can work, and language choice rarely determines success.

Porn sites go dark in France over new age verification rules

Perceived harms of youth porn exposure

  • Broad agreement that very young children shouldn’t access porn; disagreement about teenagers.
  • Some think porn is a relatively minor issue vs social media or other risks; others see links (cited in the thread) to earlier/riskier sex, body shame, aggression, and coercion.
  • Several argue “we didn’t turn out fine,” pointing to rising mental health issues and sexual dysfunction, though causality is disputed.
  • Others say teens have always found porn; exposure alone doesn’t necessarily produce extreme behavior or misogyny.

Effectiveness and unintended consequences

  • Many expect teens to bypass blocks via VPNs, mirrors, smaller or offshore sites, or sneakernet.
  • Concern that burdens will fall mainly on mainstream, moderated platforms, pushing minors toward more extreme, unregulated content.
  • Counter‑view: laws don’t need to be airtight; raising friction and delay can significantly cut average exposure.

Privacy, anonymity, and technical models

  • Strong worry about loss of anonymity and creation of databases linking identity and porn use.
  • The “double anonymity” / third‑party verifier model is seen by some as a good compromise; others worry about what the verifier and government can infer in practice.
  • Discussion of privacy-preserving tools: anonymous credentials, zero‑knowledge proofs, eID apps, Privacy Pass‑style tokens. Doubts about maturity, inclusivity, and attack surfaces (e.g., relay/KYC abuse).
  • Some suggest age-flag HTTP headers or RTA-style labelling so client devices/parents can filter, avoiding central ID checks.

Parental responsibility vs societal regulation

  • One camp: it’s primarily on parents; ISPs already age‑gate access, similar to alcohol sales.
  • Others argue parents are outgunned; society routinely uses legal guardrails (tobacco, gambling, driving) and should do so here too.

Nature of porn and changing landscape

  • Claims that modern porn is more violent, incest‑themed, and ubiquitous (phones, streaming) than past eras, making comparisons (“we watched at 13 and were fine”) questionable.
  • Suggestions for “teen‑safe” porn with strict content rules, analogous to softer vs harder drugs/alcohol.

Regulatory design, enforcement, and culture

  • Debate over standards vs detailed rules, risk of selective enforcement, and costs for small sites.
  • Some favor outright online bans or physical “porn rooms”; others see this as overreach and moral panic.
  • A few question why porn is targeted more heavily than graphic violence, and note differing European vs US attitudes toward sex.

Just how bad are we at treating age-related diseases?

Progress vs. stagnation in age‑related disease

  • Commenters agree we’ve dramatically reduced many historical killers (TB in rich countries, hookworm, black lung, infectious diseases, some cancers), but have done poorly on classic age‑related degeneration (Alzheimer’s, Parkinson’s, IPF, macular degeneration, neurodegenerative diseases).
  • Disagreement over terms like “conquered”: some see big mortality drops and early‑stage survival gains (e.g., breast cancer, prostate cancer) as success; others argue that anything still common, lethal, or disabling is not “conquered.”

Treatment vs. prevention and behavior

  • Many argue we’re far better at chronic management than curing, and “successes” often come from prevention and environment (better sanitation, nutrition, shoes, fewer miners) rather than drugs.
  • Strong thread claiming many “age‑related” diseases are largely lifestyle‑driven (diet, exercise, alcohol, obesity, smoking); counter‑arguments say age and biology still dominate and behavior change is extremely hard in practice.
  • GLP‑1 drugs are cited as evidence that behavior is biologically constrained, not just willpower.

Metrics: clinical proxies vs quality of life

  • Frustration that trials focus on surrogate markers (lesion size, biomarkers) rather than quantity and quality of life.
  • Acknowledgment that QoL studies are expensive and slow; proxies are cheaper but often weakly linked to real‑world benefit.

Neurodegeneration and aging biology

  • Alzheimer’s singled out as emblematic failure; mention of past research fraud possibly delaying progress.
  • Neurodegenerative diseases seen as fundamentally unsolved; some view curing them as tantamount to achieving biological immortality.
  • Speculation about aging as a malleable process (e.g., cell reprogramming, long‑lived species), but current interventions like young‑blood transfusions are described as weakly supported and ethically fraught.

Public health, risk behaviors, and new threats

  • Tobacco control cited as a rare behavioral success requiring massive social and political pressure.
  • Debates over vaping and cannabis: how harmful vs cigarettes, dose differences, second‑hand risk, and whether society should treat them equivalently.
  • Concern that eradication strategies (vaccines, PrEP for HIV) run into compliance, education, and misinformation barriers.

End‑of‑life, ethics, and society

  • Several comments highlight the misery of long decline (Alzheimer’s, advanced cancer), aggressive life‑prolonging care with poor QoL, and the appeal of legal assisted suicide.
  • Some see our biggest failure not as medical but philosophical: inability to accept and plan for death.

Broader social and system factors

  • Loneliness, insomnia, polypharmacy, and loss of function are seen as “new” geriatric burdens now that people outlive earlier killers.
  • Observations that national systems differ: public systems may push lifestyle change more, private ones may profit from long‑term treatment dependence.

Why I wrote the BEAM book

BEAM performance, pauses, and mailboxes

  • Commenters debate how a 15ms pause could be “post‑mortem worthy”; those familiar with BEAM note that typical latencies are in microseconds, so a jump to milliseconds can cause massive backlogs.
  • A gen_server is described as effectively a big mutex guarding shared state; if it normally serves a request in ~20µs, a 15ms stall can queue hundreds of messages.
  • Unsafe receive patterns that scan entire mailboxes become catastrophic under backlog, making processing time proportional to queue length.
  • Some systems resort to drastic recovery strategies: dropping entire mailboxes, age‑based request dropping, tuning GC around large mailboxes, and monitoring drain/accumulation rates.

Concurrency model, OTP, and failure handling

  • BEAM’s main concurrency tools noted: gen_server, ETS (in‑memory tables), and persistent_term, plus newer “aliases” to avoid message buildup from timeouts.
  • There’s discussion of where to block work (e.g., letting callers block rather than the gen_server) and how to apply backpressure instead of blindly queueing.
  • Some argue BEAM’s magic is really OTP’s abstractions (supervision trees, processes, fail‑fast semantics), which can be emulated conceptually in other languages, though often without BEAM’s preemptive, lightweight processes.

BEAM vs other runtimes and stacks

  • Historically BEAM was unique; now many problems it solves (message buses, clustering, serialization, orchestration, reliability) have a “zoo” of alternatives: Kafka/SQS/NATS, gRPC/Cap’n Proto, Kubernetes, lambdas, micro‑VMs, etc.
  • Several comments emphasize that BEAM’s 2025 advantage is integration: one coherent runtime with built‑in messaging, supervision, clustering patterns, and a consistent data model, rather than wiring many disparate systems.
  • Others counter that Kubernetes‑style stacks and language‑agnostic infra give more flexibility, and BEAM’s built‑ins (queues, DB, clustering) can be weaker than best‑of‑breed external tools.

Adoption, marketing, and ecosystem

  • Many see Erlang/Elixir/BEAM as highly underrated but note weak corporate backing and marketing compared to Java, Go, Rust, etc.
  • Some say Erlang really shines at very large scales (millions of users), and its “all‑or‑nothing OS‑like stack” plus unusual deployment model (ERTS, epmd, clustering) raises the adoption bar.
  • Others argue modern “standard” languages now handle large concurrency loads on single machines, reducing the perceived need for BEAM.

Experiences with Erlang/Elixir

  • Multiple comments praise BEAM as “alien tech” for fault‑tolerant, concurrent, networked systems; Elixir especially is highlighted for web apps (Phoenix, LiveView) and small teams managing big problems.
  • Some report painful setup and environment drift (especially with LiveView) and prefer containers to stabilize runtime expectations. Others find Elixir deployment straightforward in recent years.
  • Erlang is credited with improving developers’ thinking about immutability, pattern matching, and concurrency, but can make other languages feel clumsy afterward.

The BEAM book and technical publishing

  • The book is open‑sourced on GitHub, with paid Kindle/print versions for those wanting to support the author. Readers welcome deep, implementation‑level documentation, saying official docs are too shallow.
  • Several note that writing a book is a powerful way to truly understand a complex system.
  • Broader discussion covers traditional publishers vs self‑publishing: publishers bring marketing, editing, and print logistics but push for broad, beginner‑friendly topics; niche, deeply technical works increasingly succeed via self‑publishing, LeanPub, and similar platforms.

Cockatoos have learned to operate drinking fountains in Australia

Why cockatoos use fountains

  • Multiple hypotheses discussed: better-tasting “pure” water, elevated vantage point to watch for predators, laziness/convenience, and sheer curiosity.
  • Several commenters argue taste alone is not obvious, noting many pets prefer muddy puddles over clean tap water, suggesting “interesting” or variable water sources may be attractive.
  • Running water is suggested as a general cue of freshness (as with cats), possibly relevant here.
  • Others propose that operating the mechanism itself is mentally stimulating “play” for a very smart, puzzle-loving bird.

Playfulness, mischief, and social behavior

  • Cockatoos are repeatedly described as pranksters and “little jerks”: destroying decks, hoses, and tools; running what amounts to a neighborhood “protection racket” by threatening property until fed.
  • Stories from Australia and New Zealand extend this to other birds (kea, weka, seagulls, ibis, magpies, crows), showing coordinated theft, distraction tactics, bin-opening and even traffic-cone manipulation.
  • Cockatoos are noted as very social, often in large flocks that split and recombine, with complex interaction and apparent enjoyment of games.

Pet birds: intelligence and welfare

  • Rescue and pet owners emphasize high intelligence, curiosity, destructiveness, long lifespans, and emotional complexity.
  • Birds are likened to “flying eternal toddlers with can‑opener mouths,” needing constant stimulation and often outliving owners.
  • Several commenters say this is why they now consider long-term caging of parrots/cockatoos unethical, despite how charming they are.

Animal intelligence and avian cognition

  • Discussion touches on convergent evolution of intelligence: bird pallium vs mammalian neocortex; neuron density in bird forebrains; parallels with cephalopods.
  • Debate over whether humans underestimate bird intelligence or overestimate human intelligence; some argue cockatoos at “3‑year‑old” level, others push back on dismissing their understanding.
  • Brief disagreement on bird language abilities: parrots can mimic, but some species (e.g., African greys) show evidence of deeper semantic understanding.

Handedness and learning claims

  • An anecdotal “fun fact” claims wild cockatoos are all left-footed when manipulating food, sparking broader discussion of handedness in animals.
  • One commenter calls “learned to operate” overstated because success was ~41%; others counter that twisting a spring-loaded handle with body weight clearly qualifies as operating the device.

Cloud Run GPUs, now GA, makes running AI workloads easier for everyone

Serverless GPU model and use cases

  • Commenters interpret Cloud Run GPUs as a way to run arbitrary models (e.g. Hugging Face) behind an API, paying only when used and scaling to zero between bursts.
  • Main value seen in small/custom or cutting-edge open-weight models where managed APIs don’t exist or are too restrictive.
  • Several note this is best for bursty or early-stage workloads (new apps without clear steady traffic), not for consistently busy services where VMs+GPUs are cheaper.

Cold starts and latency

  • Cold start is a major concern. Reports for CPU Cloud Run range from ~200ms to 30s+, heavily dependent on language and whether using gen1 vs gen2 runtime.
  • For GPUs, a cited example is ~19s time-to-first-token for a 4B model including container start and model load; some see this as unacceptable for interactive UX, others say it’s fine for first-request-only or batch/agent use.
  • Model weights download and GPU memory load can significantly add to startup time; several say you’ll likely keep at least one warm instance, so “scale to zero” is not always practical.

Pricing, billing, and cost controls

  • Pricing of Google’s GPUs (especially beyond L4) is widely viewed as uncompetitive versus specialized providers; L4 on other platforms is quoted at ~40¢/hr vs ~67–71¢/hr here.
  • Cloud Run GPUs bill per use but with a ~15-minute idle window; if you get at least one request every 15 minutes you effectively pay 24/7, often several times the cost of a comparable VM.
  • Lack of hard spending caps on GCP is a major worry. Budgets and alerts exist but are delayed and can’t prevent “runaway” bills; some hack auto-disabling billing but fear breakage.
  • Limiting max instances and concurrency can cap Cloud Run service spend, but not other APIs (e.g. Gemini). Several argue real stop-loss billing is essential for individuals and small teams.

Comparisons to other providers

  • Runpod, vast.ai, Coreweave, Modal, Coiled, DataCrunch, Lambda Labs, Fly, and others are discussed as cheaper or more flexible GPU options, often with per-second billing and/or true caps or prepaid credit.
  • Modal, in particular, is praised for fast cold starts, good documentation, and scale-to-zero GPUs.

Cloud Run experience and architecture

  • Many praise Cloud Run’s developer experience and autoscaling, often preferring it to AWS Lambda/ECS/Fargate/AppRunner; some report large-scale, cost-effective production use.
  • Others report mysterious scale-outs, restarts, and outages that support couldn’t fully explain, prompting moves to self-managed VMs or Kubernetes.
  • Differences between Cloud Run gen1 (faster startup) and gen2 (microVM-based, slower startup) are noted; Cloud Run Jobs (non-HTTP batch) are highlighted.
  • Root access is not yet generally available but is being worked on; GPU types are currently limited (mainly L4), with more promised.

Local vs cloud AI and GPU market

  • Some wish for consumer-grade, local “AI appliance” hardware, arguing many LLMs can run locally if UX were better.
  • Others counter that large-scale training and heavy inference still demand cloud GPUs; GPU supply on major clouds is described as constrained and expensive, fueling the rise of “neo-cloud” GPU providers.