Hacker News, Distilled

AI powered summaries for selected HN discussions.

Page 462 of 543

Fun with C++26 reflection: Keyword Arguments

Usefulness of the showcased keyword-argument trick

  • Many see the reflection-based kwargs approach as “fun but not for real code.”
  • A common recommendation: define a separate parameter struct and pass it (optionally with default member initializers) instead of clever templates/macros; slightly more boilerplate, but far clearer and safer.
  • Several note that the need is usually not reordering arguments, but eliding some while keeping call sites readable.

Existing C/C++ techniques for named-style arguments

  • C-style patterns: argument-name comments at call sites, structs (named or anonymous), and pointers to structs, especially on embedded systems where stack usage and calling conventions matter.
  • C99 designated initializers and compound literals, and now C++20 designated initializers, already give a clear, built-in way to simulate keyword arguments via aggregates.
  • Some mention static analysis (clang-tidy’s argument-comment check) as a practical validator for commented arguments.

Reflection in C++26: promise vs complexity

  • Strong split in attitude: some are excited that compile-time reflection can replace code generators and help with serialization, UI generation, enums, and component systems; others see it as another layer of baroque syntax.
  • Compile-time reflection is praised as “zero-cost” and a better foundation than runtime reflection; others argue runtime reflection (dynamic invocation, plugin-like behavior) is more directly useful.
  • Concerns that C++ committee designs are constrained by extreme backward-compatibility, leading to opaque, over-generic interfaces and unpleasant syntax.

Design, safety, and language philosophy

  • Several posters deliberately restrict themselves to a small, simple subset of C++ (basic classes, STL, algorithms, lambdas; no raw loops/new/delete when possible) to avoid footguns and unreadable template metaprogramming.
  • Others argue complex template-heavy abstractions are sometimes justified in large codebases when they significantly reduce total code.
  • Skeptics see modern C++ evolution as chasing features other languages had decades ago, increasing complexity and technical debt; some suggest simply using languages with native kwargs and reflection (Python, JS, C#) instead.

Miscellaneous points

  • IDE parameter inlay hints are mentioned as a lightweight alternative for making positional calls readable.
  • There is side discussion on UB and lifetime around designated initializers, and on why GUIs or networking are not (or not yet) in the C++ standard library.

Hedge Funds Are Pocketing Much of Their Clients' Gains with 'No Limit' Fees

Visualization of Fees

  • Several comments focus on the article’s graph style, identifying it as similar to a Voronoi treemap (Voronoi + treemap hybrid).
  • Critics find it confusing and “overdesigned,” suggesting simple bar charts instead.
  • Defenders argue irregular areas are useful for showing many small categories without losing them in thin slices and that area-based comparison is natural for humans.

SPY, Leverage, and Risk

  • Some say this reinforces “just buy SPY” (or an ETF like VOO).
  • Counterpoint: time horizon matters—someone retiring into a crash may prefer smoother 6–8% returns with reduced drawdowns.
  • Debate on ways to get 2–3× leverage:
    • Margin is flagged as dangerous due to margin calls in crashes.
    • Leveraged ETFs (SPUU, SPXL, UPRO) criticized for daily rebalancing, volatility decay, and fees; defenders note they’ve outperformed multiples of SPY over long bull runs and can be acceptable if you understand path dependency.
    • Long-dated deep ITM calls proposed as an alternative; others respond they’re costly, complex (delta, theta), and tax-inefficient.

What Hedge Funds Are For

  • One camp: hedge funds aren’t about maximizing returns but providing low-beta, decorrelated, or market-neutral exposure, especially valued by institutions and very wealthy investors.
  • Another camp: “we hedge, not chase alpha” is viewed as marketing speak; skeptics point to long-standing data that, in aggregate, active funds don’t beat cheap indexes after fees.
  • A few note top multi-manager funds are capacity constrained and may genuinely deliver good, smoother returns, but most others mainly monetize management fees.

Fees, Lockups, and Market Discipline

  • Some argue high, “no limit” fees are a voluntary problem for rich, sophisticated investors; if they accept the terms, that’s their choice.
  • Others highlight frictions: lockups, infrequent redemption windows, and the long time needed to judge performance make it hard for investors to “just switch,” dulling competitive pressure.
  • There are calls for better monitoring tools or institutional rules favoring low-fee vehicles; pushback stresses regulators shouldn’t “protect” wealthy LPs from contracts they sign.

Institutions, Pensions, and Social Impact

  • A critical view portrays hedge funds and especially private equity as parasitic: extracting large fees from pension funds, unions, endowments, and sovereign wealth funds, then converting that into political influence.
  • Proposed remedies: restrict public/union funds to low-fee structures, bar them from PE, and manage risk via simple mixes of stocks, bonds, and cash.
  • Opponents argue this is exaggerated or “empirically false,” claiming PE often outperforms public markets and offers diversification; they blame fees and allocator decisions, not the vehicles themselves.
  • There is factual correction that US Social Security itself holds Treasuries, though many other public pension systems do invest in hedge funds/PE.

Hedging Quality and Portfolio Construction

  • Some stress that investors care about how returns are earned: low or even modest average returns can be attractive if they spike during crashes.
  • Others counter that many hedges end up being worse than the underlying risk—“portfolios are littered with bad hedges.”

Careers and Industry Structure

  • Brief side discussion notes hedge funds remain highly selective, often recruit only via referrals/outbound channels, and hire from a narrow pedigree pool.
  • For non-quant software roles, the compensation uplift vs big tech may not justify worse hours and risk, though some are drawn by ultra-low-latency and trading infrastructure work.

Musk-led group makes $97B bid for control of OpenAI

Bid, valuation, and timing

  • Many see the $97.4B offer as both late and low relative to recent paper valuations ($157B raise; rumors of $260–340B deals).
  • Others stress this is a cash bid for control, not a minority stake at a headline valuation, and that “bird-in-hand” logic could look compelling to a judge.
  • Debate over whether this creates a market “floor” for what the nonprofit’s stake is worth, or is irrelevant next to prior preferred-share valuations.

Nonprofit structure and spinout complications

  • A big thread explains: the OpenAI nonprofit owns/controls the for‑profit; any spinout requires the for‑profit to “buy itself” from the nonprofit at a fair market price.
  • OpenAI had reportedly been planning to value the nonprofit’s interest at ~$40B; commenters call this a self‑dealing discount that effectively lets insiders buy control cheap.
  • Musk’s higher offer is framed by some as a legal/optics “wrench”: if the nonprofit refuses ~$97B, it’s hard to defend a $40B internal deal as fair.

Motives behind Musk’s move

  • Several view this as a negotiation tactic or PR stunt rather than a sincere attempt to close:
    • To force a higher valuation for the nonprofit and derail or delay the spinout.
    • To undermine much higher prospective valuations (e.g., SoftBank/“Stargate”) by signaling “real” value is far lower.
    • Possibly to starve OpenAI of capital and help xAI, or as personal revenge from a former co‑founder.
  • Others argue it could also be straightforward empire-building: buying the market leader because xAI is still behind.

Is OpenAI actually worth that much?

  • Strong disagreement on fundamentals:
    • Some say ChatGPT’s brand and traffic (top global site, consumer mindshare) justify “hundreds of billions.”
    • Others note huge losses, fading first‑mover advantage, and open models like DeepSeek-R1 eroding any moat, calling current valuations bubble territory.
  • Microsoft’s role and credits are cited as both a pillar of strength and a structural complication.

Power concentration and political context

  • Extensive concern about Musk as a “single point of failure” across EVs, rockets, satellites, social media, AI, and now US government functions (DOGE).
  • Some commenters explicitly fear a techno‑authoritarian dynamic: using wealth + state position to punish enemies, steer regulation, and capture AI and government IT via chatbots.
  • Others push back that billionaire influence predates him, but agree his visibility and behavior make the problem stark.

Musk–Altman online spat and CEO culture

  • Altman’s dismissive counter‑quips and Musk’s name‑calling (“swindler”, etc.) are widely described as childish, reinforcing cynicism about modern tech leaders.
  • Several lament that CEOs behaving like influencers on X/Twitter erodes any remaining mystique or claim to exceptional competence.

NAT Is the Enemy of Low Power Devices

IPv6 vs NAT for IoT/low‑power devices

  • Several commenters argue IPv6 is the “real” fix: removes ARP/DHCP hacks, avoids NAT, enables simpler end‑to‑end addressing and easier P2P/firewall traversal (PCP, SLAAC, etc.).
  • Others push back: IPv6 stacks add complexity (NDP, ICMPv6, MLD, SLAAC, IPsec), are still not universal on small MCUs and cellular modems, and many carriers or enterprises either don’t deploy IPv6 or cripple it (e.g., disabling SLAAC, enforcing DHCPv6).
  • Privacy is debated: some see per‑device global addresses as worse than NAT; others say with privacy extensions and shared prefixes it’s roughly equivalent to IPv4+NAT. Concern remains about revealing number/timing of devices in a household.
  • IPv6 NAT exists and can be used for prefix translation or masquerade, but many consider it philosophically wrong and only necessary where networks are misdesigned (e.g., CGNAT on IPv6).

NAT, firewalls, and virtual networking

  • NAT is recognized as both enabling today’s IPv4 scale and as a fundamental break with the “network of hosts,” pushing the world into client/server asymmetry.
  • Several comments stress that the user’s real desire is a stateful firewall; NAT is often (wrongly) treated as a security feature. Equivalent inbound‑blocking IPv6 firewalls are possible (and often default).
  • CGNAT and multi‑layer NAT (ISP + CPE + mesh) break port forwarding and direct access, complicating IoT, home servers, and P2P.
  • NAT is also used as a practical tool to work around enterprise restrictions, for VMs and home networks; some prefer it because “what the outside can’t see can’t hurt it”.

Low‑power IoT specifics (sessions, power, and stacks)

  • The core IoT pain: keeping NAT bindings and secure sessions alive forces periodic traffic, which consumes power and data; tearing sessions down requires expensive TLS/DTLS handshakes.
  • Proposed mitigations in the thread: TLS/DTLS session tickets, DTLS Connection IDs, CoAP over DTLS, cellular PSM/eDRX, SMS or NIDD as out‑of‑band wake/triggers.
  • Some practitioners report it’s usually more power‑efficient to fully power‑cycle modems and accept reconnect cost, sacrificing instant cloud‑to‑device control.
  • There is strong advocacy for mature RTOS networking stacks (e.g., Zephyr, lwIP + PPP) and using separate modems rather than closed SoC+modem combos, for better updates, protocol control, and easier hardware replacement as cellular generations sunset.

Security, local‑only design, and alternatives

  • Many argue IoT devices should be LAN‑only, with a local hub or VPN (WireGuard, OpenVPN, Tailscale) for remote access; internet‑reachable consumer IoT is viewed as insecure and economically questionable without subscriptions.
  • Suggestions include Thread/mesh with a border router, strict WLAN isolation, and assuming closed‑source IoT is compromised by default.
  • Some see the article as overblaming NAT: stateful firewalls also time out, and many IoT control flows can tolerate polling and delayed updates. Others think the NAT‑timeout problem remains a real blocker for ultra‑low‑power, always‑reachable devices.

1% Equity for Founding Engineers Is BS

Employee-Owned and Cooperative Models

  • Multiple commenters ask why there aren’t more fully employee‑owned tech companies.
  • Main constraint identified: financing. To be truly employee‑owned, employees must supply capital or accept low/no salaries until profitability; most can’t or won’t.
  • Partnerships / buy‑in models (like Big 4-style partner tracks) are cited, but seen as risky for new hires and hard to scale; employees often want upside without downside risk.
  • Practical issues:
    • How to price buy‑ins for new employees.
    • How to handle exits (buybacks, forced sale at retirement, caps on ownership).
    • Ongoing fights over dilution, dividends vs salary, and reinvestment vs payouts; several anecdotes of employee‑owned firms devolving into bitter governance disputes.
  • Co-ops and ESOPs exist (e.g., food brands, consultancies, tech co-ops), but are viewed as better fits for stable, profit-sharing businesses than for high-growth tech startups.

Crypto Tokens vs Traditional Equity

  • The article’s proposed solution—replacing equity/options with a company “utility token”—draws heavy skepticism.
  • Concerns:
    • Legal status (likely a security / ICO risk, SEC issues).
    • Zero enforceable rights vs stock (no voting, no protections, acquirers can ignore it).
    • Unlimited dilution potential; no clear safeguards for employees.
    • Practical questions about lost wallets, insider trading, and aligning token price with company value.
  • Recruiters report that simply mentioning “crypto” causes over half of candidates to decline; others note euphemistic job ads hiding crypto involvement.
  • Some acknowledge VC has its own distortions and welcome experimentation, but many see tokens as re-running known scams rather than genuine innovation.

Economics of 1% for Founding Engineers

  • Numerous comments argue 1% for a true “founding engineer” is a bad deal:
    • You’re effectively investing a large salary cut and concentration of risk into a single illiquid asset.
    • After dilution, preferences, and typical acquihire dynamics, that 1% can become negligible or worthless.
  • Stories: repeated experiences of startup equity going to zero; examples where employee options were wiped while founders/investors captured all value.
  • Some founders share their own splits: first engineers getting far less than founders, but founders also taking years of low/no salary, extreme hours, and non-technical burdens (fundraising, hiring, sales).

Risk, Fairness, and Market Dynamics

  • One camp: equity should track risk; founders started before any selection filter, many fail before they can even hire, so later “founding engineers” are already joining a de‑risked entity—hence ~1% is market-clearing.
  • Other camp: early engineers often do 80–90% of founder-level work without proportionate equity; if you’re doing that much, you should be a co‑founder, not a 1% employee.
  • Disagreement on “fairness” vs “market outcome”: some say if 1% persists, it must be acceptable to enough people; others argue many candidates misunderstand options, probabilities, and dilution.

Career and Negotiation Perspectives

  • Several advise treating startup equity as a near-zero-value lottery ticket and optimizing for cash, learning, and network.
  • For high-paid big-tech engineers, staying put and investing surplus (even in index funds) often has higher expected value than 1% at a random startup.
  • Others highlight non-monetary motives: escaping big-company “existential dread,” wanting more impact, or enjoying early-stage chaos; for those people, lower pay and small equity can still be rational.
  • Repeated advice: don’t accept “founding engineer” titles plus sub-market pay; either negotiate hard (2–5%+ pre-funding, or strong re‑ups) or start your own company.

Undergraduate shows that searches within hash tables can be much faster

What the new hash table result actually is

  • Commenters clarify: the result gives sublinear worst-case probes for queries and insertions in open-addressing hash tables when the table is very full, on the order of (log(1/δ))² where δ is the empty fraction, versus prior Ω(1/δ)–type bounds.
  • High‑level idea (from paper/talk summaries): conceptually split the table into exponentially smaller regions (“layers” or “funnels”). Insertions probe a bounded number of slots in each layer, possibly skipping early free slots, and “funnel” down to smaller layers.
  • This deliberately non‑greedy probing breaks a long‑standing conjectured lower bound for greedy schemes, while preserving constant average query time even near full capacity.

Big‑O, models of computation, and “O(1) hash tables”

  • Large subthread debates whether hash tables are really O(1):
    • One side argues that in pure big‑O, as keyspace and table grow, hashing and comparisons require O(log n) work (word size, key length), so hash tables and trees both pick up extra log factors.
    • Others counter that big‑O is always relative to an assumed cost model (word‑RAM, hashing oracle, external memory, etc.), and in that standard model, “one hash + one memory access” is taken as O(1).
  • Related discussion on how these abstract models ignore cache hierarchies and memory latency, so asymptotics alone don’t predict real performance.

Practical impact and implementations

  • Several note that typical production hash tables resize well before they get “almost full” (e.g., 75–87.5% load), making worst‑case chains short; over‑allocating is often simpler and faster.
  • Concern that the new structures may have poor locality and large constants, making them “galactic” algorithms: asymptotically better but slower for all realistic sizes.
  • Others point out niche value: nearly-full, non‑resizable tables or memory‑constrained settings (databases, large in‑memory indexes).
  • Links are shared to early PoC implementations of funnel/elastic hashing and to other high‑performance tables (Swiss tables, Robin Hood, bucketed designs) for comparison; some skepticism that uniform‑probing tables like in the paper are actually used in production.

Article presentation and terminology

  • Multiple readers complain Quanta’s piece is heavy on human‑interest and light on the actual mechanism, with no real pseudocode, benchmarks, or clarity on when it helps.
  • Some object to calling this “data science” rather than theoretical computer science/data structures, speculating the term came from misreading arXiv’s cs.DS tag or editorial trendiness.
  • Others defend the profile style as celebrating a notable undergraduate result and inspiring students.

AI, originality, and prior work

  • One commenter tried major LLMs on the open problem; none rediscovered the technique, prompting remarks that “human‑driven computer science is still safe.”
  • Thread splits between:
    • Those who see this as evidence that fresh human intuition still beats current AI on deep theory.
    • Those who note AI has improved other algorithms and could plausibly find such results as models and tooling improve.
  • A long meta‑discussion explores whether breakthroughs often come from ignoring prior work vs. building on it:
    • Pro‑ignorance camp cites this result and game‑design anecdotes as benefits of not being trapped by prior “impossibility” lore.
    • Others emphasize survivorship bias: most such attempts just re‑invent or fail; the undergrad was still deeply embedded in current theory and inspired by an earlier “Tiny Pointers” paper, not working in a vacuum.

Monty Hall and counterintuitive proofs

  • A huge side‑thread uses the Monty Hall problem as an analogy: a simple, counterintuitive probabilistic result that many smart people initially rejected.
  • Debate centers on whether the classic statement is underspecified (host strategy, conditional probabilities) or well‑formed; people argue about independence assumptions and how real‑world behavior affects probabilities.
  • This is tied back rhetorically to how counterintuitive results in hashing and complexity can also clash with experts’ heuristics.

Other notable discussion points

  • Some ask for Rust or other language implementations; answers range from “be the change” to skepticism about AI‑generated implementations for a “fiddly and novel” algorithm.
  • Explanation that in some theory subfields author order is alphabetical, so the undergraduate not being “first author” is normal; contribution statements are suggested as clearer practice.
  • One commenter highlights that the work was NSF‑funded and raises concern that proposed cuts to such funding would reduce exactly these kinds of foundational advances.

CAPTCHAs: 'a tracking cookie farm for profit masquerading as a security service'

Scale, “unpaid labor”, and dystopian framing

  • Commenters highlight the study’s claim of 819M hours spent on reCAPTCHAv2, equating it (using rough lifespan math) to over a thousand human lifetimes.
  • Several frame this as uncompensated labor used to train AI and build tracking profiles, with dystopian comparisons (humans kept around just to annotate data for machines).
  • One person notes the $1T figure is mostly attributed to tracking cookies; that value is realized even without users solving the image challenges.

Security value vs. “tracking first” criticism

  • Strong disagreement here:
    • Many site operators report CAPTCHAs (especially reCAPTCHA) dramatically reduce spam, card‑testing fraud, and brute‑force login attempts. They stress security is about raising cost/probability, not absolute prevention.
    • Others say most bots are trivial and could be stopped by any low‑effort measure (simple math question, honeypot field, rate limiting), so Google-scale CAPTCHA is overkill and unnecessary data exposure.
    • Critics argue the paper’s “worthless for security” framing ignores this practical middle ground where CAPTCHAs stop unsophisticated bots but not well-resourced ones or paid-solving services.

Alternatives and technical debates

  • Suggested alternatives: homegrown CAPTCHAs, simple questions (“what color is snow?”), CSS-hidden honeypots, rate limiting, proof-of-work (e.g., mCaptcha, Tor’s PoW), Cloudflare’s “Attestation of Personhood,” behavioral/ML-based bot detection, and just turning forms off or centralizing comments.
  • PoW gets heavy pushback:
    • Hardware disparity makes it cheap for attackers with GPUs/ASICs but painful for users on old phones.
    • Some argue algorithms like RandomX narrow this hardware gap; others doubt practicality in JS on low-end devices.
  • Accessibility and cultural bias in puzzles are major concerns; “simple” questions can exclude disabled users or those from different regions/cultures.

Privacy, law, and the open web

  • Several EU-based commenters say reCAPTCHA is effectively or explicitly illegal under GDPR decisions; some governments nonetheless use it (and Cloudflare) for essential services.
  • There’s tension between:
    • Users who now abandon any site that shows a CAPTCHA.
    • People warning this response disproportionately harms small sites that can’t “eat” bot traffic like big platforms.
  • Some propose regulating both invasive CAPTCHAs and disruptive bot traffic; others are pessimistic about enforceability across jurisdictions.

How reCAPTCHA works and its “elegance”

  • Explanations note dual use: mixing known and unknown images to both validate users and gather new labels, plus extensive device/behavioral fingerprinting.
  • A minority admire reCAPTCHA’s “elegant” multi-purpose design (spam defense, AI training, tracking), while others call it a global-scale Trojan horse for unavoidable Google scripts.

The Anthropic Economic Index

Perception of the Index (PR vs Insight)

  • Many see the index as primarily PR/marketing to justify valuations and influence regulators or “be in the room” for government money.
  • Others argue it’s valuable market research and a rare, concrete look into millions of real-world AI interactions that should be applauded.
  • Some think the target audience is journalists and business leaders, not technical HN readers; others complain the “index” is never clearly defined as a single interpretable metric.

Data Quality, Methods, and Trust

  • Concerns: Anthropic has incentives to cherry-pick or “game” numbers, and outsiders have little way to detect it.
  • Counterpoint: metrics are mostly relative shares, reducing incentive to fake; being caught would hurt reputation.
  • Methodological critiques:
    • Only web chats, no API usage, so automation-heavy workloads are largely invisible.
    • Classifying all game-debugging or dishwasher questions as “occupational tasks” in specific professions is seen as dubious.
    • Counting “conversations started” biases toward desk jobs and doesn’t distinguish one-off tests from sustained use.

Usage Patterns and Sector Penetration

  • The dataset confirms what many suspected: usage is dominated by programmers and technical tasks; Claude in particular is seen as a “coding model.”
  • Surprisingly low use in law, medicine, and finance; explanations include liability from hallucinations, conservative cultures, and lack of good workflows/interfaces.
  • Several note massive usage via third-party tools (e.g., coding IDEs) that the study omits, skewing the picture.

Economic and Labor Implications

  • The reported 57% “augmentation” vs 43% “automation” is read in two ways:
    • Optimistic: tools mainly boost worker productivity rather than replace jobs.
    • Pessimistic: mass displacement is coming anyway; current layoffs and hiring freezes are cited as early evidence.
  • Debate over whether LLMs have yet produced measurable macro productivity gains; some see mostly “better search,” others report significant staff cuts in marketing/support and multi-role automation in small businesses.

Privacy and Ethics

  • Anxiety over reuse of user chats, though others stress that only anonymized metadata, not raw text, was published and that Anthropic described a privacy-preserving pipeline.
  • Broader skepticism about “ethical AI” and public-benefit structures, especially given partnerships with defense/government firms.

Surnames from nicknames nobody has any more

Patronymics and surname origins

  • Extensive discussion of patronymic systems: Icelandic (-son / -dóttir), historical Scandinavian practice, Welsh “ap/ab ” (Upjohn, Powell, Pritchard, Pugh, Bowen), Irish/Scottish “Mac/Mc” and “O’…”, Norman “Fitz…”, Spanish/Portuguese -ez/-es, Ukrainian -enko, Persian -zadeh, Arabic bin/ibn and abu, Japanese -rō, etc.
  • Several commenters describe how patronymics “froze” into fixed family surnames in Scandinavia and elsewhere, often due to 19th–20th c. legislation or administrative needs (tax, conscription).
  • Some note issues when patronymics coexist with fixed surnames (siblings with different -son/-dóttir names, confusion in non-patronymic societies).

What counts as a “family name”?

  • Debate over whether Icelandic -son/-dóttir are truly “family names” since they’re not inherited across generations, vs. the pragmatic view (whatever sits in the “surname” field on a passport or form).
  • Examples from passports and phonebooks: Iceland sorts by given name; surnames there function mainly for disambiguation, as elsewhere.
  • Separate discussion of Slavic gendered surnames (-ski/-ska, -ov/-ova) and Lithuanian endings differing by gender and marital status.

Nicknames turning into surnames

  • Many English surnames are shown to derive from rhyming or shortened nicknames: Richard → Rick → Dick → Dixon; Robert → Rob/Bob → Robinson, Dobson; Simon → Simme → Simpson; Bartholomew → Bat/Bate → Bates/Beattie; Roger → Hodge/Dodge → Hodgson/Dodgson; Theobald → Tibb → Tibbs; Hugh → Hud → Hudson, etc.
  • Realizations that Nixon = Nick’s son, Robinson ≈ Robertson, Dawson = David’s son, Harris/Harrison from Harry/Henry, etc.
  • Long digression on traditional English nicknames: William → Will/Bill, Margaret → Meg/Peg, Mary → Molly/Polly, John → Jack, Henry → Harry/Hank/Hal, Elizabeth’s many variants, etc.

Letters, diacritics, and spelling drift

  • Argument over whether characters like ö, é are “separate letters” or base letters with diacritics; examples from Icelandic, French, German, Hungarian, etc., including sorting conventions.
  • Discussion of how diacritics are treated when keyboards or systems can’t easily produce them, and of alternate spellings (e.g. Hülsbeck/Huelsbeck).
  • Note that “nickname” itself comes from rebracketing “an ekename” → “a nekename”.

Marriage, gender, and surname transmission

  • Historical English usage of “Mrs John Smith” and analogues, still visible in wedding invitations and older etiquette; some find this offensive or obsolete.
  • Anecdotes of banking and paperwork problems when checks are written to non-existent “Mrs Husband’s Name”.
  • Comparisons to Spanish dual-surname systems, Norwegian two-surname children, hyphenated or merged surnames, and proposals for “union names” or random inheritance of one surname from each parent.
  • Descriptions of culturally specific systems: Lithuanian gendered and marital suffixes, Hungarian women historically adopting husbands’ full names, Japanese terms like 奥さん (okusan).

Diminutives and augmentatives in English

  • Historical English diminutive -kin noted (napkin, pumpkin, munchkin, bodkin, firkin, catkin), but essentially no longer productive in everyday speech.
  • Modern productive diminutives are mostly -y/-ie (Johnny, kitty, tummy), plus -let (applet, hamlet), -ling (duckling), -ette (cigarette), occasionally -o (kiddo, doggo); mostly used for children, informality, or specialized jargon.
  • Thread compares this to very productive diminutives/augmentatives in Romance languages (-inho/-inha, -ito/-ita, -ón/-ona), which English lacks.

Immigration, transcription, and myth-busting

  • Multiple family anecdotes of surnames changing spelling (e.g., -sen to -son, farm names altered, Slavic diacritics lost) through bureaucratic error, low literacy, or later voluntary “Americanization”.
  • Others stress historians’ consensus that Ellis Island officials did not routinely rename immigrants; changes mostly came from immigrants themselves or later generations.
  • Transliteration from non-Latin scripts (Russian, Arabic, Greek) shown to be inconsistent across languages and time.

Other notable curiosities

  • Occupational surnames and roles: Peterman linked to saltpeter work; Baxter as female Baker; King/Lord/Virgin as actor roles in medieval mystery plays.
  • Examples of culture-specific kinship or ordinal naming (Roman numerals in praenomina, Japanese numbered sons, English Junior/Chip/Trip/Skip patterns).
  • Several commenters express surprise at how many “ordinary” English surnames encode forgotten nicknames or family relationships once you know the patterns.

Why hasn't commercial air travel gotten any faster since the 1960s? (2009)

Fuel economy, physics, and why planes aren’t faster

  • Core answer from the article and the thread: fuel economy dominates. Drag – especially wave drag near Mach 1 – rises sharply with speed, so fuel per passenger‑mile increases steeply.
  • Modern high‑bypass turbofans are much more efficient than 1960s turbojets, but airlines have deliberately slowed cruise from early jet speeds to around Mach 0.8–0.83 to hit a fuel‑cost optimum.
  • Concorde is cited as proof that faster is possible but uneconomic: it needed huge fuel, had narrow, uncomfortable cabins, and worked only as a niche, premium product on a few routes.
  • Similar patterns are noted in shipping (“slow steaming”) and high‑speed rail (practical ceiling ≈300–350 km/h). Beyond that, energy use and infrastructure costs explode for modest time savings.
  • Many commenters argue that what’s improved since the 1960s is efficiency, safety, and comfort options, not cruise speed.

Total journey time vs cruise speed

  • For many trips, door‑to‑door time is dominated by:
    • Travel to far‑out airports
    • Early arrival buffers for check‑in and security
    • Boarding, taxiing, and baggage claim
  • Doubling cruise speed might cut a 6‑hour flight to 3–4 hours, but the fixed 3–4 hours of “airport overhead” remains. For medium‑haul, this often makes flying little better than driving or trains.
  • Several people argue the real gains would come from:
    • Faster ground links to airports
    • Smoother security and boarding
    • Better scheduling, fewer connections, less hub‑and‑spoke

Security, TSA, and “user‑space” slowness

  • Many say travel feels slower mainly because of post‑9/11 security: long, unpredictable TSA lines; liquids and shoes rules; ID checks; more time buffers.
  • There’s strong disagreement on value:
    • One side: cockpit doors plus changed passenger attitudes already solved the hijacking problem; much of TSA is security theater and a jobs/lobbying machine.
    • Other side: even imperfect screening and visible measures deter organized plots; hijackings used to be routine, now they’re vanishingly rare.
  • Pre‑check/Clear are seen by some as rational risk‑based screening; by others as paid bypasses that prove the inconvenience isn’t truly necessary.

Comfort, economics, and “enshittification”

  • Most passengers choose cheapest tickets; airlines optimize for cost per seat, not speed or comfort.
  • Comfort has diverged by class: lie‑flat business/first has improved dramatically; economy has seen tighter pitch, more fees, and more friction (baggage fees, overhead fights, boarding chaos).
  • Several note that for many, better seats and simpler processes matter more than shaving 30–60 minutes off flying time.

Alternatives and future concepts

  • High‑speed rail is cited as a more pleasant medium‑distance option where available; some prefer long train or car journeys to avoid airport hassles.
  • Supersonic (e.g., new startups) and “Earth‑to‑Earth” spaceflight are discussed as possible rich‑niche futures, but commenters doubt near‑term economics, safety, and environmental acceptability.

The hallucinatory thoughts of the dying mind

Personal accounts of dying visions and “good deaths”

  • Many share bedside stories: dying relatives talking with deceased parents, spouses, miscarried children, or childhood friends, often with visible joy or peace.
  • These episodes bring comfort to survivors, especially when prior dementia or Alzheimer’s had limited communication.
  • Several describe “beautiful” deaths at home or in hospice with family present, versus regretful stories where denial or distance left people to die largely alone.
  • One commenter asks where to share such experiences; another recounts an unhappy, highly medicalized end and missing the final call.

Sleep, hypnagogia, and everyday hallucination-like states

  • Multiple users connect deathbed experiences to hypnagogia: bizarre but convincing thoughts as we fall asleep or wake.
  • People recall sleep-talking, “sleep-writing,” or phone conversations drifting into nonsense, then snapping back.
  • Anecdotes include using this state creatively (e.g., Dali’s spoon trick).

Medical and neurological factors

  • Several describe delirium from infections (especially UTIs), strokes, metabolic issues, or brain injury causing vivid hallucinations and conversations with the dead.
  • Some note altered mental status in the elderly is standardly treated as a metabolic/infectious problem, though this wasn’t always communicated to families.
  • Delirium is framed as an “organic psychosis” arising from widespread brain dysfunction; antipsychotics have mixed results.
  • “Terminal lucidity” is discussed: brief, striking returns of clarity and even physical capability just before death, sometimes seen as challenging simple “brain failure” narratives, sometimes as a last mobilization of reserves.

Near-death experiences and function

  • NDEs are reported as intensely structured and often comforting; one recalls being told death is like birth into another world.
  • Some suggest such experiences might ease dying or elicit care from others.
  • Others argue evolutionary “just-so stories” for NDEs feel strained, noting the paradox that many survivors lose fear of death, which seems maladaptive.

Spiritual vs materialist interpretations

  • Sharp debate: some insist all such phenomena are hallucinations because “the supernatural isn’t real”; others argue we lack a complete model of reality, so hard-line dismissal is unwarranted.
  • Discussions range over nondual Buddhist ideas, Tibetan bardo, empirio-criticism, and whether “values” or “significance” can exist in a purely physicalist universe.
  • One thread claims consciousness cannot simply “arise from the brain,” citing popular-level arguments; others firmly defend materialism.

Caregiving ethics and communication

  • Hospice advice to affirm dying visions is contrasted with dementia-care guidance that warns against “therapeutic lying” and recommends reflection/redirection instead.
  • Several caregivers and clinicians emphasize not arguing with dying or delirious patients; gentle presence, touch, and reassurance are seen as most helpful.

Consciousness, hallucination, and AI analogies

  • Commenters note that ordinary perception is already a constructed narrative; dying brains under stress may simply produce more fragmented stories.
  • Comparisons are drawn between human speech and LLM token prediction, suggesting much of our own talk is similarly automatic, which feeds into broader debates about mind, self, and agency.

What about K?

Learning K and Related Resources

  • Several commenters recommend “Q for Mortals” and the KX “phrases” / idioms collections as more example‑rich companions to the linked intro.
  • Nial and other Iversonian/array languages (APL, J, BQN, Uiua) are mentioned as alternatives, with guidance that if you learn one, the others are easier to pick up.
  • For beginners, BQN and J are often suggested due to open tooling; Dyalog APL is seen as best tooled but commercial.

Intended Use Cases and Comparisons

  • K is described as a fast vector/array language, primarily used with kdb+ for time‑series and tabular analytics in finance.
  • Compared to SQL, K/q can express complex queries in very few tokens; compared to Python+Pandas/R+tidyverse, they are more focused, less “ecosystem rich,” but often faster.
  • Some see it as ideal for data wrangling and quant research, not for full‑stack applications or nanosecond‑level trading engines.

Readability, Terseness, and Culture

  • The “readability is a property of the reader” claim provokes debate. Many find K unreadable or “write‑only”; practitioners insist it becomes natural with practice, like regex or Chinese script.
  • The JSON implementation in K is cited as an extreme example: some see it as IOCCC‑style obfuscation, others as compact, readable code for experts.
  • Terseness is defended as increasing the amount of code a programmer can keep in working memory; critics argue brevity in characters is meaningless versus clarity in tokens.

Performance and Benchmark Debate

  • K/Shakti marketing claims around 50GB CSV loading far outpacing Polars spark skepticism, especially given a DeWitt‑style anti‑benchmark clause.
  • A Polars engineer independently tests the same benchmark and describes their SIMD + multithreaded CSV parser, getting ~7–8s for full parse on large machines.
  • Even skeptics concede that if K’s 1.6s claim is accurate, it’s extraordinary, but independent verification is lacking.

Ecosystem, Versioning, and Licensing

  • There is concern over nine mutually incompatible K versions and heavy, Oracle‑like licensing for q/kdb+.
  • Some report real‑world pain: proprietary binaries, expensive training, upgrades for memory leaks, and coupling one’s livelihood to a niche, closed system.
  • Others use K/q recreationally (e.g., Advent of Code) and enjoy its expressiveness but warn against basing a career or large codebases on it.

I built an open source AI tool to find my autoimmune disease

Patient Empowerment and Positive AI Use‑Cases

  • Several commenters describe using ChatGPT/Claude to interpret oncology reports, ER notes, lab panels and imaging, then using that understanding to ask better questions and advocate for themselves or relatives.
  • LLMs are praised for: translating jargon to lay terms, summarizing scattered records, surfacing possible conditions or tests, and preparing patients for specialist visits or hospital rounds.
  • Some say AI helped them or relatives move toward correct diagnoses (e.g., MS, ankylosing spondylitis, Crohn’s-related arthritis, dyshidrotic eczema) faster than or alongside doctors.

Skepticism About the Tool and AI Diagnosis Claims

  • Multiple commenters argue the showcased tool is mainly re-packaging information created by doctors; AI’s “diagnosis” came only after ~$100k of tests and specialist notes.
  • Critics say the story is vague about the exact autoimmune disease, the timeline, and what the model actually output (likely a broad “possible autoimmune condition” among others).
  • Some suspect the post is more about promoting an AI side project than demonstrating a genuine diagnostic breakthrough.

Fragmented Healthcare Systems and Missed Diagnoses

  • Many describe years of being bounced between specialists, each focused on one organ system, with no one owning the “big picture” (common for autoimmune, EDS/hEDS, ME/CFS, endometriosis, chronic pain).
  • Structural issues cited: underfunded public systems, private‑equity and hospital consolidation, short appointments, heavy administrative overhead, and limited incentives to pursue low‑probability “zebras.”
  • Commenters emphasize the need for self‑advocacy or a dedicated case manager; some think AI could partly fill that coordination role.

Risks of Self‑Diagnosis, Hallucinations, and Hype

  • Several doctors and technically literate users worry about AI reinforcing hypochondria, over‑testing, and doctor‑shopping, similar to but stronger than WebMD or TikTok self‑diagnoses.
  • LLMs are described as performing like a “recent graduate”: useful for checklists and differentials but prone to confident errors and missing nuance in lab variability.
  • Concern that desperate patients might over‑trust vague AI suggestions like “you may have an autoimmune disease” and pressure doctors into unnecessary or harmful procedures.

Privacy, Regulation, and Data Ownership

  • Debate over whether privacy rules (e.g., HIPAA) truly prevent doctors from using cloud LLMs with patient consent, or whether institutions are simply risk‑averse.
  • Many insist PHI should not be sent to commercial models; they favor self‑hosted or fully local tools and formats like FHIR to integrate personal records safely.
  • Some argue current regulation and institutional conservatism already impede potential benefits, while others stress that cautious rollout is appropriate.

Lifestyle, Genetics, and the Limits of Medicine

  • Thread includes long side discussions about diet (keto/carnivore, gluten sensitivity), micronutrients, MTHFR and other SNPs, and whether AI can help individuals navigate these complex, poorly understood areas.
  • Others push back that much online gene/diet discourse is alternative‑medicine hype with weak evidence, and that over‑medicalization and overtreatment are already serious problems.

Broader Debate on AI’s Role in Society

  • A large subthread veers into AI in the arts: some see it as just another tool in a long history of tool changes; others see it as theft and deliberate devaluation of creative labor.
  • Similar anxieties are projected onto medicine: will AI be used to augment diagnosis and care, or primarily to cut costs and deny treatments?
  • Overall, commenters see AI in healthcare as both promising and fraught: potentially transformative for organizing information and widening access, but dangerous if oversold or deployed under perverse economic incentives.

Advertising Is a Cancer on Society (2019)

Advertising, surveillance, and capitalism

  • Several commenters see commercial surveillance as the real “cancer,” with advertising as its economic engine; state surveillance often piggybacks on the same data flows.
  • Others argue the article is really attacking capitalism itself: ads are integral to a system that treats attention and persuasion as core economic activities.

Zero‑sum competition and “Moloch trap”

  • Many agree with the essay’s claim that much advertising is zero‑ or negative‑sum: firms escalate spend mainly to cancel out each other, while platforms extract rents.
  • Super Bowl and big‑brand campaigns (Coke/Pepsi, airlines, luxury cars) are cited as prototypical awareness wars where consumers already know the product exists.
  • A minority argues this is a “shallow” view: competition via advertising can still reveal options and drive product improvement.

Do ads ever do good?

  • Defenders highlight roles like: announcing genuinely new products or services, local businesses reaching nearby customers, theater tickets, public‑health campaigns, cancer screening.
  • Critics counter that informing and persuading are in conflict when the informant profits from overuse; in practice, most spend goes to manipulation, not neutral information.
  • Several take an absolutist stance: even the “good” cases aren’t worth the systemic harms and they’re willing to “throw the baby out with the bathwater.”

Psychological manipulation and cultural damage

  • Strong agreement that modern ads systematically exploit cognitive biases, create dissatisfaction, and induce anxiety (“your life will suck unless you buy…”).
  • Examples include pharma “mini‑movies,” nostalgia mining (Disney, Nintendo, retro music and film scenes), and targeted financial‑product ads aimed at young adults’ insecurities.
  • Some broaden the critique to all “manipulation through information” (politics, religion, escapist entertainment); others insist advertising is uniquely industrialized psychological abuse.

Children, status, and inequality

  • Deep concern over advertising aimed at children, who can’t distinguish mascots from culture; parents debate shielding vs. supervised exposure.
  • Status marketing (phones, cars, luxury brands) is blamed for reinforcing class markers and cruelty toward the poor.

Regulation, bans, and free speech

  • Proposed responses include: banning or heavily restricting whole categories (gambling, drugs, kids’ ads, billboards), defining and outlawing bias‑exploiting techniques, or even banning commercial advertising outright.
  • Opponents warn this would require draconian control over corporate speech and blurred lines with art, reviews, sponsorships, and basic product information. There’s sharp disagreement over whether corporations should have speech rights at all.

Business models and alternatives

  • Many lament that ad‑funding “enshittified” the web yet admit there’s no widely deployed replacement: micropayments are blocked by transaction costs, subscriptions increase inequality, and users often choose “free with ads” over paying.
  • Some see the core issue as user disempowerment and DRM: without control over software and media, people can’t meaningfully choose what ads to accept.
  • A few content creators describe direct experience: adblocker uptake can kill revenue, but advertisers largely refuse non‑tracking arrangements.

Coping strategies and limits

  • Technically savvy users report the modern web as “borderline unusable” without aggressive ad‑ and tracker‑blocking; others differentiate “tolerable” (opt‑in, skippable, contextual or entertaining) from intolerable (forced, loud, deceptive) formats.
  • There is scattered but strong support for pushing back at least on the worst forms—especially those harming kids, the poor, and vulnerable groups—even among people who accept some advertising as unavoidable.

I built an AI company to save my open source project

Overall reception

  • Many commenters enjoyed the article and praised the perseverance, storytelling, and focus on “real” AI that optimizes operations instead of generating text.
  • Several people had successfully used the predecessor (OptaPlanner) in production and expressed intent to use Timefold in future projects.

What Timefold is (constraint solver / “AI”)

  • Commenters explain it as a constraint solver / operations research tool, not a typical ML model.
  • It supports hard, soft, and now “medium” constraints to reflect must-have, nice-to-have, and “unassigned work” penalties.
  • Under the hood it uses meta-heuristics such as simulated annealing, tabu search, and especially Late Acceptance; base algorithms are described as a small fraction of the total work.
  • There is debate about what counts as “AI”; some view rule-based / OR systems as classic AI (GOFAI), others note the label is now partly marketing-driven.

Modeling constraints & usability

  • Multiple people note that the hardest part is translating messy business logic (dates, skills, exceptions) into constraints.
  • Timefold’s “constraint streams” are described as more human-readable than low-level MILP formulations, but the learning curve is still seen as steep.
  • One user spent 40+ hours to build an MVP for employee scheduling and felt documentation lacks a clear mental model plus many end‑to‑end real-world examples.
  • The team acknowledges past pain points (e.g., a removed project-scheduling example) and emphasizes new quickstarts, REST APIs (field service routing, shift scheduling), and educational material.

Vehicle routing & distance matrices

  • A side thread covers vehicle routing: SaaS distance matrices (e.g., Google) are considered expensive at scale.
  • Alternatives mentioned include straight-line distances, offline OpenStreetMap routers, OSRM, GraphHopper, and custom caching / precomputation.
  • Timefold’s own routing APIs sit atop OSRM or other providers with caching and incremental requests.

Adoption challenges & human factors

  • Several commenters confirm that many hospitals, logistics firms, and others still schedule manually despite decades-old algorithms.
  • Reasons cited: lack of in-house OR expertise, failure-prone and expensive projects, messy real-world constraints (e.g., snow, narrow roads), and resistance from experienced planners.
  • Successful deployments require involving domain experts, treating the solver as “assistant” rather than replacement, and supporting real-time replanning and manual overrides.

Startup economics and open-source business

  • A long subthread debates founder compensation: early founders often live off savings and later take below-market salaries, with equity as primary compensation.
  • Some see this as necessary “skin in the game”; others view it as undue leverage and a path to personal bankruptcy.
  • There is concern about open-source “moats” and whether a larger player could simply host Timefold; commenters suggest the hosted models, APIs, and accumulated expertise may be the defensible layer.

Red Hat / IBM and product evolution

  • Some lament Red Hat/IBM killing or sidelining products (including OptaPlanner), and describe OptaPlanner as powerful but difficult and “consulting-heavy.”
  • The Timefold team responds that Timefold is designed to be easier, with higher-level abstractions and REST APIs, but acknowledges not all problems will ever be trivial.

Building a personal, private AI computer on a budget

Quantization, precision & model behavior

  • Many argue P40’s poor FP16 isn’t critical because local setups usually run quantized models (Q4–Q8) to fit VRAM, often with negligible loss at Q6–Q8.
  • Quantizing the KV cache (context) can greatly expand context length and reduce memory, but quality impact is model- and task-dependent.
    • Some models (e.g., Command-R) handle KV quantization well, others (e.g., Qwen) can “go nuts,” especially on context‑sensitive tasks like translation or evaluation; more forgiving for coding/creative use.
  • There’s confusion over precision actually used at inference and how “standard” KV quantization is; consensus: it’s supported widely but not universally safe to enable.

Performance, tokens/sec & usability

  • 4 tokens/sec on a 671B model is seen by some as “runs but unusable,” others say it’s fine for async, deep or overnight jobs, or agentic workflows.
  • For interactive coding or long back‑and‑forth chats, many want ≥10–40 tok/s; sub‑10 tok/s feels sluggish, especially with large outputs.
  • Single-user home setups are typically batch=1; many cloud comparisons note that for $2k of API usage you can often buy billions of tokens at high speed, so local heavy models only “win” if you have sustained, high usage or strict privacy needs.

Hardware tradeoffs: GPUs, Apple silicon, and “budget”

  • Used server GPUs (P40, M40, K80, P41, etc.) offer lots of VRAM cheaply but bring driver pain, missing CUDA/compute features, high power, and often poor perf per watt; some are effectively “toy slow.”
  • Consumer GPUs (3090, 3060, 1080 Ti, 4090) generally outperform P40‑style cards, but high‑VRAM models are expensive; multi‑GPU setups must watch PCIe bandwidth and sharding strategies.
  • eGPU over USB4/Thunderbolt can work surprisingly well if the whole model fits in VRAM (LLMs see negligible perf loss; ~10% for some PyTorch workloads).
  • Apple M‑series (especially Mac mini / Studio) are highlighted as compelling: unified memory, decent bandwidth, low power, very simple setup.
    • Counterpoints: weaker memory bandwidth vs high‑end NVIDIA, no CUDA so many research/codebases don’t “just work,” and some generative image models (e.g., Flux) are much slower.
  • Several commenters feel calling a dual‑P40, ~€1700 build “budget” is misleading; “real budget” is closer to a single mid‑range GPU or repurposed existing hardware.

Cloud vs local: cost, privacy, and risk

  • Many see local LLMs as an enthusiast hobby with questionable ROI; renting GPUs or using APIs is usually cheaper per unit of useful work, especially as models and hardware obsolete quickly.
  • Some mix approaches: a local smaller model handles private/tool-calling tasks, escalating non‑sensitive heavy work to cloud models.
  • Arguments for local:
    • Strong privacy (avoiding ToS changes, data leaks, opaque “shadow prompting” and provider-side guardrails).
    • Predictable spend vs cloud “tail risk” from misconfigured GPU instances.
  • Arguments for cloud:
    • Better models, higher speed, no hardware/driver headaches, easy to switch as new models appear.

Practical advice & ecosystem state

  • VRAM needs exceed raw model size due to KV cache and context; over‑provisioning is common practice. Quantized variants (Q4/Q5/Q6) on HuggingFace often list real RAM requirements.
  • Tiny 7–8B models on phones or low‑RAM laptops are often judged “tinkering only” for serious coding, though some report acceptable work use with careful model choice.
  • Tools discussed: Ollama, llama.cpp (including distributed/RPC mode), ktransformers for MOE offload, Intel’s IPEX‑LLM, plus various UIs (OpenWebUI, LibreChat).
  • Overall sentiment: homelab AI is fun and educational, but for most people it’s still a niche, fast‑moving, anecdote‑driven space rather than a clear economic win.

SQL pipe syntax available in public preview in BigQuery

Origins of pipe syntax

  • Several comments trace pipe-style composition back to Unix pipelines in the 1970s.
  • The specific |> function-pipe operator is linked to ML-family languages (Isabelle/ML, F#, OCaml) from the 1990s, later popularized in F# and then picked up by Elixir.
  • People distinguish Unix “bytes in, bytes out” pipes from higher‑level, function-chaining pipes in functional languages.

Comparisons to other pipe / query languages

  • Kusto Query Language (KQL) is repeatedly cited as a strong precedent; some call it “the best query language,” others see it as only an incremental improvement over SQL.
  • PRQL and Malloy are mentioned as cleaner, pipe-oriented languages that compile to SQL; PRQL’s function system and extensibility are highlighted, as is a DuckDB extension.
  • LogsQL, ClickHouse’s KQL mode, and LookML/Malloy are given as further examples of pipe-based or transpiled query syntaxes.
  • Some note that InfluxDB’s Flux (also pipe-based) was dropped due to complexity and low adoption.

Perceived benefits of BigQuery’s pipe SQL

  • Many are enthusiastic: they find the stepwise, linear workflow more natural for data exploration and cleaning, especially for iterative filtering, aggregation, JSON/array handling, and DBT macros.
  • Advantages cited:
    • Unifying WHERE/HAVING/QUALIFY.
    • Reducing nested CTEs and ugly subqueries; easy to comment out or reorder stages.
    • Closer to data-frame/tidyverse / KQL mental models and easier integration with host languages.
    • Table-valued functions plus pipes allow reusable higher-order transformations.

Concerns and drawbacks

  • Some worry this further fragments SQL dialects and increases complexity; they’d prefer higher-level languages that transpile to SQL.
  • Others argue transpilers/ORMs have leaky abstractions and poor optimization compared to engine-native syntax.
  • A few find pipeline style harder to read than modular CTEs because it increases “state in your head,” especially for long chains; best practice suggestions include ending with an explicit SELECT to clarify outputs.
  • There is debate over whether this is essentially “SELECT * in disguise” and whether overuse will harm clarity.
  • Some simply prefer traditional SQL aesthetics and don’t feel a need for pipes.

Standardization and adoption

  • Several commenters hope this syntax spreads to Postgres, MySQL, SQLite, and note early support in Databricks and experiments in SQLite and DuckDB.
  • Others argue the real win would be a standardized, engine-aware pipelined query language used across databases, rather than many slightly different SQL extensions.

Nvidia Security Team: “What if we just stopped using C?” (2022)

Scope and Reality of “Stopping C” at Nvidia

  • Several commenters note the title is misleading: Nvidia did not drop C wholesale; SPARK/Ada is used in specific, security‑critical components.
  • Case study material (linked in thread) indicates SPARK is used mainly for:
    • GPU firmware image authentication and integrity checks
    • BootROM and secure monitor firmware
    • Parts of an isolation kernel for an embedded OS on a RISC‑V root‑of‑trust core
  • Overall usage is described as “significant in important places” but a tiny fraction of Nvidia’s total codebase; most drivers and CUDA stack remain C/C++.

Marketing vs Engineering Substance

  • Some see the piece as AdaCore marketing, arguing external consultants can recommend radical rewrites without bearing migration cost.
  • Others counter that multiple Nvidia security engineers, training investment, and multi‑year adoption suggest real internal buy‑in, not just PR.

Why SPARK/Ada for Firmware

  • SPARK offers:
    • Strong typing, range types, and rich numeric/packet modeling
    • Formal proofs of absence of run‑time errors and selected functional properties
  • This is seen as a good fit for small, stable, security‑critical firmware and isolation kernels, less for large, fast‑moving codebases like drivers or CUDA libraries.

Syntax, Approachability, and Interoperability

  • Debate over Ada’s “unapproachable” syntax:
    • Some argue syntax is a red herring compared to safety/contracts; others claim a curly‑brace dialect would greatly boost adoption.
  • Interop concerns:
    • Ada interoperates well with C, and to a degree C++, but heavy modern C++ (STL, unique_ptr, templates) complicates clean boundaries.
    • Serialization/deserialization at language boundaries can add overhead and complexity; this is framed as a general multi‑language problem, not Ada‑specific.

Formal Verification vs Testing and C

  • Some commenters applaud emphasizing provability over testing for high‑integrity code; others warn verification only proves conformance to a (possibly flawed) spec and doesn’t cover timing side channels, microarchitectural flaws, etc.
  • Comparisons raised:
    • SPARK vs Rust: SPARK can express and prove more properties (e.g., range types, functional proofs), while Rust primarily targets memory safety.
    • Verified C paths (CompCert, Frama‑C, proprietary tools) are mentioned; several people who tried both say SPARK’s type system makes proofs easier than for pointer‑heavy C.
  • Skeptics doubt any language will truly “replace C” for OS/low‑level work; they see this more as adding a verified island inside a C sea.

AI, Proof Assistants, and the Future

  • Some argue the real future path to pervasive verification is AI‑assisted proof and code generation, not niche languages.
  • Others are cautious:
    • Current LLMs struggle with non‑mainstream domains (Ada/SPARK, advanced math, borrow checkers) and often hallucinate APIs or incorrect arguments.
    • AI‑generated code in non‑verified languages may increase subtle concurrency and logic bugs if developers over‑trust completions.
  • A promising niche proposed: AI as a helper for generating verification annotations and boilerplate in already‑formal ecosystems (SPARK, Coq‑based systems), where results are still checked by deterministic tools.

GPU Security and Threat Model

  • Some initially question why Nvidia needs strong security: “they just make graphics cards.”
  • Counterpoints:
    • Modern GPUs carry RISC‑V roots of trust akin to T2/ME/PSP, enforce secure boot chains, isolation, and feature segmentation.
    • GPUs are central to national‑scale AI infrastructure and used in safety‑critical domains (automotive, drones); compromise of firmware or RoT is high‑impact.
  • Discussion touches on GPU rootkits and DMA capabilities; concerns about “kludgey” GPU architectures and mailbox/VRAM transfer bottlenecks surface.

SPARK Tooling, Licensing, and Activity

  • Clarifications:
    • SPARK and GNATprove are open source and freely available; commercial “SPARK Pro” mainly adds support, newer releases, and “wavefronts” (details not fully explained in thread).
    • The project’s commit history is cited as evidence it’s actively developed, countering claims of being “largely untouched.”
  • Alire is highlighted as the de‑facto package manager/distribution channel for the free toolchain.

Language Design and Safety Features

  • Ada/SPARK features praised:
    • Range types, derived types, subtype predicates, and bounds‑checked arrays/strings that prevent many classes of errors by construction.
    • Ease of modeling bit‑packed registers and network/packet formats compared to many mainstream languages.
  • Some view these strong types as “creepy” or alien compared to plain integers; fans argue they eliminate whole bug categories and are closer to what safety‑critical domains need.

Alternatives and Broader System Ideas

  • Suggestions include:
    • Using verifiable C subsets and verified compilers instead of switching languages.
    • Avoiding opaque binary blobs entirely by distributing intermediate representations and compiling on‑device; critics label this unrealistic due to performance, IP, and ecosystem hurdles.
  • There is recurring tension between “ideal” verified stacks and messy real‑world constraints (existing C++ ecosystems, proprietary drivers, hardware quirks).

Rust kernel policy

Governance and Policy Clarity

  • Several commenters feel the Rust-for-Linux effort lacks a clear, enforced policy: who decides what, what maintainers must accept, and how cross-language interfaces should work.
  • Some argue Linux’s distributed governance (subsystem maintainers with a lot of autonomy) makes it hard to impose a coherent Rust strategy without stronger direction from the top.
  • Others counter that Linus’s approach—allow Rust, but let subsystem maintainers decide—is appropriate given project size and complexity.

Linus’s Role and Maintainer Conflicts

  • Strong criticism that the current “political mess” is largely due to Linus approving Rust without clearly defining what that means in practice.
  • One side wants him to either explicitly back Rust (and overrule blocking maintainers) or clearly say “no Rust” to avoid wasting effort.
  • Others reply that he must trust maintainers, can’t centrally micromanage every subsystem, and that some behind-the-scenes coordination is already happening.

Technical and Toolchain Issues

  • Concerns about mixing GCC (for C) and LLVM/Clang (for Rust and bindgen) in kernel builds; some kernel developers historically oppose mixed toolchains.
  • Rust’s use of unstable language features is seen as a risk; defenders note that:
    • Kernel has long depended on non-standard GCC extensions.
    • Rust and kernel teams coordinate to stabilize what the kernel needs.
  • Questions around static linking of libgcc and long-term compatibility of Rust-generated code.

Rust vs. C: Safety, Complexity, and Culture

  • Pro-Rust arguments:
    • Memory safety is increasingly mandated (governments, industry); Linux must compete.
    • Rust is easier to write correctly than C; C is “simple to learn, hard to use safely.”
    • Rust is particularly valuable for drivers and DMA-safe abstractions.
  • Skeptical views:
    • Rust feels complex/“PhD-level” to some; fear existing C maintainers won’t adopt it.
    • Cross-language codebases and bindings are called “cancerous” or maintenance nightmares.
    • One commenter argues Linux’s unstable APIs, monolithic design, and GCC focus make deep Rust integration fundamentally mismatched; a full Rust OS (e.g., Redox) is seen as cleaner.

Project Identity and Support

  • Confusion over who “we” is in the Rust-kernel policy; some want clearer attribution and authority.
  • Clarified in-thread: Rust-for-Linux is led by kernel folks, not the Rust Foundation, but the Rust project treats “R4L on stable Rust” as a flagship goal and actively supports needed language work.
  • There is debate over naming individuals publicly due to concerns about harassment and misplaced responsibility.

Long-Term Outlook

  • Some predict old-guard C maintainers may eventually leave as Rust use grows and new generations of developers arrive.
  • Others think Rust will remain limited to drivers and select subsystems, never supplanting C wholesale.
  • Overall, commenters see high technical promise but unresolved governance, social, and tooling tensions.

The state of Rust trying to catch up with Ada [video]

Why Ada Didn’t “Win” Broad Adoption

  • Historical timing: universities were teaching Fortran, Simula, Pascal, etc. long before Ada; if Ada was your “first language” that likely means a late-starting CS program.
  • Unix shipped with C and its own source; C was easy to bootstrap and port, Ada compilers were heavy, slow, and often expensive extras on Unix.
  • Commercial compilers: high prices, opaque licensing, and hardware demands limited hobbyist and startup use.
  • Ecosystem bias: Ada positioned as “mission critical / aerospace / defense,” not for web, games, or general apps, so it never built broad mindshare.
  • Chicken‑and‑egg: mainly “specialists” use it, so tooling and UX for newcomers lagged, which in turn discouraged wider adoption.
  • ACATS public conformance suite is both a strength and barrier: a new compiler starts with thousands of failing tests and no customers until it passes.

Safety Features, Subtypes, and Pattern/Liquid Types

  • Ada has long had range types and subtypes (and later dynamic predicates), allowing types like “angles 0..360” or “UUID_String” with embedded format checks.
  • These constraints can be enforced at runtime and, with SPARK/verification tools, often proven at compile time.
  • Rust’s “pattern types” / liquid types proposal aims for similar refinement-style types, expressed as patterns over existing types.
  • Current Rust niches (e.g., NonZeroI8, Option<&T> occupying same size as T) rely on compiler magic; pattern types would let users define such optimized, constrained types themselves (e.g., “BalancedI8” excluding INT_MIN).
  • Debate over utility: some see runtime-checked subranges as crucial for invariants; others find them historically underused or awkward without good verification tooling.

Rust vs Ada: Domains, Tooling, and Safety-Critical Use

  • Many see Rust and Ada as overlapping in “C/C++ replacement” and safety‑critical domains; Ada additionally has deep roots in hard real‑time control loops where post‑init allocation is forbidden.
  • Others argue Ada’s memory management model makes “browser‑like” workloads awkward, while Rust is built for heavy dynamic allocation with graceful failure.
  • Safety‑critical industries require certified/qualified toolchains; Ada vendors already offer such compilers and proof tools, largely proprietary.
  • Rust is building similar infrastructure (Ferrocene, safety‑critical consortium, MISRA‑aligned guidelines, model checkers like Loom).
  • Integer overflow semantics are contrasted: Ada treats overflow as a constraint violation; Rust uses debug‑panic vs release‑wrap plus explicit checked/saturating APIs, which some consider too opt‑in.

Syntax, Ergonomics, Adoption, and Community

  • Strong disagreement over readability: some find Pascal/Ada “BEGIN/END” clearer than C braces; others say that syntax “hurts their eyes” and prefer C‑style.
  • Rust deliberately adopted C‑like surface syntax as “weirdness budget” management to ease onboarding; internally it borrows heavily from ML/FP ideas.
  • Several commenters argue Rust’s success owes more to marketing, large‑company backing, platform roles (like Swift/Kotlin), and ecosystem investment than to purely technical merit.
  • Rising interest in Rust and “memory‑safe C replacements” has renewed attention to Ada as prior art; some are annoyed by Rust evangelism that appears to rediscover decades‑old ideas without credit.
  • Rust’s learning curve is widely acknowledged as steep; advice includes treating it initially like a mostly immutable, value‑passing language and adding advanced features later.