Hacker News, Distilled

AI powered summaries for selected HN discussions.

Page 286 of 531

Fully homomorphic encryption and the dawn of a private internet

Performance and scalability limits

  • Many commenters argue current and foreseeable FHE is orders of magnitude slower (often cited ~1000×, possibly far worse) than plaintext due to bootstrapping and huge ciphertexts.
  • Even without bootstrapping, ciphertext blowup (often ~10³× larger) implies massive extra memory bandwidth and compute that hardware advances alone are unlikely to erase.
  • Latency impacts are framed as unacceptable for most user-facing tasks (e.g., milliseconds → seconds/minutes; 30s → hours).

Search, databases, and index problems

  • Fully private search over large indexes is highlighted as especially hard:
    • Naively, encrypted-key search is O(n) instead of O(log n); recent PIR-style schemes with polylog(n) queries require extreme preprocessing and storage blowups (petabytes for modest DBs).
    • For “FHE Google,” either the server must encrypt huge indexes per client or do near-linear work per query, both viewed as impractical.
  • Some systems mix FHE/PIR with partial leakage (e.g., hashed prefixes, subsets, anonymous networks), trading strict privacy for performance.

Use cases and economics

  • Strong consensus: generic “private internet” via FHE is not economically viable soon. A privacy‑first Google‑scale service would be vastly more expensive and far slower; few users would pay enough.
  • People often prefer free, data‑harvesting services; privacy‑preserving alternatives already exist but remain niche.
  • FHE is seen as promising for narrow, high‑value, low‑throughput tasks: finance, regulated data sharing, some medical or government/military computations, “data owner vs model owner” scenarios.

Alternatives to FHE

  • Confidential computing (TEEs: SGX/TDX/SEV/ARM TEE) is repeatedly described as the only realistic way to do private LLM inference and many cloud workloads, despite hardware‑trust issues and past breaks.
  • Specialized searchable encryption and PIR schemes can give near‑plaintext search performance for specific patterns, with FHE reserved for small filtered subsets.
  • Self‑hosting and encrypted backups are presented as a simpler, cheaper privacy route for many personal data uses.

How FHE works & security nuances

  • Multiple explanations outline homomorphism: operations on ciphertext correspond to operations on plaintext, enabling arbitrary circuits via additions/multiplications.
  • Some struggle with the intuition that “being able to compute means weaker encryption”; others note:
    • FHE is malleable (not NM‑CCA2) but can still be IND‑CPA/semantically secure.
    • Computation may leak allowed structure (e.g., which operations were run) but not plaintext, and “circuit privacy” research aims to hide even that.

Privacy, incentives, and politics

  • Commenters doubt big providers will voluntarily adopt FHE that blocks data harvesting without strong regulation or market pressure.
  • Governments may resist or demand backdoors; export controls and surveillance incentives are seen as major non‑technical obstacles.

NIH is cheaper than the wrong dependency

Paid Dependencies, Vendor Lock‑In, and Longevity

  • Several comments argue paid dependencies are often risky: incentives for lock‑in, single support source, and danger when the vendor folds or is acquired.
  • Counterpoint: paying can be good if it guarantees people whose job is to maintain and support the code; open source can also abandon users.
  • Many recommend only paying for components that implement open standards with alternative implementations, to preserve an escape route.

Dependency Minimalism, Vendoring, and NIH

  • Strong support for “dependency minimalism”: fewer moving parts, less that can break, easier long‑term maintenance.
  • Common heuristic: “Can I vendor it?” If yes, it may be better to copy small libraries into your tree and own them. Vendoring is framed as accepting maintenance responsibility, not just snapshotting.
  • Forking or vendoring dependencies (and even their transitive deps) is suggested to avoid disappearing repos and to understand build chains.
  • Others argue a strict “zero dependencies” stance is overreaction: most small, well‑scoped libraries are harmless and save time.

Evaluating Dependencies: Risk, Scope, and Lock‑In

  • Suggested process: only use OSS, review license, ubiquity, community health, security history, and ease of replacement.
  • Only adopt dependencies you’d be willing and able to maintain or fork. Some advocate preemptive forking; others say that’s too heavy and prefer vendoring or caching.
  • Ubiquity and scale are seen by some as a safety signal; others report severe bugs and slow responses even in big‑company libraries.

Where NIH Makes Sense (and Where It Doesn’t)

  • NIH is seen as appropriate when:
    • You need only a narrow subset of functionality, can implement it in ≤ a day, and want tight control.
    • The domain is core to your business (e.g., a custom ledger, highly tuned storage, niche graph DB, embedded time‑series engine).
  • Generally discouraged for: databases, cryptography, OSes, or large game engines—unless you truly have unique constraints and deep expertise.

AI and NIH

  • Some teams use LLMs to generate internal tools or codegen utilities instead of adding runtime dependencies.
  • Concerns: AI‑generated clones of libraries may inherit old CVEs, introduce subtle bugs, and lack automated vulnerability tracking.

Organizational and Human Factors

  • Dependencies can reduce time‑to‑market, which often dominates business decisions.
  • Long‑lived or safety‑critical systems (energy, finance, compliance) tend to avoid sprawling dependency trees due to security, audit, and maintenance costs.
  • Multiple commenters stress that hidden long‑term costs of upgrades, security fixes, and lock‑in are routinely underestimated.

USB-C hubs and my slow descent into madness (2021)

One‑cable dream vs messy reality

  • Many comments echo the article’s theme: people want a single USB‑C cable for power, display, network, and peripherals, but get there only after multiple failed hubs/docks.
  • DisplayLink-based docks are seen as a compromise: they enable many monitors (useful for M1 Macs) but introduce compression artifacts and DRM/streaming issues.
  • Thunderbolt 3/4/USB4 docks are praised as the only consistently reliable option, but they’re expensive, often need proprietary power bricks, and sometimes still have firmware or refresh‑rate quirks.
  • Monitor-integrated USB‑C hubs and daisy-chained displays are viewed as the cleanest desk setup, but are limited by macOS (no DP MST on some MacBooks, scaling options).

OEM/ODM reality and Realtek skepticism

  • Multiple hubs turn out to be the same OEM “Goodway” design rebranded at different prices. This is described as standard industry practice, not exceptional.
  • Realtek USB NICs are frequently blamed for flakiness (driver issues, feature gaps, weird failure modes like loops or pause frames killing networks), especially on Linux/BSD.
  • Others argue the silicon is often fine and the real problem is low-end manufacturers cutting corners on integration and firmware.

USB‑C spec complexity and broken devices

  • Long subthread on devices that only charge with their original USB‑C cable: explanation centers on missing or incorrect CC resistors and lack of USB‑C “dead battery mode,” making them depend on out‑of‑spec A‑to‑C cables that always supply 5V.
  • Some note that many products (including well-known brands and even Raspberry Pi revisions) have shipped with non‑compliant USB‑C implementations.
  • Debate over whether USB‑PD’s “0V by default until negotiated” is a design mistake or necessary to avoid dangerously tying two power sources together.

Power negotiation and PD hubs/chargers

  • USB‑PD power strips/hubs often momentarily cut power to all ports when a new device is plugged in, hard‑rebooting things like Raspberry Pis.
  • People report this behavior even on modern GaN chargers from major brands; opinions differ on whether it’s unavoidable or just poor design.
  • Some call for user-selectable behavior (e.g., a switch to favor existing loads vs new ones).

Cable capabilities, labeling, and testing

  • A recurring frustration is not knowing what any given USB‑C cable supports (data rate, video, PD wattage, Thunderbolt).
  • Suggested mitigations: personal labeling systems, USB testers/analyzers, and relying only on clearly logo‑marked cables.
  • Some see variable cable capabilities as a spec failure; others argue it’s a necessary tradeoff for backward compatibility, cost, and cable length flexibility.

Hub design gaps

  • Many consumer hubs provide mostly USB‑A and just one usable USB‑C data port, contradicting the “USB‑C everywhere” vision.
  • A few niche products (USB‑C–only hubs, Thunderbolt/USB4 hubs with multiple downstream C ports) are mentioned, but they’re rare or pricey.
  • There is interest in more integrated solutions: docks in monitors, desks, floor boxes, and better “USB power strips” for charging only.

My favorite use-case for AI is writing logs

AI for logging and boilerplate

  • Many commenters like using AI for tedious, low-level tasks: log lines, docstrings, CLI glue, and repetitive boilerplate.
  • They see value in “sketch the logic, let AI add the polish,” especially across many languages/projects where logging patterns differ.
  • Others argue that if logging is truly that mechanical, macros, snippets, or AOP-like approaches should handle it deterministically instead of LLMs.

Cognitive load, tooling, and determinism

  • One camp emphasizes that reducing cognitive overhead (remembering exact logger APIs, formatting, project-specific conventions) meaningfully improves focus and flow.
  • The opposing camp argues this is what editors, LSPs, snippets, and static tools are for; an LLM is an overkill “fuzzy hammer” where deterministic autocomplete and templates are more reliable and energy-efficient.
  • There’s tension between favoring highly predictable tools vs accepting probabilistic assistance that must be verified.

Programming as craft vs problem-solving

  • Some see complaining about logging syntax as disliking “real programming” and basic competence. They take pride in carefully hand-written, context-rich logs.
  • Others separate “liking programming” from liking its minutiae: they’d rather focus on solving business problems than on remembering logging variants, and view AI as another abstraction layer like garbage collection or sort routines.
  • Broader discussion touches on career motivations: many developers are in it for pay and outcomes, not love of the craft; AI shifts value from “knowing how” toward “knowing what you want.”

Logging best practices and pitfalls

  • Multiple comments critique the use of Python f-strings in logs:
    • They eagerly format even when the log level filters them out, harming performance in hot paths.
    • They break aggregation in tools like Sentry that rely on static format strings.
  • Suggested alternative: classic parameterized logging ("msg %s", var) or libraries that support lazy {} formatting.
  • Some highlight log level discipline (avoiding noisy or misleading error logs) and warn against logging secrets such as Redis URLs.

Libraries, models, and reliability concerns

  • JetBrains’ local models and Cursor-style completions are praised as helpful, but also noted for occasionally plausible-yet-wrong suggestions.
  • Debate over loguru: some like its ergonomics and features; others dislike its nonstandard formatting syntax and backtrace style.
  • Several caution that AI-written logs and code must still be reviewed; unreliable or misleading logs can severely complicate debugging.

People kept working, became healthier while on basic income: report (2020)

Effects of basic income on work and life choices

  • Many commenters note that multiple pilots (including this one) consistently report: reduced stress, better health, and most recipients continuing to work.
  • Critics emphasize the flip side: if ~75% “kept working,” then ~25% stopped, despite knowing the program was temporary; they see this as a large effect on labor supply.
  • Supporters counter that about half of those who stopped working went back to school, or into caregiving or parenting, which they view as socially valuable, not “idleness.”
  • There is deep disagreement on whether most people would still choose to work under a lifelong UBI: some cite pensions and early retirees as evidence many will quit; others argue UBI amounts are “basic,” not enough to fund a comfortable retirement.

Short-term pilots vs long‑term, universal UBI

  • Many argue that time‑boxed, non‑universal pilots cannot show true long‑term behavior: people assume payments will end and act differently than under a lifelong guarantee.
  • The specific Ontario/Hamilton study is criticized for: small, self‑selected survey (217 of ~4,000), no proper control group, advocacy-style presentation, and a design that clawed back benefits at 50% of earnings (more like a negative income tax than true UBI).

Funding, taxes, and macroeconomic risks

  • Concerns: a real UBI would cost on the order of trillions annually, require major tax increases, and might reduce labor supply at both ends (recipients and high earners).
  • Some fear inflation, especially in rents and essentials, as landlords and producers capture extra cash; others reply that UBI is largely a redistributive transfer, not new money, so aggregate demand need not surge.
  • Debate over whether UBI could replace existing programs (welfare, SNAP, disability, Social Security, Medicaid) or whether healthcare and high‑variance needs still require separate systems.

Targeted welfare vs universal cash vs alternatives

  • Some prefer expanding the Earned Income Tax Credit / negative income tax as a work‑incentivizing “UBI-lite.”
  • Others argue existing means‑tested systems are complex, stigmatizing, and create benefit cliffs; universality simplifies administration and reduces perverse incentives.
  • A competing proposal is guaranteed public employment at a living wage; critics say that’s hard to design, match to abilities, and may devolve into make‑work bureaucracy.

Moral framing and political resistance

  • One camp frames UBI as empathy and social insurance in a surplus society; another emphasizes “forced redistribution” and selective empathy (“my tax dollars”).
  • There is recurring reference to “welfare queen”–style narratives, and disagreement on whether UBI should be justified by compassion, self‑interest (reduced crime/health costs), or system‑stability arguments.

Anthropic tightens usage limits for Claude Code without telling users

Usage patterns and how limits are hit

  • Some users on $20–$100 plans hit limits in 1–3 hours, even with a single Claude Code instance and modest prompts; others on the same plans never come close and are surprised.
  • Heavy users describe:
    • Long-running “agentic” workflows with sub-agents, multi-step planning docs (hundreds of lines) and implementation phases, often costing the API-equivalent of $15–$45 per feature.
    • Multiple parallel Opus/Sonnet sessions (4–8+) running for hours, or even 24/7, on tasks like large refactors, migrations, debugging, test fixing, data analysis, etc.
    • Workflows where Claude repeatedly re-reads large files or entire folders, causing big token burn.
  • Others see frequent 429/529 errors or early downgrades to Sonnet and suspect dynamic throttling by time of day, account, or region.

Pricing, transparency, and business model

  • Many complain that limits were tightened silently, with confusing in-product messaging and no clear remaining quota indicators; some only infer costs with tools like ccusage.
  • There’s broad agreement that $100–$200 “Max” subscriptions can easily yield hundreds or thousands of dollars of API-equivalent usage, implying heavy subsidization.
  • Competing narratives:
    • “Uber/drug dealer model,” “enshittification”: underprice to hook users, then ratchet limits/prices.
    • Counterpoint: this is rational loss-leading in a space where compute costs will fall and models will get more efficient.
  • Some see flat-fee “unlimited” plans as inherently unsustainable and expect eventual convergence on metered pricing or stricter caps.

Productivity gains vs. skepticism and overreliance

  • Enthusiasts say Claude Code massively boosts throughput (often 2–3×), offloads boilerplate, accelerates learning of new patterns, and enables experimentation across stacks and domains.
  • Others find it boring, brittle, or slower than simply coding and searching themselves; several report loops, wasted tokens, or degraded code quality that then demands heavy human cleanup.
  • There’s a cultural clash around “vibe coding”:
    • Critics worry about skill atrophy and about projects becoming impossible when limits hit.
    • Supporters argue that as long as you understand and review the code, it’s a power tool, not a crutch—and that not using LLMs at all is now self-handicapping.

Lock-in, reliability, and alternatives

  • Some users now fear dependence on a third‑party tool with opaque, moving limits and compare it unfavorably to owning a keyboard or compiler.
  • Others rotate between providers (Claude, Gemini, Kimi, etc.) or look to local/open models (Llama 3.x, DeepSeek, R1 + tools like Goose/Ollama) to mitigate vendor risk.
  • Several note growing reliability and UX issues in Claude clients (slow history, cryptic errors, poor usage visibility) and ask Anthropic to prioritize stability and clearer quota communication.

The daily life of a medieval king

Reliability and Purpose of the Account

  • Several comments question how literally the king’s “day in the life” can be taken.
  • The text is seen as partly idealized and didactic: a model of how a “good” king should behave, possibly contrasted with a mentally ill successor.
  • Some liken it to modern PR / lifestyle pieces that smooth over messy reality.
  • Others frame it as medieval propaganda: informative about ideals, but weak as a strict diary of actual behavior (e.g., daily church attendance, listening to commoners).

Comparison with Other Medieval Rulers

  • This king is described as relatively sedentary; other monarchs spent more time hunting, on campaign, in mobile courts, or at seasonal residences.
  • Discussion notes that many medieval rulers weren’t linguistically or culturally aligned with their “national” subjects (e.g., French‑speaking English kings).
  • Legal reforms like mandating contracts in plain English are cited as examples of practical, competence‑oriented governance.

Daily Routine, Wine, and Leisure

  • Commenters highlight the late start, limited work hours, and structured mix of piety, audiences, and rest.
  • The morning glass of light, “well cut” wine draws attention; multiple replies clarify this meant wine diluted with water, a common historical practice.
  • The jewel‑admiring segment is compared to modern car or collectibles culture—time with close companions plus conspicuous luxury.

Modern vs Medieval Living Standards

  • A substantial subthread compares a 15th‑century king to someone in today’s global 5th wealth percentile.
  • Consensus: modern poor almost certainly have longer life expectancy and better medical options, but a king had unmatched food security and status (though tightly constrained life choices and obligations).
  • Several people reflect on “living like a king” today via indoor plumbing, refrigeration, varied diets, and drastically lower child mortality.

Health, Life Expectancy, and Medicine

  • Cited data for medieval English monarchs gives an average death age in the 40s; commenters stress how much this overstates life expectancy at birth due to high child mortality.
  • Others emphasize that even elites were vulnerable to infections that are trivial today and to violent deaths in war, hunting, or coups.
  • There’s debate over painkillers: one side claims “no painkillers,” others note long‑known remedies like opium, nightshades, willow bark, and herbal medicine.

Food Security, Preservation, and Diet Variety

  • Kings had guaranteed food but limited by pre‑Columbian ingredients and seasonality.
  • Participants list extensive preservation techniques (salting, pickling, drying, smoking, fermenting, sugaring, confit) and long‑keeping crops, arguing winter diets could still be varied by their standards.
  • A side debate challenges the common trope that medievals avoided water in favor of alcohol and disputes the idea that people were constantly drunk.

Global Hunger and Malnutrition Debate

  • One thread claims hunger has worsened for ~25 years and cites large annual death tolls from hunger and related disease.
  • Others contest this, pointing to data showing decades of improvement, a plateau, and only recent reversal; they demand more precise sourcing for large mortality numbers.
  • Malnutrition’s changing nature is noted: undernutrition now coexists with rising obesity, even in poorer countries.

Work Hours, “Four‑Hour Days,” and Fulfillment

  • Some note the king seems to do ~4 hours of overt “work,” comparing this to modern managers or knowledge workers.
  • Others argue much of what looks like leisure (religious services, public displays, audiences) was in fact required labor to maintain legitimacy.
  • There’s a parallel drawn to hunter‑gatherer “original affluent society” claims, with pushback that “work” is often defined too narrowly in those arguments.

Power, Delegation, and Productivity

  • A long branch contrasts how kings/CEOs accomplish much via delegation versus ordinary individuals.
  • Commenters debate the value of human secretaries versus tools like email and calendars: some see tech as clearly superior; others emphasize that a good assistant exercises judgment and acts as a force multiplier.
  • The underlying point: people with large staffs and authority can “get a lot done” without personally doing most tasks.

Monarchy, Government Ideals, and Contact with Subjects

  • Several note that this king appears to spend more time hearing petitions than some modern politicians do with constituents, though population and logistics were very different.
  • Comparisons are made between monarchs and mafia dons in terms of personal rule and patronage.
  • A philosophical thread argues that absolute monarchy has very high variance (best and worst regime type), while democracies are more middling but safer.
  • Others draw an analogy to modern civics teaching: idealized descriptions of how systems should work (separation of powers, accountability) often diverge sharply from practice, yet revealing those ideals is still historically informative.

Ask HN: What Pocket alternatives did you move to?

Major Alternatives People Chose

  • Wallabag is the most-cited replacement.

    • Reasons: no ads, Pocket import works well, self-hosting possible, good content extraction, ePub export (especially for Kobo), KOReader integration.
    • Criticisms: Android/iOS apps feel basic/outdated; Firefox-based server-side scraping sometimes fails; self-hosting can be heavy for weaker NAS boxes.
  • Instapaper frequently mentioned.

    • Pros: simple, stable, “doesn’t try to do too much,” direct Pocket import, Kindle digest, IFTTT support.
    • Several people say it’s far ahead of Pocket in handling paywalled content; others had previously left Instapaper for Pocket.
  • Raindrop.io also popular.

    • Pros: smooth Pocket import, good tagging, clean UI, open API; free tier sufficient for some.
    • Used both personally and for work, sometimes alongside other tools.
  • Readwise Reader / Readwise.io

    • Seen as “power user” oriented; strong highlighting and TTS, good on e‑ink, praised ongoing development.
    • One person left it due to ugly/unchanging UI.

Self-Hosted & Open-Source Options

  • Wallabag, Readeck, Karakeep, Shiori, Linkding, Linkwarden, Readeck+Kobo mod, and others are used by people who don’t trust hosted services to persist.
  • Karakeep praised for AI tagging + Meilisearch, but criticized as immature: SQLite-only, limited export options, tag overload.
  • Shiori liked for simplicity, multi-DB support, and local copies.

Native / Minimalist Approaches

  • Some abandon third‑party services entirely:
    • Browser bookmarks + Reading List, or RSS readers’ built‑in “read later.”
    • iOS tricks: Safari Reader + print-to-PDF, or just Safari Reading List (with debate over offline reliability and PDF usability).
    • Plain-text lists, email-to-self workflows.

Knowledge-Management & Obsidian-Based Setups

  • Several use Obsidian (web clipper, ReadItLater, Slurp, Relay) to store cleaned markdown locally, often synced across devices and used collaboratively.
  • Others fold articles into broader tools like Notion, DEVONthink, Zotero, or Eagle.

New / Niche Services & Personal Projects

  • Mentioned: Fika, Folio, Cubox, Curio, Fabric, CouchReader, Histre, Mozaic, Veo3, link.horse, linksort, bukmark, simple PWAs and Python/Docker apps, ArchiveBox lists.
  • Some actively building Pocket replacements; questions arise about offline storage, export, APIs, and e‑reader/Firefox support.

Meta Reflections

  • Debate over “just Google pocket alternatives” vs real-world feature gaps.
  • Many admit they rarely read what they save; others argue long-term bookmarking is still highly valuable.

Nintendo Switch 2 account bans continue: warning after buying old copy of Bayo 3

Scope of the “Ban” (Brick vs Online Lockout)

  • Several commenters stress the console isn’t technically bricked: it loses online services (multiplayer, updates, redownloads) but still runs existing games.
  • Others argue that for people who specifically bought the device for online play or digital access, this is “effectively bricked,” functional difference or not.
  • There’s confusion over whether banned users can redownload purchased titles or only keep what’s already installed; behavior seems restrictive and is viewed as “very heavy” even if not full bricking.
  • Some point out that most people cannot or will not install custom firmware as a workaround.

Used Games, MIG Flash, and Who Gets Punished

  • A key concern: Nintendo bans hardware when the same cartridge ID is seen multiple times (e.g., MIG Flash dumps), but the system can’t reliably tell pirate from innocent second-hand buyer.
  • Many see punishing both sides (or any side when it’s ambiguous) as morally unacceptable; some say Nintendo should punish neither if they can’t distinguish.
  • Commenters recall prior warnings that buying used Switch 2 carts could become a “roulette” because of this detection.
  • There’s disagreement about the specific incident:
    • Some say this case is clearly tied to the owner using a piracy-focused device and is being misrepresented as a used-game problem.
    • Others cite reports of innocent users being banned over used carts and bans later being reversed, suggesting a real risk. Exact timeline/evidence is described as unclear.

Resale, Digital Lock‑In, and Trust

  • Many see this as part of a long pattern: hostility to modding/piracy, erosion of resale rights, and digital purchases tied to fragile online services.
  • Shutting down Miiverse and older eShops is cited as evidence Nintendo doesn’t honor long-term digital ownership or platform features.
  • Some predict or fear a shift toward carts as mere license keys, further undermining physical media.

Consumer Reactions and Alternatives

  • A sizable group vows to skip Switch 2, avoid online-dependent games, or rely on emulation/PC/Steam Deck instead.
  • Others argue Nintendo still has no real substitute: unique first‑party games, kid‑friendly ecosystem, portability, and massive sales suggest most buyers accept — or ignore — these risks.

The patterns of elites who conceal their assets offshore

Moral views on offshoring and tax evasion

  • Many see elites’ offshore concealment as straightforward theft from society: they benefited from public systems then avoid contributing back.
  • Others argue that rich actors are primarily protecting themselves from “state expropriation” (asset seizure, arbitrary taxation, nationalization), not street crime.
  • There’s tension between “end corruption” vs “let me use the same tricks”; some lament that many commenters want access to loopholes rather than enforcement against them.
  • A minority defend the right to “escape” the state if one can, framing the state as just another power bloc using coercion.

Wealth, value, and exploitation debates

  • Long subthread on whether wealth creation is zero-sum:
    • One side: value depends on finite resources/energy; creating wealth in one place effectively devalues something elsewhere.
    • Other side: value is not bounded by resources; technology and productivity massively expanded living standards, so “the pie can grow.”
  • Arguments over whether all large fortunes are inherently exploitative: some say no individual can morally earn billions; others emphasize risk-taking, capital, and productivity multipliers (e.g., “lazy guy with an excavator” outproducing many with shovels).
  • Transactions as value-creating vs value-destroying:
    • One camp: every voluntary transaction creates mutual surplus.
    • Critics point to fraud, lemons, advertising arms races, and crime as transactions that destroy or externalize value.

Mechanics and incentives of offshore finance

  • Offshore structures serve tax minimization, asset protection, and anonymity: shielding against governments, lawsuits, spouses, and even heirs.
  • For the ultra-wealthy, funds often never “re-shore”: offshore entities directly purchase companies, jets, yachts; individuals then lease or borrow against these assets.
  • Repatriation sometimes happens via political “tax holidays,” especially in the U.S.
  • Some note ordinary pensions and non-criminal investors also route via offshore hubs for neutrality and legal predictability.

Enforcement, politics, and journalism

  • Skepticism that elites who control enforcement (politicians, legislatures, courts) will meaningfully crack down on havens they themselves use.
  • Discussion of blacklists: current EU list is small and omits major players (US, UK, Switzerland, Luxembourg, China), likely due to political power.
  • Investigative consortia (e.g., Panama Papers) are praised as crucial; collaboration is partly for safety after journalists have been killed over such work.

Offshoring, civil liberties, and activism

  • Some argue offshore and crypto-like tools are also vital for activists in authoritarian or repressive contexts, where domestic accounts can be frozen.
  • Example given of a UK political group having its bank account frozen; advice: keep funds offshore in a different jurisdiction.

Conceptual/terminology debates

  • “Elites” criticized as a mislabel for billionaires/oligarchs; some reserve the term for the educated professional “top 20%” who work for the ultra-rich.
  • “Civic participation” discussed via the Singapore example: low protest activity and constrained civil liberties despite low corruption.

My experience with Claude Code after two weeks of adventures

Enthusiastic Views and Use Cases

  • Several commenters say Claude Code (CC) feels like “changing jobs” or working with a capable junior/mid engineer: great at boilerplate, CRUD, refactors, debugging, and greenfield apps in mainstream stacks (TS/Next.js, Laravel, Ruby, Python, Go).
  • Reported wins include: Airflow DAGs, billing dashboards, cloud cost analyzers, DB rebuild tools, mobile apps with nontrivial features, and data‑migration pipelines, sometimes in days instead of weeks.
  • People like using CC to overcome “white page” inertia, explore unfamiliar tech, and offload tedious work while keeping interesting problem‑solving for themselves.

Negative Experiences and Limitations

  • Others find CC slow, fragile, or unusable on complex, legacy, or niche codebases (C/C++ systems, Win32, Haskell/Bazel/Flutter, Godot, etc.).
  • Common failure modes: incomprehensible or redundant code, runaway edits, “grep hell”, invented APIs, non-durable designs, and gaslighting-like insistence on nonexistent functions.
  • Visual/UI tasks and subtle design details (e.g., a wrong-looking SVG icon) often slip past its “self‑verification”.

Best Practices and Workflow Patterns

  • Strong consensus that success depends heavily on:
    • Good tests and fast feedback loops; otherwise it may “fix” tests instead of code.
    • Structured codebases with clear conventions.
    • Per-folder CLAUDE.md / repomaps, explicit architecture and business rules.
    • Using plan mode, small tasks, explicit constraints, and having it keep a journal of issues/decisions.
  • Many treat it as an intern: you must review every change “like a hawk” and be ready to discard bad runs.

Comparison with Other Tools

  • Cursor: praised for in‑editor UX and tab completion; some find Cursor+Claude better than CC alone, others say CC’s agentic behavior and tool use are clearly superior.
  • Other tools mentioned: Cline, Aider, opencode, RepoPrompt, Kimi‑K2, Gemini 2.5, GPT‑4.1/o3, Grok 4; quality varies by model and task.

Impact on Developers and Jobs

  • Strong debate over whether CC makes juniors unnecessary and eventually lets a few seniors oversee AI‑generated code for many.
  • Pushback that CC doesn’t learn, repeatedly makes the same mistakes, and that mentorship is an investment in future seniors.
  • Some seniors report real speedups (1.5–10× by their own metrics); others cite studies or experience where perceived gains hide slower real progress.

Costs, Limits, and Hype

  • Rate limits on Pro vs Max tiers frustrate heavy users; some route through Bedrock or APIs.
  • Worry that current prices are “Uber‑era subsidies” and may spike later.
  • Many complain that rave posts lack concrete details (stack, size, tests, diffs), fueling suspicion of influencer hype and calling for more rigorous, specific case studies.

Apple Intelligence Foundation Language Models Tech Report 2025

Contributor List & Anti‑Poaching Speculation

  • Some suspect the long, randomly ordered contributor list is meant to blunt poaching.
  • Others note large projects often use flat or alphabetical ordering, and with many non‑Latin names, English alphabet order is arbitrary anyway.
  • People point out it’s easy to find core researchers via references or LinkedIn, so any anti‑poaching effect is weak at best.

On‑Device 3B Model & Hardware

  • The ~3B-parameter model already runs on current iPhones, iPads, Macs and visionOS in the 26 betas, accessible via the Foundation Models framework and even Shortcuts.
  • Users report decent latency (few seconds) and that it can act as an offline chatbot despite being tuned for summarization, extraction, and writing tools.
  • Discussion of ANE vs GPU: GPU tends to be faster but less efficient; ANE is optimized for low power. Fitting 7B+ models in 8 GB RAM is seen as technically impressive but impractical for real use.

Siri & Apple Intelligence User Experience

  • Many comments complain that Siri still fails simple or multi-part requests, especially compared with ChatGPT or Gemini, and call Siri “a joke.”
  • Some note current Apple Intelligence models already power features like notification summaries and writing tools, but users often find these underwhelming or intrusive.
  • Multiple insiders-style comments describe attempts to bolt an LLM onto legacy Siri as an integration nightmare (multi‑turn flows, “smart” UI, privacy, backwards compatibility), arguing a full reset is needed.

Apple’s AI Strategy, Privacy, and Data Centers

  • Strong split in perception:
    • One side sees Apple as badly behind frontier models, over‑promising with “Apple Intelligence,” and now partly retreating to OpenAI/Anthropic APIs and legal/PR positioning (e.g., “responsibly sourced” training data, Private Cloud Compute).
    • The other side argues Apple should not chase frontier models; their differentiator is privacy, on‑device inference, and tight OS integration, leaving generic chatbots to apps.
  • Private Cloud Compute is praised as a technically strong privacy design, but skeptics doubt scalability and note that truly powerful, tool‑using agents will likely require large cloud models with lots of tokens and user context.

Training Data, Applebot, and Robots.txt

  • Apple claims not to use user private data, to filter PII/profanity, and to honor robots.txt for Applebot.
  • Critics argue Apple scraped before clearly documenting AI training use, making later “you can opt out” messaging feel disingenuous.
  • Others respond that robots.txt has long been the standard opt‑out mechanism and that expectations for advance crawler disclosure are new and mostly born of anti‑LLM sentiment.
  • Some propose adding crawler “categories” (e.g., search vs. LLM‑training) to robots.txt to give publishers finer control.

Alt Text, Accessibility & Training

  • The paper’s use of image–alt‑text pairs sparks debate: alt text is heavily advocated for accessibility, yet is now prime supervised data for vision‑language models.
  • Some see this as “free annotation labor” for AI; others argue it’s still morally consistent to write alt text for disabled users even if it’s scraped.
  • A few note they already use LLMs to draft alt text but still review edits manually.

Developer Experience & Structured Output

  • iOS developers are enthusiastic about the Foundation Models framework: typed Swift structs, guided/structured output, streaming partial fields, and a clean bridge to external models.
  • Commenters compare this to “structured output” / grammar‑based sampling already available elsewhere, noting that forcing strict structure can reduce model quality and sometimes needs a two‑pass “think then structure” approach.

Model Updates & LoRA Adapters

  • People wonder how often on‑device models will change; the base weights are gigabytes, so frequent silent updates seem unlikely.
  • Apple appears to rely on LoRA‑style adapters for specialization; these must be retrained per base‑model version, suggesting model changes will likely track OS point releases, not constant churn.

Is Apple Behind or on Its Own Track?

  • Critics frame Apple’s current AI as a “slow‑motion train wreck”: small models, weak Siri, lack of headline features vs. ChatGPT/Gemini, and RAM‑constrained hardware.
  • Defenders counter that:
    • Apple has always moved slowly and then integrated deeply;
    • They don’t need to win the model race, only own the device, UX, and privacy story;
    • Users who want frontier chatbots can easily install apps or use the OS‑level ChatGPT integration.
  • There’s broader pessimism about Apple’s product coherence post‑Jobs, contrasted with arguments that partnering (e.g., with OpenAI) fits a long history of Apple leveraging outside tech while keeping tight control of the platform.

All AI models might be the same

Architecture vs Data and Model Limits

  • Some argue architecture “doesn’t matter,” and convergent behavior mainly reflects shared data; others strongly disagree, likening this to saying algorithm choice is irrelevant.
  • Critics note current LLMs are largely feed‑forward with discrete training cycles, unlike brains’ continuous feedback and learning, and our limited understanding of memory and neural dynamics may hide better architectures.
  • Skeptics of Transformers claim true AGI will likely use very different mechanisms.

Shared Semantic Spaces and “One Model” Hypothesis

  • The “Mussolini or Bread” game is used to argue for a shared latent semantic space across people and models; many find this compelling but limited to overlapping knowledge and culture.
  • Several commenters point out logical flaws in the game’s reasoning (e.g., many non‑person concepts could still be “closer to Mussolini than bread,” non‑transitive relations).
  • Some see the effect as mostly due to similar training corpora and consistent estimators, not deep Platonic structure.

Diffusion Models, Compression, and Memorization

  • A highlighted paper shows optimal diffusion models act like patch mosaics from training data; using this directly would produce huge, unwieldy systems.
  • Others caution against taking the “patch mosaic” metaphor too literally: real models aren’t perfectly minimized, are overfit on small benchmarks, and succeed largely because imperfect training enables interpolation, correction, and decomposition tasks.
  • Debate continues on whether convergent representations can yield smaller models, or inherently push toward larger architectures approximating the data.

Language, Translation, and Non‑Human Communication

  • There’s extensive debate over whether shared embedding spaces could let us translate whale or lion communication without a “Rosetta stone.”
  • One side emphasizes shared physical world and experiences (hunger, sun, movement) and the possibility of mapping contexts and abstractions across species.
  • The other stresses that meaning is deeply tied to lived, species‑specific experience (Wittgenstein’s lion), that some concepts may be untranslatable or extremely lossy, and that current methods already struggle across human cultures.
  • Relatedly, people discuss universal grammar, animal symbolic ability (e.g., apes, dolphins, elephants), and projects like dolphin‑focused LLMs; views range from “humans just have special grammar hardware” to “other species lack only an effective, fitness‑linked naming system.”

LLM Capabilities, Hallucinations, and Domain Use

  • In practice, LLMs sometimes fail at simple semantic games (Mussolini/Bread) without heavy prompting.
  • A user report on a backup‑software assistant shows plausible but hallucinatory instructions; they conclude domain‑specific LLMs need strong fact‑checking and that good documentation plus search often remain superior.
  • Others note that different models often give remarkably similar answers, which they attribute to similar architectures and overlapping corpora.

Intelligence, Learning, and AGI Debates

  • Some commenters see LLMs as brute‑force reverse‑engineered human brains, matching input–output behavior over huge datasets.
  • Others insist LLMs “don’t think, learn, or exhibit intelligence,” comparing them to static books, pointing to lack of persistent self‑updating without retraining.
  • Opponents counter with analogies to neurons (simple units giving rise to emergent cognition) and argue that training + context + external memory already approximate forms of learning.
  • Dynamic tokenization and continual‑learning schemes are discussed as necessary steps toward more “alive” systems.
  • There is disagreement over whether LLMs can ever yield AGI, with some viewing Transformers as a dead end and others treating them as early approximations of a single underlying “intelligence.”

Ethics, Alignment, and Platonic Forms

  • Some tie the convergence of concepts (“dog,” “house,” etc.) to Plato’s Forms and speculate about a learnable “Form of the Good” that could aid alignment.
  • Others note that moral notions (abortion, animal testing, etc.) are highly contested even within one culture, so a universal “Good” embedding is dubious.
  • A few liken this to Jungian archetypes or to deep, overloaded words in natural language.

Implications for Open Source and Future Work

  • If all large models converge on essentially the same representation of the world, one strong open‑source model might eventually substitute for proprietary ones.
  • Suggested empirical tests include training small models on very different corpora (e.g., disjoint historical traditions) to see whether their embeddings can be mapped into a common space.

ChatGPT agent: bridging research and action

Real‑world actions and liability

  • Many see the “order 500 stickers” demo as a milestone: text -> physical goods (“voice to stuff”).
  • Others note similar pipelines (voice → code → 3D print) have existed for years; this is more polish than a conceptual first.
  • Concerns raised about mis-orders (e.g., 500k stickers): who pays? Discussion touches on indemnity clauses in OpenAI’s ToS and the practical backstop of credit card limits and merchant checks.

What is an “agent”? Competing definitions

  • Several conflicting definitions circulate:
    • Technical: “models using tools in a loop” or “tools + memory + iterative calls.”
    • OpenAI-style: “you give it a task and it independently does work.”
    • Older meaning: scripted workflows, where the human designs the steps.
  • Some argue real “agency” for users will be when systems can negotiate messy real-world tasks (refunds, cancellations, appointments, disputes), not just run tools.

Usefulness in everyday life

  • Mixed views on personal utility:
    • Optimists imagine agents booking dinners, babysitters, travel, doing price monitoring, building spreadsheets, etc.
    • Skeptics highlight trust, integration, and nuance: knowing partner preferences, personal history, social dynamics. Hard parts of life are about values, not operations.
  • Some argue the best UX is not fully autonomous “one‑shot” tasks but an ongoing assistant that asks brief questions and executes, like a good human PA.

Error rates and the “last 2%”

  • The spreadsheet example (98% correct, 2% manual fix) triggers a long debate:
    • Critics say finding subtle 2% errors can take as long as doing the whole task, and compounding errors across multi‑step agent workflows can make results unusable.
    • Others note humans also make frequent mistakes; verification can be cheaper than doing work from scratch, and for many business tasks 95–98% accuracy is economically acceptable.
    • There’s broad agreement that LLM outputs must be treated like work from a very junior hire: useful, but requiring careful review and good tests.

Security, privacy, and prompt injection

  • Strong worry about giving agents access to email, calendars, payments, and logins.
  • Prompt injection via web pages, calendar invites, or emails could exfiltrate sensitive data or trigger harmful actions.
  • OpenAI’s promise of “explicit confirmation” for consequential actions is questioned: how reliably can an LLM detect what’s consequential?
  • Some foresee a new wave of agent-targeted attacks and blackmail‑style scenarios.

Web integration, blocking, and on‑device agents

  • Past OpenAI “operator” bots are reportedly being blocked by major sites (e.g., job boards, marketplaces), undermining shopping and job‑application use cases.
  • People expect a shift toward agents running through the user’s own browser, IP, and cookies (extensions, local “computer use”) to evade datacenter blocking and robots.txt.
  • This raises separate risks of account bans and makes agent–website business relationships (or profit‑sharing) a possible future battleground.

Hype, limitations, and comparisons

  • Some see this as the first “real” agent product from a major lab; others note that similar agentic CLIs and desktops (Claude Code, Gemini, etc.) already exist and can be re‑created with a model, a loop, and a few tools.
  • There’s a recurring “last‑mile” concern: going from 90–95% to 99% reliability on open‑ended tasks may be extremely hard, and many demos appear curated along happy paths.
  • Debate continues on whether current LLM progress is hitting diminishing returns versus still on a steep, transformative trajectory.

Impact on work and jobs

  • Some think this mostly accelerates workers (fewer hours, same headcount); others expect AI‑first firms to outcompete “paper‑shuffling” incumbents, killing many white‑collar roles.
  • Several commenters expect bosses to use any time saved to demand more output, not more leisure, and worry about growing technical debt and low‑quality outputs as agents proliferate.

Ask HN: What are your current programming pet peeves?

Documentation, Search, and Learning Curve

  • Strong frustration that in 2025 it’s still hard to find clear docs for basic features; search is dominated by SEO farms, AI sludge, bootcamp blogspam, and outdated answers.
  • Python splits opinion: some say the built-in help() and official docs are excellent once you learn their structure; others find them dense, spec-like “walls of text” and want concise quick-reference tables instead.
  • PEPs are praised as good references but criticized as serving as de‑facto user docs.
  • Similar issues in other ecosystems (Ruby, Free Pascal/Lazarus): official docs often only list arguments without explaining meaning, use, or design.
  • W3Schools/GeeksForGeeks are called out as high-ranking but shallow and incomplete; MDN-style docs are considered far more informative.
  • Frequent complaint about libraries lacking READMEs that explain audience, context, and realistic examples.
  • Several people wish API docs explained high-level design and rationale, not just tautological method descriptions.

Shells, Terminals, and Tooling

  • One thread laments that shells still look the same and asks for a Python-like shell experience.
  • Others defend traditional shells (bash/zsh/fish) as better suited for process orchestration and pipelines than Python; Python as a login shell is described as clumsy.
  • Alternatives like xonsh and nushell are suggested as “Python-ish” or more structured shells.
  • Debate over whether using a shell counts as “programming.”

Churn, Deprecation, and Ecosystem Instability

  • Major peeve: libraries/frameworks constantly deprecating APIs; even projects untouched for months feel “ancient,” especially in NPM/Node ecosystems.
  • Old browser JS still works, but modern build chains (Gatsby/Node) often won’t build without painful upgrades.
  • Complaints about moving targets in both tech and product requirements; developers feel forced to chase shifting APIs and specs.
  • Monorepos are said to complicate documentation and deployment, requiring extra tooling layers.
  • Some want drastic simplification (e.g., dropping legacy x86 modes).

Types, Config, and Abstraction Choices

  • Static typing sparks strong disagreement:
    • One side finds complex type systems overbearing and misaligned with how they think and work.
    • Others say most bugs in dynamic-language codebases are effectively type errors and prefer static or gradual typing.
  • Preference expressed for inferred, non-ornate type systems (Hindley–Milner-like) over “do everything in the type system” designs.
  • Frustration with poor domain modeling: objects full of optional fields, or logic expressed via untyped hash maps/dicts and YAML configs.
  • YAML DSLs in particular are criticized as fragile, hard to tool, and harder for LLMs to handle compared to real scripting languages.

AI, Code Generation, and Developer Experience

  • Mixed views on LLMs:
    • They’re seen as useful for navigating bad documentation, dependency hell, and obscure errors.
    • But AI-generated code from colleagues is described as overengineered, subtly buggy, and poorly understood by its “authors.”
  • Some argue AI should focus more on reviewing human code than generating it.
  • Worry that improving public docs just feeds training data; predictions of more paywalled/“paid web” content.
  • Annoyance at AI tools that don’t auto-retry failed requests and that lag behind current language versions.

Product Websites, Process, and Social Frictions

  • Strong irritation with developer-product sites aimed at executives: buzzwords instead of clearly stating “what it does” and “what it costs.”
  • Dislike of being forced to provide contact info to get technical details, which signals an unpleasant sales funnel.
  • Complaints about bug reporting: project-specific trackers, rigid templates, and stale bots closing issues. Idea floated for an independent, cross-product bug tracker curated by users.
  • Social/process peeves dominate for some: moving requirements, pressure to estimate and then compress timelines, and being responsible for both defining and building evolving products.

Miscellaneous Pet Peeves

  • Odd indentation widths; unpleasant large legacy C codebases with nonstandard styles and heavy macro/makefile usage.
  • Git is seen by some as overcomplicated and overdue for a more intuitive replacement; alternatives like fossil and jj are mentioned positively.
  • Node/JS ecosystem churn, function coloring, JS vs TS debates (with some asserting TS + LLMs is now the clear choice for serious work).
  • Complaints about debugging multiprocessing, opaque build systems (e.g., CMake), null-terminated strings, ambient-privilege OSes, fragile web-app build pipelines, AI service UX (e.g., Gemini quirks), and even specific syntax like elseif.

New colors without shooting lasers into your eyes

Effect on different kinds of color vision

  • Several red–green colorblind and deuteranomalous commenters report that the illusion works, often producing a vivid blue‑green or lighter green halo, though likely different from “normal” trichromats.
  • Others with color weakness see only a pale halo or no “new” color at all; one suggests heavy cone overlap might blunt the effect.
  • Some speculate it should work for anomalous trichromats but likely fails for true dichromats (protanopia/deuteranopia). One suggests it might even be used as a diagnostic test, though that’s unproven.
  • There is disagreement over some physiological claims in the article about how deuteranomaly works.

How the illusion works and how to view it

  • Many clarifications: the black bar is just a countdown; the effect comes from saturating cones via prolonged fixation on the central dot and then watching the shrinking circle/afterimage.
  • Moving eyes or blinking reduces the effect; some users “refresh” the countdown or change viewing distance to intensify it.
  • Some see nothing, question whether it truly goes “outside” the natural color gamut, or argue it’s just a standard negative afterimage, not a fundamentally new color.

Subjective experiences and comparisons

  • Many report a striking, “magical,” ultra-saturated teal/green/cyan halo, sometimes grainy or “TV snow”-like, sometimes reminiscent of psychedelic “ultragreen” experiences.
  • Others see different hues (yellowish, orange, purple) depending on custom color combinations, lighting, display settings, or even contact lenses.
  • Comparisons are made to Shepard tones (endless pitch illusion), op art, James Turrell installations, and high-contrast typography causing persistent afterimages.

Color vision biology and evolution

  • Extended discussion on overlapping cone sensitivities, gene duplication on the X chromosome, and differences between Old World and New World primate color vision.
  • Mention of rare human tetrachromacy, bird tetrachromacy, and the challenge of describing a four-dimensional color experience.

Plants, spectrum, and environment

  • Side thread on why plants are green, chlorophyll absorption bands, solar spectrum peaks, and hypotheses about energy variance, noise reduction, and historical evolutionary contingencies.
  • Another tangent: water’s transmission window defines “visible light,” plus how atmospheric filtering shapes both photosynthesis and visual systems.

Tell HN: Notion Desktop is monitoring your audio and network

Privacy concerns and monitoring behavior

  • Many commenters find the idea of a note-taking app watching microphone and “network ports” inherently creepy and disproportionate to its purpose.
  • People worry about highly sensitive audio potentially being captured, habits inferred from microphone usage, and the possibility of data eventually being sold or repurposed (even if not today).
  • Several draw a distinction between:
    • Local-only detection of activity for UX, vs.
    • Exfiltrating raw data or detailed usage logs to servers.
  • Some argue that even local monitoring of meetings without explicit, up-front consent is unethical and erodes trust.

Notion’s explanation of the feature

  • Company employees explain:
    • Audio recording only occurs when users explicitly start “AI Meeting Notes”.
    • For desktop “meeting detection” notifications, the app detects that some process is using the microphone and matches process names (Zoom, etc.); it does not inspect audio.
    • They state there is no network traffic monitoring or port analysis; earlier support wording was called a misunderstanding.
    • Users can disable “Desktop meeting detection notifications” in settings, but it ships enabled by default.
  • On macOS, microphone usage would be visible via OS indicators; commenters note they haven’t seen Notion active there except when explicitly recording.

Opt-in vs opt-out and trust

  • The biggest flashpoint is that meeting detection is opt-out:
    • Employee admits PMs avoid opt-in because features then see very low usage.
    • Many argue this is precisely why privacy-sensitive features must be opt-in, with clear, contextual explanation.
  • Some suggest better patterns: first-launch privacy screens, inline “this feature needs X, enable?” prompts, or at least a one-time “turn this off” button.
  • A few defend the implementation as a common, benign pattern (similar to other “meeting-aware” apps), saying the outrage is disproportionate.

Platform, sandboxing, and OS behavior

  • Multiple commenters say they avoid native/Electron wrappers altogether, preferring browser use (especially on Linux) to reduce permission surface.
  • There’s discussion of macOS’s “Local Network” permission:
    • Local-network prompts are often triggered by Electron/Chromium defaults (mDNS, WebRTC) and do not equate to packet sniffing.
    • True packet inspection still requires more privileged APIs.
  • Others note that non-sandboxed desktop apps can still inspect open file descriptors and see which processes use the mic or sockets.

Alternatives and product tradeoffs

  • Several users love Notion’s information architecture and collaboration but hate its performance and now question its privacy posture.
  • Numerous alternatives are discussed (Obsidian-based setups, Anytype, Loop, Affine, NocoDB, Docmost, XWiki/Nextcloud/wiki.js, etc.), with tradeoffs around UX, database features, multiplayer, and self-hosting.
  • Some advocate contributing to open-source tools instead of building yet another proprietary system, and call for Notion to offer end-to-end encryption or open-source clients to rebuild trust.

Show HN: Conductor, a Mac app that lets you run a bunch of Claude Codes at once

Purpose and Workflow Model

  • Conductor lets users run multiple Claude Code agents in parallel by creating isolated git worktrees per session.
  • Supporters say this solves conflicts when multiple agents edit the same repo and makes better use of “agent lag” time by letting them switch between tasks.
  • Skeptics argue they can already do this with multiple terminals, tmux, and manual worktrees; for them, Conductor “gets in the way” compared to keyboard-only workflows.

Git, GitHub, and Local Repos

  • Initial behavior was to clone repos from GitHub with broad OAuth permissions and no obvious privacy policy, which drew strong criticism (full read-write on all repos, org settings, deploy keys).
  • The author responds that this was due to OAuth limitations and says they are moving to a GitHub App with fine-grained permissions and also supporting local GitHub CLI auth.
  • Several users want purely local git support with existing checkouts, no mandatory GitHub integration, and no repeated dependency installs; setup scripts (copying node_modules, env files, running installs) are suggested but seen as obscure.
  • Worktrees cause friction for untracked files (env files, submodules) and PR workflows; some consider the extra complexity not worth it for most AI coding tasks.

Features, UX, and “Feel”

  • Requests include:
    • Changing the default branch for new workspaces.
    • Custom “Open in…” commands (e.g., SourceTree).
    • Embedded terminal, “plan mode,” message queuing, and multi-repo workflows.
  • Some users feel Conductor loses the “feel” of native Claude Code: streaming output, escape-to-interrupt behavior, and tight terminal interaction.

Alternatives and Ecosystem

  • Multiple alternatives are mentioned: Crystal, Claude Squad, Plandex, container-use, par, vibe-tree, hydra, autowt, etc., many of them open source and focused on simple worktree or container management.
  • Some prefer minimal tools that “do one thing well” and work against existing local repos.

Broader Reflections

  • There’s debate about whether parallel agents truly increase productivity versus the human review bottleneck.
  • Concerns are raised about AI coding tools and data privacy in general; others respond that enterprise contracts, on-prem/cloud provider hosting, and not storing secrets locally mitigate risk.

How I Use Kagi

Yandex Partnership & Geopolitics

  • Major thread focus is Kagi’s continued use of Yandex as a data source and paying them for API access.
  • Some users cancelled subscriptions or won’t subscribe because they do not want any of their money flowing into the Russian economy during the invasion of Ukraine; Yandex is described as deeply entangled with the Russian state and propaganda.
  • Others argue this is disproportionate or impractical given that many Western companies and governments also fund or enable wars; accusations of “whataboutism” fly in both directions.
  • A recurring split:
    • One camp prioritizes moral boycotts, even if inconsistent or small in impact (“I draw a line somewhere”).
    • Another camp stresses moral complexity, inevitable complicity, and says a good, small, privacy‑respecting search engine may do more net good than harm.
  • Kagi’s founder defends the Yandex integration as ~2% of costs and justified purely on search quality, not politics; argues that once geopolitics drives inclusion/exclusion of sources, it stops being a neutral search engine.
  • Some users find this principled and “refreshing”; others see it as technocratic, evasive, or outright disqualifying, and request at least a per‑user “no Yandex” toggle.

Search Quality, Features & Comparisons

  • Many subscribers report Kagi consistently matching or beating Google on difficult, specific, or technical queries, especially when Google’s “fuzzy intent” and SEO spam get in the way.
  • Kagi’s killer features for fans:
    • Per‑user block/boost of domains and “lenses” (custom source filters like “Academic”), which can also constrain its AI assistant.
    • Ability to globally nuke sites like Pinterest, tabloids, or obvious AI‑spam; some say going back to Google + uBlock feels awful after getting used to this.
  • Some users see little or no difference versus Google or DDG once ad‑blocking is enabled, and don’t find a subscription justified.
  • Performance is occasionally cited as slower than Google; Kagi staff engage and offer to investigate.

AI, LLMs & Changing Search Habits

  • Several users now lean on LLMs (ChatGPT, Claude, Perplexity) for many queries and feel that “search is being squeezed out.”
  • Others highlight Kagi’s integrated AI features (? quick answer, !ai to send to Kagi Assistant with multiple models) as a strong hybrid approach.
  • Some cancel Kagi because their primary “search” is now an LLM; others subscribe to both Kagi and AI tools and see Kagi’s higher‑quality web retrieval as a good substrate for AI.

Blocklists, Content & UX Gaps

  • The article’s shared blocklist is controversial: some see it as over‑broad (blocking large news sites, social networks, Amazon, etc.), mixing “SEO spam” with sites the author personally dislikes.
  • Users advise treating such lists as starting points, not gospel, and emphasize downranking over hard blocks.
  • Kagi’s weak spots repeatedly mentioned: maps and local business/restaurant search, where Google Maps is still preferred.
  • Other frictions: required sign‑in for any search, lack of easy prepaid/anonymous options beyond a constrained Privacy Pass, and Yandex being a hard blocker for some otherwise enthusiastic users.

Mistral Releases Deep Research, Voice, Projects in Le Chat

Model Release Fatigue & How People Cope

  • Many describe “model release fatigue”: constant switching between Claude, GPT, Gemini, Llama, Mistral, etc. creates context overload with only marginal real-world benefit.
  • Coping strategies mentioned:
    • Pick 1–2 vendors and stick with them unless a big shift happens.
    • Use AI mainly on “fringe” tasks (Excel, scripts, glue work) while keeping core workflows mostly traditional until the field stabilizes.
    • Accept that chasing “the best” all the time is unsustainable and often distracts from doing actual work.

Local vs Hosted Models & Hardware

  • Several commenters happily run local models (Qwen, Whisper, etc.) via Ollama/LM Studio for coding and experimentation.
  • Others argue local GPUs are economically unjustified if the model runs only a small fraction of the time, suggesting shared/collective infra.
  • Debate on whether consumer hardware (VRAM) will evolve fast enough for large models to be “mid-tier local” this decade.

Competition, Innovation, and ‘Copying OpenAI’

  • Some claim the entire industry just clones OpenAI’s product set (chat, voice, deep research). Others counter that:
    • Labs continually copy and leapfrog each other (e.g., agentic protocols, world models, novel attention mechanisms).
    • From the outside everything looks like f(string) -> string, but training data, tools, and UX differ in practice.

Deep Research Features

  • Mixed views on “deep research” tools:
    • Several praise OpenAI’s for market and engineering tradeoff studies, calling it like having a junior researcher.
    • Others say OpenAI’s is actually worst in their tests; Anthropic, Kimi, Perplexity and others do better on their queries.
    • Common complaint: all vendors produce verbose, “AI-slop” style reports when users often want concise comparisons.

Voice, Speech, and Image

  • Mistral’s Voxtral STT is seen as a strong entrant but critics note the marketing didn’t compare against all top open ASR models.
  • Confusion/disappointment that Mistral’s “Voice mode” seems to be dictation, not a full real-time conversational voice agent.
  • Image editing via Le Chat (likely Flux Kontext under the hood) impresses many: precise localized edits, good preservation of the rest of the image; main drawbacks are resolution and small artifacts (e.g., book titles, shadows).

EU Angle & Vendor Lock-In

  • Some celebrate Mistral as evidence the EU is “waking up” and plan to switch from US providers; others point out Mistral’s US investment and infra ties, questioning how “European” it truly is.
  • A few say geopolitical/ethical concerns about data and regulation will matter more over time, with interest in credibly open, well-sourced datasets.

Productivity, Jobs, and Coding with AI

  • Split sentiment on coding productivity:
    • Many report clear speedups for boilerplate, script writing, and navigating huge APIs; non-users risk lagging behind over time.
    • Others argue aggregate productivity gains remain unproven and that “vibe-coded” AI-heavy codebases may create long-term maintenance nightmares.
  • Some advocate deliberately not using LLMs to preserve the joy of programming; others note that for employees, ignoring productivity tools can be career-risky.

UX, Pricing, and Practical Impressions

  • Several say Mistral’s Le Chat UX is among the best: fast responses, stable UI, projects/libraries, optional web search.
  • Users like the growing competition: frequent promos keep premium models cheap, and meta-routers (litellm, OpenRouter) make model hopping easier.
  • A recurring theme: if you just picked a solid model last year and stuck with it, you probably didn’t miss much—except for a few standout releases (e.g., specialized reasoning or coding models).