Hacker News, Distilled

AI powered summaries for selected HN discussions.

Page 346 of 535

ChatGPT Is a Gimmick

What “gimmick” means in this debate

  • Many interpret “gimmick” as tech that looks impressive but doesn’t fundamentally change important work; others say that bar is too high given how much value they personally get.
  • Some argue the author really means: AI cannot replace the human, effortful process of learning and thinking, especially in education, not that it has zero utility.

Education, cheating, and learning

  • Several commenters agree that students using LLMs to bypass learning is a real problem, analogous to older forms of cheating but cheaper and easier.
  • Others ask for empirical evidence that cheating has increased, not just shifted form.
  • Strong concern that edu hype frames AI as making learning “effortless,” which is seen as dishonest and infantilizing.
  • Pushback: AI can be a powerful tutor, explainer, and “rubber duck,” especially for those without access to good teachers; the right question is how to teach students to use it well.

Practical usefulness vs. frustration

  • Heavy users describe clear wins:
    • Drafting and tightening emails, docs, and reports.
    • Explaining math/physics proofs, checking derivations, and catching reasoning mistakes.
    • Coding scaffolds, refactors, one-off scripts, and unfamiliar stack “on-ramping.”
    • Advanced thesaurus, language learning aids, data cleanup, and search replacement where web search is SEO-spammed.
  • Others say they “can’t get anything valuable”: hallucinated APIs, kernels, flags, math errors, overconfident nonsense, and brittle “agentic” loops. They see usefulness only when answers are easy to verify.
  • A recurring heuristic: LLMs help when it’s slow to produce a plausible answer but fast to check it; they’re poor when verification is hard.

Reliability, hallucinations, and epistemics

  • Multiple comments stress hallucinations, fabricated citations, and misrepresented sources as dangerous, especially where domain knowledge is weak.
  • Some suggest treating LLMs as “compressed Google” or recommendation engines: good at pointing to ideas, not as authorities.
  • Others complain about the obsequious style and tendency to reinforce user assumptions instead of challenging them.

Hype, markets, and capitalism

  • Many see a gap between corporate/VC “near-AGI” hype and the modest, fragile reality of current tools.
  • Some worry AI is pushed to cut labor costs and concentrate wealth, using unpaid human-created training data, while displacing illustrators/writers.
  • Others argue markets will eventually sort out what genuinely works, but that hype-driven overinvestment and “enshittification” are already obvious.

Use in creative/ordinary life

  • Examples include cooking with limited ingredients, generating language-learning imagery, creative brainstorming, and casual explanation of physics or philosophy.
  • Critics call many of these uses “gimmicky” or niche; proponents counter that small, messy boosts across many tasks are still transformative.

Kotlin-Lsp: Kotlin Language Server and Plugin for Visual Studio Code

Motives, Timing, and Kotlin Adoption

  • Many see this as JetBrains finally yielding after years of resisting an official LSP to protect IntelliJ’s advantage.
  • Suggested triggers: perceived stagnation or slight decline in Kotlin usage, Kotlin’s image as “Android-only,” VSCode’s massive share (especially with AI/LLM tools), and developers unwilling to adopt ecosystem-locked languages.
  • Some argue prioritizing IDE lock‑in over Kotlin’s ubiquity was shortsighted; others say JetBrains must optimize for IntelliJ revenue, not raw Kotlin usage.

Business Model and Ecosystem Debates

  • Debate over whether growing Kotlin usage benefits JetBrains: critics say it creates cost without direct revenue; defenders cite brand, mindshare, education, and long‑term conversion to paid tools.
  • Comparisons: Apple (closed but wildly successful) vs. Borland/Delphi and sponsors that abandoned NetBeans/Eclipse.
  • Clarification that Kotlin and IntelliJ are overwhelmingly developed by JetBrains, not funded by large Google contracts.

Partial Open Source LSP and Architecture

  • Initial backlash to the “partially closed-source” status.
  • JetBrains explanation: current LSP implementation leans heavily on internal IntelliJ/Fleet/Bazel infrastructure; goal is to iterate quickly, then decouple and fully open it.
  • Some see this as pragmatic early release (“better some OSS now, all later”) rather than a red flag.

Editors, IDEs, and Lock-In

  • Strong split between “best specialized IDE per language” (IntelliJ/Android Studio/Xcode, etc.) and “one editor for everything” (VSCode, Helix, Zed, Emacs).
  • VSCode praised for multi-language workflows and LSP support in a single instance; criticized for fragile extensions and weaker debugging/profiling than JetBrains.
  • JetBrains criticized for product fragmentation (different IDEs per language) and brittle project configs; others still prefer its deep tooling.
  • Emacs users are particularly excited—Kotlin support via LSP unblocks them.

Kotlin vs Java and Language Lock-In

  • Java 21/23 raises the question: why Kotlin? Supporters cite null safety, mutability semantics, terseness, better functional features, top-level functions, unsigned types, and iterator ergonomics.
  • Several correct the misconception that Kotlin is proprietary; it’s Apache 2.0 licensed.
  • Broader debate on ecosystem lock‑in: some avoid Kotlin/C#/Apple stacks entirely; others point out every language has an ecosystem, and C#/.NET and Kotlin are now broadly cross‑platform.

Limitations and Concerns About This LSP

  • Current LSP only supports JVM‑only Gradle projects; lack of Kotlin Multiplatform support is a blocker for some and is explicitly awaited.
  • Worry that JetBrains might abandon this like the Kotlin Eclipse plugin; others note that plugin was maintained for years, but had a small audience.
  • Isolated reports of compatibility issues with Cursor’s older VSCode base.

Getting a paper accepted

Science vs. “Research Game” and Careerism

  • Many commenters distinguish between “doing science” (discovery, understanding) and “playing the game” (papers, grants, prestige).
  • Some see the article’s branding/marketing advice as further evidence academia prioritizes careers over discovery, especially in U.S. PhD cultures.
  • Others argue this has always been true to some degree; science has patrons and politics, yet still progresses.
  • A nuanced view: career optimization has produced both major advances and a lot of low‑quality, hype‑driven work.

Communication, Branding, and Clarity

  • Broad agreement that clearer writing, better figures, and shifting from “what we did” to “why it matters” genuinely improves papers and helps readers.
  • Several worry that in ML, “branding” (catchy names, punchy titles, flashy graphics) has become central, blurring good communication with salesmanship.
  • Some say similar branding exists in other fields (e.g., characteristic figure styles in chemistry) and does affect recognition.
  • Debate over whether peer review rewards truth or mainly rewards plausible, clearly packaged claims; code and artifacts often escape serious scrutiny.

Peer Review, Randomness, and Politics

  • Multiple experiences of acceptance/rejection feeling largely stochastic once a basic quality threshold is met.
  • “Author engineering”: having a well‑known coauthor or insider often helps “enter” a field; double‑blind review only partially mitigates this.
  • Reviewers are overworked and uncompensated; many skim quickly, making clarity and first‑page impressions disproportionately important.
  • Some conclude peer review is primarily a social filter; others defend it as an imperfect but still “strong” filter against obvious errors and fraud.

Ideas, Experiments, and What Counts as Science

  • One line of argument: ideas are cheap; real science is rigorous experimentation/theory plus faithful reporting, including failures.
  • Counterpoint: high‑impact, genuinely surprising ideas are rare; “obvious next step” ideas plus solid execution still move fields.
  • Several lament that negative or non‑working results are rarely published, causing duplicated effort and slower progress.

Metagame Advice and Career Strategy

  • Commenters extend the article’s theme: optimize time for visible outputs (papers, grants, high‑impact collaborations, conferences).
  • Engineering that “makes things actually work,” cluster tooling, teaching improvements, and careful packaging are often invisible to academic gatekeepers.
  • Some suggest these “invisible” skills are better rewarded in industry or entrepreneurship, where working systems can be directly monetized.

Gemini Diffusion

Performance, hardware, and speed

  • Commenters are struck by the demo speed; some compare it to Groq/Cerebras and wonder how well diffusion will map to SRAM-heavy or local hardware.
  • Several note that diffusion likely uses more total compute than comparable autoregressive (AR) models but does it in far fewer, parallelizable steps, trading wall‑clock time for throughput.
  • Concern that this parallelism may saturate accelerators quickly and reduce batching efficiency for cloud providers, while being a big win for self‑hosted/local inference.

Coding assistance and large codebases

  • Mixed experiences: some say LLMs excel at greenfield code and small refactors; others report “steaming pile of broken code” even for simple CRUD tasks across multiple frontier models.
  • A recurring pain point is refactoring or re‑architecting ~1k+ LOC files or multi‑file patterns; users report hallucinated APIs, broken layering, and missed edits unless heavily constrained and iterated.
  • Others counter that careful workflows (spec docs, implementation plans, step‑wise patches, tools like Aider/Cline/Continue/Cursor) can make LLMs very effective, but they require significant “prompting and glue”.

Institutional knowledge and “negative space”

  • One thread emphasizes that models can’t see what is absent from a codebase—architecture choices, rejected libraries, etc.—and this “negative space” carries critical design signal.
  • Some argue this should be documented (design docs, ADRs, comments about rejected options), but others note that in reality most codebases don’t capture this, and much tacit knowledge remains in developers’ heads.
  • Ideas include mining git history, Jira tickets, issue trackers, and meeting notes to approximate institutional memory, though several worry about noise and scale.

Diffusion vs autoregression: mechanisms and trade‑offs

  • Multiple comments clarify that diffusion here replaces AR, not transformers: these are likely encoder‑style transformers trained with heavy masking/BERT‑like objectives.
  • High‑level explanation: start from heavily masked or noisy sequences; the model repeatedly “denoises” them, progressively refining all tokens in parallel. Unlike AR, earlier tokens can be edited later.
  • Claimed benefits: faster generation, less early‑token bias, potential for better reasoning/planning per parameter, and the ability to revise intermediate text.
  • Skeptics question whether diffusion can match AR on “output quality per compute”, especially for strictly sequential causal data (code, math, time series), and note a lack of detailed public training/inference specs yet.

Safety, determinism, and alignment

  • Early access reports mention strong prompt‑injection vulnerabilities (e.g., safety refusal bypassed by roleplay framing), interpreted by some as under‑baked alignment on this experimental model.
  • Others emphasize that diffusion LLMs can still be deterministic given fixed seeds and controlled hardware, with the usual floating‑point caveats.

Community reaction and meta‑discussion

  • Many see Gemini Diffusion as one of the most important I/O announcements, especially for code generation; others note similar prior work (e.g., Mercury) and frame this as mainstream validation rather than novelty.
  • Several defend the blog post as adding value over the official DeepMind page (demo video, cross‑vendor comparisons, curated explanations), pushing back on claims it’s “blog spam”.

I have tinnitus. I don't recommend it

Personal experiences & impact

  • Many commenters have tinnitus, from mild “sound of silence” only noticeable in quiet, to constant, loud ringing or pulsatile (heartbeat-synced) noise.
  • Onset varies: since early childhood, after a single loud event, gradually over years, or suddenly (e.g., after illness, meds, or vaccines).
  • Impact ranges from “barely notice it unless reminded” to severe anxiety, sleep loss, difficulty with conversation (especially calls and noisy rooms), and, in extreme cases cited, suicidality.
  • A recurring theme: the brain can learn to filter it out over months/years; distress is often highest near onset and decreases with habituation.

Causes and triggers discussed

  • Loud sound is the dominant cause: concerts, clubs, bands, PA systems, gunfire, machinery, sirens, and even a screaming laptop drive.
  • Military service (artillery, helicopters, defective earplugs) is frequently mentioned.
  • Other triggers: childhood ear infections, eardrum rupture, snorkeling/barotrauma, TMJ/TMD, neck posture, jaw clenching, stress, high blood pressure, COVID infection, and some drugs (antibiotics, valproate, chemo, Zyban/Wellbutrin).
  • Some report onset soon after COVID vaccination; others associate theirs more with WFH headphone use. Causality is acknowledged as unclear.

Coping strategies & therapies

  • Common tools: white/brown noise (fans, apps, myNoise), meditation, distraction, and avoiding silence.
  • Several describe cognitive-behavioral / “tinnitus meditation” approaches: deliberately focusing on the sound to change the brain’s threat response, then learning to shift attention away. Some report major benefit; others online reportedly fear it might worsen things.
  • EMDR and trauma-focused therapy are linked in papers connecting tinnitus with PTSD, complex trauma, and adverse childhood experiences.
  • Physical approaches: neck and jaw work (stretching, massage guns, posture changes), night guards for bruxism/TMJ, and in a few cases upper-cervical chiropractic.

Devices and tech aids

  • Bimodal/bilateral stimulation devices (e.g., Lenire, and an anticipated but delayed Susan Shore device) get mixed reports: “much quieter” for some, “made it worse” for others.
  • Hearing aids, especially high-frequency loss–targeted (e.g., Lyric), are reported to reliably reduce or eliminate tinnitus for some.
  • AirPods Pro and similar devices: personalized audiogram-based EQ and hearing-aid features can improve hearing clarity; not claimed as a tinnitus cure but sometimes change its character.

Prevention and sound environments

  • Strong consensus: wear earplugs at concerts, clubs, shooting ranges, even with lawn equipment and vacuum cleaners; carry plugs or use ANC as routine.
  • Several argue concert and venue volumes are unnecessarily, even dangerously, high and call for regulation.
  • Rule of thumb voiced: if sound is painful or leaves ears ringing, damage is occurring.

Research, psychology & side threads

  • Links shared to treatment trial trackers, support forums, EMDR/tinnitus studies, and a hyperacusis definition.
  • One thread debates whether tinnitus is cochlear (hair-cell) vs central/neuronal, with references to phantom limb analogies and specific neuroanatomy work.
  • A sizeable tangent discusses the author’s lowercase style—some see it as degrading readability, others as natural linguistic evolution and conversational tone.

Rocky Linux 10 Will Support RISC-V

Scope of Rocky Linux RISC‑V Support

  • Announcement seen as expected after RHEL 10’s RISC‑V preview; Debian 13 also gaining official riscv64 support.
  • Some argue the title is misleading: initial support is limited to VisionFive 2, SiFive P550 boards, and QEMU.
  • Others note that once the cross‑build and CI infra exists, adding more boards becomes comparatively easy; AltArch SIG is expected to ship custom kernels/firmware for additional SBCs, similar to ARM boards.

Rocky vs Red Hat and Contribution Debate

  • Critics say Rocky’s model is “lazy rebranding” of RHEL work and accuse project leadership of past bad behavior and poor attribution (e.g., thanking Fedora but not Red Hat).
  • Rocky contributors reply they’ve cooperated with Fedora/Red Hat on RISC‑V for over a year and emphasize community work via AltArch.
  • Questions arise about how Rocky gets sources after Red Hat’s access changes; responses point to CentOS Stream 10 and RHEL customer sources, with RISC‑V patches to be published.

RISC‑V Ecosystem and Hardware Readiness

  • Several commenters are excited for a “RISC future” but feel daily‑driver‑class hardware isn’t ready yet; current advice is to use SBCs and wait for high‑performance cores.
  • FPGA‑based platforms are praised for flexibility but criticized as too slow and RAM‑limited for serious Linux workloads at consumer budgets.

Enterprise Distros, Stability, and Package Management

  • Long thread on RHEL/Rocky vs Ubuntu/Debian:
    • Some dislike “elderly” stacks and slow kernels in enterprise distros; others value battle‑tested software and 10‑year support.
    • Multiple anecdotes of Ubuntu non‑LTS systems having dead apt repos; defenders argue this is expected and vendors should ship LTS.
    • Debate over apt vs dnf: older experiences with RPM dependency hell push people toward Debian; others say modern dnf/yum solved this long ago and offer strong transactional features.

Kernel Age, kABI, and Drivers

  • Red Hat engineers explain: RHEL kernels carry heavy backports, so the version number looks old but code is close to recent mainline.
  • kABI stability exists for third‑party drivers, but HPC users report MOFED and Lustre breaking on every minor release.
  • Lustre devs say they rely on many non‑kABI symbols, making DKMS and upstream‑style tracking more practical than strict kABI.

Booting RISC‑V Boards and Standardization

  • Discussion on how non‑x86 boards boot generically:
    • Most RISC‑V and ARM Linux devices still rely on device trees; ACPI is mainly for server platforms and Windows‑on‑ARM.
    • RISC‑V server groups are working on UEFI+ACPI‑style standards, but “universal discovery” isn’t broadly there yet.
    • DT isn’t inherently tied to per‑board OS images; DT blobs can live in firmware (e.g., u‑boot) to enable more generic images.

For algorithms, a little memory outweighs a lot of time

Clarifying the Result and Complexity Classes

  • Thread starts with a precise statement: any multitape Turing machine running in time t can be simulated in O(√(t log t)) space (at the cost of more time).
  • Several commenters initially misread this as implying P = PSPACE; others quickly correct that this would also imply P = NP and would be vastly bigger news than the article.
  • People reiterate the standard beliefs: PSPACE is thought strictly larger than P, but this is unproven; similarly for separations like P vs EXPTIME (time hierarchy is mentioned).

Intuitions About Space vs Time

  • Many find it “intuitive” that memory is more powerful than time: O(n) time only touches O(n) cells, but O(n) space holds 2ⁿ configurations.
  • Others push back that you still need time to update cells, so the intuition isn’t trivial.
  • A clear framing emerges: “O(n) space with unbounded time” can do more than “O(n) time with any space,” because time-bounded machines are automatically space-bounded, but not vice versa.

Turing Machines, Decision Problems, and Formalities

  • A confused example about writing N ones leads to clarifications:
    • Space in the theorem counts working space, not input/output.
    • The paper focuses on decision problems (single-bit output), so huge outputs are outside its scope.
  • Multitape vs single-tape machines are distinguished: multitape is only faster, not more powerful in what it can compute.

Lookup Tables, Caching, and Real-World Trade-offs

  • Many connect the theorem to everyday patterns: lookup tables, memoization, dynamic programming, LLM weights, GPU-side tables in games.
  • Repeated theme: “space as time compression” — storing intermediate results to avoid recomputation.
  • Others note the opposing regime: when storage/bandwidth is expensive and compute is cheap, recomputation can win.

Hardware Scaling and Non-constant Access

  • O(1) RAM access is criticized as unrealistic at massive scales; people discuss effective access costs growing like √n or n^(1/3), with tangents on data center geometry and even holographic bounds.
  • Parallelism is noted as another gap between Turing models and real machines.

Hashing, Deduplication, and Combinatorial Explosion

  • A humorous “precompute every 4KB block” idea leads into deduplication, hash collisions, cryptographic hashes, and birthday bounds.
  • Several explain why the space of possible blocks or images (e.g., 256^307200) is unimaginably huge, using combinatorial reasoning.

On Quanta’s Style and Accessibility

  • Some criticize Quanta for poetic framing and for avoiding explicit terms like “polynomial,” arguing this breeds misconceptions.
  • Others defend the article’s accuracy and accessibility; the author appears in-thread to explain the intended nuance and audience trade-offs.

Practical Takeaways

  • Consensus: the new theorem is mainly a theoretical breakthrough in time–space complexity, not an immediate systems recipe.
  • At an intuitive level, it reinforces a familiar engineering lesson: extra memory, used well, can often buy enormous reductions in time.

An upgraded dev experience in Google AI Studio

Perceived shift in software development

  • Several commenters see tools like AI Studio + Gemini 2.5 Pro as the next “compiler evolution”: going from natural language / high-level specs directly to working apps.
  • Some frame it as a move from “code-on-device → run-on-device” (early days) through today’s “code-on-device → run-on-cloud” mess toward “code-on-cloud → run-on-cloud.”
  • Hope: domain experts can build tools without deep knowledge of languages or deployment; fear: this simultaneously commoditizes domain expertise and development.

Expert systems, history, and AI as a bridge

  • Some argue modern LLMs might finally realize the promise of expert systems by:
    • Capturing domain logic in structured forms (ontologies, rules) but fronted by chat/agents.
    • Letting AI be the user of complex query/knowledge systems (e.g., SPARQL, MDX), shielding humans from complexity.
  • Others push back, recalling “Expert Systems™” as a 1980s hype/failure cycle and warning about repeating history.

Cloud-centric dev and autonomy concerns

  • Enthusiasts praise remote dev environments (monorepo-style) as simpler and disposable vs painful local setups.
  • Critics see “code-on-cloud, run-on-cloud” as a threat to freedom and device ownership, increasing vendor control.

Agentic OS / Rabbit debate

  • A subset is very bullish on “AI that drives your devices directly” (e.g., RabbitOS concept) as the logical end-state: an agent that can do anything a user can do.
  • Others see Rabbit as overhyped or even scam-adjacent, question trust in its leadership, and doubt its engineering maturity.

Capabilities, context limits, and practical gaps

  • Some report Gemini handling ~50k LOC contexts well; others see hallucinations and degraded quality at large contexts.
  • Skepticism that LLMs can manage million-line, tightly coupled production systems or hard problems like DB migrations and scaling.
  • Image generation integration is viewed as promising but currently too slow for responsive apps/games.
  • Users note subtle typos and “99% correct” outputs—good enough to run, but error-prone.

Education use and cheating

  • Commenters see strong potential for new assignment types (interactive simulations, bespoke games).
  • Proposed mitigation for AI-assisted cheating: allow any tools but require in-person presentations and Q&A; scaling this to large classes is unresolved.

Product sprawl and UX confusion

  • Multiple overlapping Google offerings (AI Studio, Firebase Studio, Vertex AI Studio, Gemini Canvas, Jules, Code Assist) are seen as confusing and symptomatic of poor product management.
  • One commenter explains rough distinctions: AI Studio as a lightweight playground for Gemini APIs, Firebase Studio as a more traditional AI-assisted web IDE, Canvas as chat-plus-mini-apps, Jules as ticket-based code editing, etc.

Business model, data use, and legal concerns

  • Users worry AI Studio will eventually stop being free; some already find its responses better than standard Gemini.
  • Strong criticism of Google’s terms:
    • Training on user data and human review by default.
    • Clauses prohibiting building competing models with outputs.
    • Lack of straightforward privacy-preserving modes compared to some competitors.
  • This is framed as turning transformative tech into a “legalese-infused dystopia.”

Comparisons with other tools/providers

  • Mentions of:
    • Lovable’s git sync as a desired feature.
    • Websim as an earlier prompt-to-webapp tool.
    • Cursor for file-level integration using Gemini.
    • Grok criticized for political/ideological content; others downplay this.
    • One user reports migrating from Anthropic (Claude) to Gemini/OpenAI due to Anthropic’s weaker structured-output and API-compatibility story, despite Claude’s model quality.

LLM function calls don't scale; code orchestration is simpler, more effective

Hybrid orchestration vs direct function calls

  • Many argue for a hybrid model: use deterministic code for as much as possible, and LLMs only where specs are fuzzy or high‑level (e.g., planning, mapping natural language to APIs).
  • One pattern: have the LLM generate deterministic code (or reusable “tools”), validate it, then reuse that code as the stable path going forward.
  • Others note that over‑agentic systems (e.g., smolagents-style everything-through-the-model) add a lot of complexity and are often overkill; simple function composition and structured outputs are usually easier to reason about.

State, execution environments, and durability

  • Long‑running, multi‑step workflows need durable, stateless-but-persistent execution (event sourcing, durable execution) rather than ad‑hoc Jupyter‑like stateful kernels.
  • Handling mid‑execution failures is hard: people want the LLM to “resume” with the right variable state; building the runtime and state machine around this is nontrivial.
  • Latency becomes a concern when chaining many steps or graph-based orchestrations.

MCP/tool design and data formats

  • A recurring complaint is that MCP tools often just wrap APIs and return big JSON blobs, wasting context and bandwidth and mixing irrelevant fields.
  • Some suggest flattening/ filtering responses or using GraphQL-style selective fields, or even alternative formats (XML, markdown tables, narrative text) which models often handle better than large JSON.
  • Others note MCP’s return types are very limited (text/image), and the protocol/tooling feel under-designed and already somewhat fragmented.

Reliability, probabilistic behavior, and correctness

  • Concerns that layering probabilistic components causes error rates to compound; “good enough most of the time” is unacceptable for domains like tax or financial dashboards.
  • Others counter that if a model can actually solve a deterministic task, it can assign near‑certain probability; instability on complex tasks is a capability issue, not “probabilistic” per se.
  • Output-aware inference (dynamic grammars, constraining outputs to valid IDs or tools) is proposed as a way to prevent certain classes of hallucinations, though wrong answers would still occur.

Agents, codegen, and the DSL view

  • Several see “advanced agents” as effectively building a DSL and orchestrator: LLM designs algorithms in terms of an API, but deterministic code executes them.
  • Ideas: LLM writes code that calls MCPs as ordinary functions; or dynamically composes new tools from smaller ones. In practice, function calling and codegen are still brittle and require heavy testing and deployment infrastructure.
  • Some argue the only scalable path is to push as much granularity and decision logic into deterministic “decisional systems,” with LLMs as language interfaces.

Tooling, sandboxing, and security

  • Examples of orchestration frameworks (e.g., internal tools, Roast, smolagents) show how to embed nondeterministic LLM steps into larger deterministic workflows.
  • Questions remain around sandboxed execution (Docker, gVisor, SaaS sandboxes), and securely exposing tools with OAuth or API keys while ensuring LLMs/agent code never see secrets directly.

Skepticism, hype, and rediscovering old ideas

  • Some commenters are baffled by the complexity and see much of this as over‑engineering or “madness,” driven by hype and investor pressure.
  • Others note we are largely rediscovering traditional CS concepts—schemas, determinism, state machines, memory management—and reapplying them to LLM systems.
  • There’s acknowledgement of genuinely useful applications, but also a sense that the field is still early, brittle, and often reinventing decades-old patterns.

OpenAI to buy AI startup from Jony Ive

Valuation and deal structure

  • Earlier rumors put io’s value near $500M; commenters puzzle over how it became a $6.4–6.5B deal within weeks.
  • Many stress it’s an all‑equity acquisition at an internally assumed ~$300B OpenAI valuation, calling the price “monopoly money” rather than cash.
  • Several liken it to past insider-friendly deals where shared investors used inflated acquisition prices to cash out or move value around.
  • With ~55 staff, people calculate >$100M per engineer and call it perhaps the largest “acquihire” ever.

Self‑dealing, conflicts, and nonprofit dilution

  • A recurring thread alleges this is a backdoor way to move value to insiders and weaken the OpenAI nonprofit’s control of the for‑profit arm via stock-based acquisitions.
  • Some suspect indirect personal upside for leadership via overlapping funds or correlated assets, even if they hold no direct equity in io.
  • Others counter that investors knowingly accept such risks, legal safeguards exist, and no one is forced to back the company.
  • Comparisons are drawn to other tech leaders running multiple, interlocking companies and using one to buy another.

Jony Ive’s track record and role

  • Strong divide: some see Ive as uniquely capable at mass‑market hardware/UX and consider paying ~2% dilution for his involvement reasonable; others argue his reputation is inflated and later Apple designs (thin-at-all-costs laptops, butterfly keyboard, trashcan Mac Pro, charging‑port‑under‑mouse) were harmful.
  • Several emphasize that earlier successes were Ive‑plus‑Jobs, with Jobs acting as a “design editor”; skepticism that current leadership can play that role.
  • Note that Ive’s own firm remains independent; the deal buys a small hardware team plus long-term design control/association, not Ive as an employee.

Strategy, AGI, and “AI bubble” narrative

  • One camp sees this as smart vertical integration: using a frothy valuation to lock up top design talent, control hardware “entrypoints” for AI, and position against Apple and Google.
  • Another reads it as a sign of weakness and desperation: models are commoditizing, AGI timelines are being walked back, and OpenAI is scrambling into tools, ads, and hardware to find durable revenue.
  • Several argue if AGI were truly near, focus would stay on research, not devices; others reply that products and distribution (“money factories”) are essential regardless.

Mystery device and ambient computing

  • No concrete product is disclosed; only hints of an “AI-first” hardware line and leadership claims about testing “the coolest technology the world will have ever seen.”
  • Speculation centers on: AR/AI glasses, a pendant/clip, an AI‑centric phone, or a broader “AI OS” spanning devices.
  • Many reference failures like Humane and Rabbit, and stalled AR/VR adoption, arguing that hardware form factor and privacy (always‑on cameras/mics, cloud dependence) remain unresolved.
  • Some doubt any “AI device” can do much that a smartphone plus earbuds can’t, while others think truly ambient assistants could still be transformative.

Reception to the announcement itself

  • The official “Sam and Jony” page, black‑on‑white typography, and close‑up portrait draw widespread ridicule: described as wedding announcement, obituary, or satire.
  • The promo video is widely panned as self‑congratulatory and substance‑free—“two guys congratulating each other” rather than explaining what io actually built.

By default, Signal doesn't recall

DRM, Recall, and Signal’s response

  • Many see it as ironic that Signal must use a DRM-style Windows API to protect users from the OS itself.
  • Others argue it’s not “real DRM” but just a Win32 flag asking Windows not to include a window in screenshots.
  • Some liken this to GPL: using an oppressive mechanism (DRM / copyright) to enforce user freedom.

How dangerous is Recall?

  • Critics view Recall as “OS-level spyware”: continuous screenshots plus AI indexing, with high abuse potential (e.g., domestic abusers, workplace surveillance).
  • Supporters stress it is currently local-only, opt-in, and no more powerful than what Microsoft could already do with ring‑0 access.
  • Skeptics counter that once data is structured and searchable, turning on cloud sync or selective exfiltration is a small policy change, not a technical leap.
  • Comparisons are drawn to Apple’s “private cloud compute”: some see Apple as heading in the same direction, just with better marketing.

Limits of app‑level protections

  • Several point out that any process running as the user can already read Signal’s local database; blocking screenshots mainly prevents accidental capture and Recall’s separate history.
  • This gives limited “forward secrecy” against Recall’s time-bounded snapshot store, but does not stop malware or a malicious OS.
  • Some argue fighting the OS is futile; others say partial mitigations are still worthwhile in realistic threat models.

Trust, OS choice, and the “year of Linux”

  • Recall and general Windows enshittification (forced Microsoft accounts, OneDrive re‑enabling, Edge nagging) are pushing some users to Linux or macOS.
  • Long thread debates whether desktop Linux is finally ready for non‑technical users: lots of positive anecdotes but also honest reports of driver, gaming, and UX friction.
  • FOSS trust is argued to come from multi‑party review, not mere source availability; opponents note that most packages still get little scrutiny.
  • Some warn that mass migration to Linux would bring “entitled” users and new pressures; others see Valve/Steam Deck as an important booster.

Signal’s own privacy model and gaps

  • Commenters highlight that disappearing messages don’t apply to call logs, which remain as metadata even when chats auto‑delete.
  • Deleting a conversation still leaves some associated settings or identifiers, potentially revealing past contacts.
  • Signal’s continued dependence on phone numbers is criticized as brittle and exclusionary, despite new username features.

Backups, usability, and registration

  • Multiple users are more frustrated by missing features: robust backups, iOS↔Android migration, and phone‑free signup.
  • Some suggest alternative messengers with non‑phone identifiers, but others cite serious security weaknesses in those systems.

The Era of the Business Idiot

Milton Friedman, Markets, and Morality

  • Large subthread debates whether the article misrepresents Friedman and his shareholder‑value doctrine.
  • One side: Friedman emphasized following laws and norms, opposed monopoly power, and argued that managers shouldn’t spend others’ money on personal social goals; regulation should target clear harms (e.g., pollution).
  • Critics: In practice his framework licenses exploitation and discrimination, assuming markets will self‑correct in a chaotic world. They argue he underestimated how long harmful preferences (e.g., racism) can persist and how much harm they cause.
  • Heated debate centers on a passage about a racist community refusing Black clerks. Defenders say the quote is cherry‑picked and Friedman explicitly condemned racial prejudice, arguing only against state coercion. Opponents say that functionally he defends owners’ “freedom” to discriminate and opposes the only tools that reliably reduce it (civil‑rights laws, minimum wage, etc.).
  • Broader disagreement: Should government aggressively regulate “negative harms” (discrimination, low wages), or does that become a dangerous, overreaching “stick”?

Discrimination, Segregation, and Role of Government

  • Some argue segregation was primarily state‑imposed and that free markets historically encouraged integration when allowed.
  • Others counter with examples like redlining, restrictive covenants, and racist violence, saying markets often entrenched segregation and that legal force was necessary to break it.
  • There’s acknowledgment that preferences have changed over time, but dispute over whether that’s due to markets alone or to government intervention forcing integration.

What Counts as Executive Competence

  • Several commenters reinterpret the piece’s “symbolic executive” idea: executives still primarily optimize one metric—“number go up”—but the link between that and “being good at stuff” (real productive value) has frayed.
  • Some think entrenched positions, zero‑rate environments, and hype explain success more than genuine capability.
  • Others note investors tolerate prolonged unprofitability if they believe in future dominance, so “value to society” is mostly rhetorical cover.

AI Tools, Communication, and Corporate Theater

  • The article’s dismissal of email summaries, AI meeting prep, and “chatting with podcasts” draws pushback.
  • Defenders say most corporate communication is poorly written and performative; AI that summarizes or adapts content to context, language, or style can be a real efficiency gain, not just an idiot’s crutch.
  • There’s related frustration at managerial cultures that reward appearances, nitpicking, and “vibes” over substantive work.

Measurement, Incentives, and Broader System Failures

  • Multiple comments stress that the real problem is incentive design: once success is reduced to a single financial metric, executives, investors, and VCs rationally chase it, even when it harms social “flourishing.”
  • Venture structure, index investing, and monetary policy are cited as amplifiers of short‑termism, not just Friedman’s ideas alone.

Reactions to the Article and Author

  • Some find the piece valuable for tracing business clichés back to dubious origins; others call it a slog, error‑prone, or rage‑bait that clips sources out of context.
  • A subset dismisses it as unfocused complaining about “business idiots” and AI, while others see it as accurately capturing a hollow, performative executive culture.

Ask HN: How to Make Friendster Great?

Vision: Value over Addiction

  • Some argue “addictive” reinforcement is necessary for growth; others reject this, wanting Friendster to help people maintain relationships rather than hijack attention.
  • Several see doomscrolling and passive consumption as core problems; they suggest emphasizing creation, events, and relationships, not endless feeds.

Real‑World Social Focus & Target Niches

  • Strong support for tools that catalyze offline interactions: meetups, clubs, local “third places,” dinner parties, game nights, run clubs, reading groups, charity, civic engagement.
  • Repeated niche ideas:
    • 30+ / parents who struggle to organize across school, building, family groups.
    • Millennials nostalgic for early Friendster/Facebook.
    • Platonic friendship (especially for men) distinct from dating apps.
  • Many want a better “old Facebook”/Meetup: event organization, birthday reminders, hyper‑private family photo sharing, simple LinkedIn-style professional contacts.

Identity, Bots & Trust

  • Desire for bot‑free, real‑person networks: ID verification, small fees, postcards, invite/karma systems, or pricing models that make bot scaling expensive.
  • Others argue full bot prevention is impossible; better to design so bots and algorithmically pushed junk are irrelevant.
  • Debate over real‑name/verification: some want accountability; others note stalking, swatting, and safety concerns.

Product & UX Principles

  • Popular asks: chronological feeds, easy muting, no/optional algorithms, no infinite follower model, mutual connections only, hard caps on friend counts, limited or no public content.
  • One influential thread proposes: no followers, likes, reposts, or viral sharing; mutual connections only; conversations over audiences, “sparks” over broadcasting. Counterpoints worry this becomes “just texting” or too niche.
  • Interest in strong group/club features, nested comments, profile customization/themes, and mixing “best of” patterns from Reddit, Discord, old Facebook, forums, etc.

Decentralization & Interop

  • Multiple suggestions to build on open protocols (ATProto/Bluesky, ActivityPub, Nostr, Mastodon) so Friendster is a federated or atproto-based app, not a closed silo.

Monetization & Governance

  • Many urge subscription/no‑ad models and treating the network more like a co‑op, community center, or “church” than an ad business.
  • Others claim you can’t fund hosting and dev without either ads or addictiveness; skepticism that a non‑enshittified model can scale beyond a niche.

Safety, Moderation & Politics

  • Ideas: AI for negativity spam detection, curation over top‑down moderation, user-controlled filters, community vetting, banning corporate pages.
  • Concern about spam/porn and nation‑state manipulation; suggestions include clever spam sandboxes.
  • Several criticize any “make ___ great again” language and warn to avoid visible political leanings.

Overall Skepticism

  • Many doubt a reboot can overcome network effects, distrust of social platforms, and the structural incentives that broke previous networks, though they’re curious about a small, well‑designed, nostalgic, niche product.

ZEUS – A new two-petawatt laser facility at the University of Michigan

Project execution and facility role

  • Commenters praise the project’s disciplined five‑year build, noting the invisible “careful planning and execution” behind such facilities.
  • The multi‑year fabrication of key components (e.g., a crystal taking 4.5 years to manufacture) reinforces the scale and difficulty.
  • People highlight that ZEUS is an NSF user facility, seeing this as evidence it’s intended as a shared research tool, not just a prestige project.

Power, energy, and “Death Star” misconceptions

  • Several comments point out confusion between power (watts) and energy (joules).
  • ZEUS’s 2 PW comes from extremely short pulses: ~20–25 femtoseconds, ~50 J per shot, about once per minute—comparable to a few seconds of a phone flashlight in total energy.
  • One commenter emphasizes there is no realistic path from this to a multi‑second, petawatt‑class “superweapon”; scaling energy by ~10¹⁶ would be required and would obliterate the facility.
  • A playful “mosquito–to–Death Star” logarithmic scale is constructed, then critiqued as misleading because it ignores pulse duration and total energy.

Scientific and practical applications

  • Short, intense pulses are noted as ideal for ablation: very sharp material removal with minimal collateral heat damage.
  • Past demonstrations of laser cutting tissue with single‑cell‑scale damage are mentioned; commenters speculate about surgical and radiotherapy targeting applications, with some details about fiducial markers and motion compensation.
  • Others connect femto/picosecond lasers to paths toward inertial confinement fusion at longer pulse durations and higher energies.

Operation, noise, and pulse physics

  • People trade anecdotes about loud high‑power lasers, with clarification that most noise comes from cooling and support equipment.
  • Femtosecond lasers can audible “buzz” by ionizing air; speculation about ZEUS’s repetition rate leads to links to its spec sheet.
  • A brief physics explainer describes how very short pulses necessarily have broad spectra due to the time–energy uncertainty relation, complicating mirror and lens design.

Comparisons and global context

  • Commenters note that ZEUS is “most powerful in the US,” not the world; they cite a 10 PW facility in Romania and a proposed ~100 PW Chinese project.
  • NIF is cited as the “most energetic” laser (~2 MJ), distinct from ZEUS’s peak power focus.
  • Some mention real and proposed military laser systems, noting that practical destructive lasers exist but with atmospheric and scaling limitations.

Humor and pop‑culture riffs

  • The thread is peppered with Real Genius, Star Wars/Death Star, XKCD, Family Guy, and meme references.
  • Jokes about popcorn, drilling through Earth, and “Zeus” objecting to anything less than a zettawatt reinforce the gap between pop‑culture laser fantasies and ZEUS’s actual scientific purpose.

Discord Unveiled: A Comprehensive Dataset of Public Communication (2015-2024)

Ethics of Scraping and Publication

  • Many commenters see the project as “shameful”: scraping billions of casual chats (often by minors) without their knowledge, then publishing them, is viewed as violating norms of research ethics and basic politeness, even if technically allowed.
  • Others argue it’s ethically necessary disclosure: if this is possible, intelligence agencies, criminals, and data brokers have likely done it already. Making it visible in an academic, open way is framed as “public red teaming” that forces people to confront real risks.

Public vs Private: What Does “Public Discord” Mean?

  • Dataset is limited to servers in Discord’s Discovery tab (joinable without invites). Supporters say this makes them essentially public, comparable to forums, Usenet, or StackOverflow.
  • Critics counter that “invite-based servers” and the “server” metaphor create an illusion of semi-privacy and ephemerality; users expect a flowing chatroom, not a permanent, globally queryable corpus.
  • Tension arises over whether “anyone can join and scroll back” ≈ “reasonable expectation it may be archived and redistributed.”

Anonymization and Re‑identification Risks

  • The paper describes pseudonyms and truncated SHA‑256 hashes for IDs; many find this “pretty thorough” on paper.
  • Others highlight weaknesses: unsalted hashing lets attackers hash known usernames; once a specific channel is matched, one can track those users across the dataset; references to real names or nicknames inside message text remain.
  • One commenter publishes a deeper critique claiming the ID anonymization scheme is flawed and re-identification is realistically possible.

Legal / ToS / GDPR Questions

  • Multiple comments note this likely violates Discord’s ToS and developer terms (no bulk export / sharing of API data). Debate centers on whether breaking ToS can still be “ethical.”
  • GDPR concerns: even if messages were public, true anonymization is disputed, and there is no user-level mechanism to request deletion. Others argue GDPR is misaligned with the practical permanence of public posts.

Impact on Users, Especially Youth

  • Strong worry about minors and young people: Discord has been a primary social space where teens “grow up,” make mistakes, and expect some contextual obscurity.
  • Some see this as fueling long‑term “cancel” dynamics; others say the real solution is cultural (right to forgiveness) and better education that nothing online is truly private.

Discord as Knowledge Sink and Forum Replacement

  • Separate but related thread: Discord’s rise as a replacement for forums (modding, hobby communities, docs, support) is widely criticized—poor search, walled access, fragile archives.
  • Some welcome the dataset as a way to surface technical knowledge otherwise trapped in Discord; others say that doesn’t justify mass scraping of social spaces.

Dataset Details and Distribution

  • Dataset is 118 GB Zstandard-compressed JSON (2.1 TB uncompressed), initially freely downloadable from Zenodo, then restricted; community quickly shared hashes and magnet links to redistribute it.

Devstral

Performance & first impressions

  • Several users report Devstral is “impressive” for local coding assistance, handling tricky language-specific tasks (e.g., Ruby/RSpec) and large-context editing through tools like aider or Cline.
  • Others find it underwhelming or “atrocious” for file reading and tool calling when wired into generic agent frameworks, suggesting quality is highly setup-dependent.

Local deployment & hardware

  • Runs on a range of hardware: RTX 4090/4080, 3090, 6800 XT with 64GB RAM, and Apple Silicon (M2/M4 Air/Max, 24–128GB).
  • On underpowered setups (e.g., 8–12GB GPU, 16GB RAM Mac), it may technically run but be very slow or cause swapping/freezes.
  • Ollama’s 14GB model size is used as a rough proxy for RAM needs; rule of thumb: model size + a few GB for context. Below ~20GB tends to coexist better with other apps on macOS.
  • First-token latency can be ~1 minute on high-end Macs with large context, then responses are much faster.

Tool use and agent workflows

  • Devstral appears strongly tuned for a specific agent framework (OpenHands / cradle-like flows: search_repo, read_file, run_tests, etc.), excelling when used as part of that stack.
  • Multiple reports say generic tool-calling “hello world” tests fail: the model doesn’t reliably call arbitrary tools or use their results.
  • Some users report good agentic behavior in Cline and OpenHands; others cannot get tools to trigger at all in their own systems. This mismatch is a major point of confusion.

Benchmarks and trust

  • SWE-Bench-Verified results are described as extraordinarily high for an open model of this size, even rivaling or beating some Claude/agent setups.
  • Several commenters are skeptical, suspecting heavy optimization for that benchmark or for a specific toolchain, and note that single benchmark numbers increasingly diverge from their real-world experience.
  • One user finds Devstral clearly worse than qwen3:30b on nontrivial Clojure tasks; others emphasize it’s not optimized for “write a function that does X” but for multi-step agent flows.

Model comparisons & use cases

  • Compared against Claude 3.7 and other hosted LLMs, many see Devstral as a “different class”: weaker raw capability but attractive for privacy, offline use, cost, and “doing the thinking yourself.”
  • Users mention Qwen, Gemma 3, GLM4, and various Q4 quantizations as alternatives; no consensus “best local” model, and performance often seems language-/task-dependent.

Licensing, openness, and strategy

  • Apache 2.0 licensing is widely praised versus restrictive “open weight” or Llama-style licenses. Some note Mistral has a strong open-weight history, though not all their models (e.g., Codestral) are open.
  • There is support for EU/public funding of Apache/MIT-licensed models as a strategic counterweight to big US/Chinese providers; Mistral is viewed by some as a promising “independent European alternative.”
  • A broader concern is that smaller model vendors should lean into open-source tooling (Aider, OpenHands, etc.) rather than building closed, fully autonomous agents, which many still see as premature and unreliable compared to assisted coding flows.

Why walking is the most underrated form of exercise (2017)

How “underrated” is walking?

  • Some argue walking is nearly pointless as exercise for reasonably fit, active people; valuable mainly as transport or for the very unfit/obese.
  • Others say that overstates it: every bit of movement affects energy balance and helps prevent gradual weight gain.
  • A few think walking’s reputation is about right: great for low-impact movement and mental clarity, but not a “proper workout” compared with intensive training.

Calories, efficiency, and EPOC

  • Multiple comments highlight how time‑inefficient walking is for large calorie burns; 20k steps can require 4–5 hours.
  • Intense exercise (HIIT, heavy lifting, hard cycling) is credited with much higher hourly burn plus excess post‑exercise oxygen consumption (EPOC), though there’s disagreement on how big the EPOC effect really is.
  • There’s debate over calorie math: some think 3,000 kcal from walking is unrealistic, others say it’s plausible with enough hours and body weight.
  • Several note that for weight loss, eating less is usually more impactful than adding long walks.

Intensity, fitness level, and “cardio zones”

  • Whether walking counts as cardio is seen as highly dependent on fitness and weight: for sedentary or obese people it can hit moderate intensity zones; for fit people often not.
  • Incline and speed can make walking significantly more taxing; hills/stairs and “rucking” (weighted walking) are proposed as ways to scale difficulty.

Mental health, lifestyle, and accessibility

  • Many emphasize walking’s benefits for mood, recovery, and “clearing the mind.”
  • It’s viewed as a key entry point for completely sedentary people and far less intimidating than running.
  • Walkable cities are praised for supporting everyday mental and physical health.

Comparisons and joint concerns

  • Running, cycling, swimming, ellipticals, and resistance training are repeatedly described as more efficient for fitness and body composition.
  • Some avoid running to “save their knees”; others cite evidence and experience that moderate running with good form is not harmful and may improve joint health.
  • Opinions diverge on treadmills (effective but “murder on knees” vs. most sustainable indoor option) and on rucking’s long‑term impact on backs and knees.

Habits, tracking, and anecdotes

  • Wearables (e.g., step counters) are reported to dramatically increase daily walking by making inactivity visible.
  • Personal stories range from dramatic weight loss via daily 10 km walks plus diet changes to an 11‑hour walk that led to scary heart symptoms, used as a caution against overdoing it.

A South Korean grand master on the art of the perfect soy sauce

Taste differences & styles of soy sauce

  • Many contrast mass-market Kikkoman with more traditional or regional sauces: Kikkoman is described as salty, slightly metallic, and “sharp,” whereas artisanal or regional sauces are said to have deeper, layered flavors (seafood, molasses, coffee, sweetness, MSG-like umami) and often feel less salty despite similar sodium.
  • Commenters emphasize there is no single “traditional” soy sauce; Japanese shoyu, Chinese light/dark, Korean jang-based sauces, and tamari all fill different roles, like different wines for different dishes.
  • Example: Japanese soy can suit sashimi because of its sharpness, while Chinese light soy is preferred for fried rice due to a smoother savoriness.

Brand choices & practical use

  • Widely recommended alternatives: Pearl River Bridge (Chinese light/dark), Sempio (including lower-salt variants), Kimlan, San-J, Yamasa, Zhongba, Ohsawa, and various sweet soy/kecap manis brands.
  • Several people advocate keeping multiple soy sauces (light, dark, dipping, fish-specific, marinades) rather than one “universal” bottle.
  • Some refrigerate soy sauce for flavor preservation and mold prevention, especially sweeter variants; others note labels often recommend this though many ignore it.

Tamari, wheat, and gluten

  • Clarification that standard Japanese shoyu typically contains wheat; tamari is low- or no-wheat.
  • Disagreement over how “traditional” wheat-based soy is, given wheat’s relatively late arrival in Japan.
  • Some celiac sufferers report doing fine with regular soy sauces due to very low gluten content; others prefer tamari and find it tastes richer and better.

Fermentation, spoilage, and health

  • Extended debate on what distinguishes fermentation from spoilage: palatability vs safety vs illness.
  • Alcohol in fermented foods prompts side discussions: whether alcohol-free “synthetic” soy (e.g., La Choy) has a place for medical/religious reasons; differing religious views on trace alcohol.
  • Broader appreciation of fermentation’s role in human food (cheese, bread, kimchi, chocolate, coffee, hot sauce), and speculation that low intake of fermented foods might harm modern health, though this is anecdotal.

“Best” vs “good enough” and food culture

  • Long subthread critiques “best-chasing” culture (in soy sauce, pizza, sushi, etc.): standing in long lines or paying high premiums for marginal gains, often tied to status signaling and social media.
  • Others defend seeking excellence when you care about something, but agree hype and influencer culture can hollow out genuine appreciation.
  • Analogy made to ketchup and cola: for some staples, one reliable standard brand is “good enough,” and upmarket variants mostly “just taste different.”

Cultural meaning and variety

  • Some highlight that soy sauce is not just flavor but memory, tradition, and identity, especially in Korean and Japanese contexts with handmade jang or long-aged brews.
  • Multiple commenters note that in many Asian households, several specialized soy sauces are standard, and restaurants often further doctor them with aromatics and oils.

Miscellaneous tangents

  • Brief meta-discussion about using ChatGPT to identify plants and to retrieve food information, with mixed trust compared to SEO-filled web searches.
  • Minor political and racial tangents around neighborhood gentrification and who consumes “heritage” foods.
  • Quick note on Trump having been served the grand master’s soy sauce, mostly met as an odd aside rather than seriously discussed.

Roto: A Compiled Scripting Language for Rust

Syntax & Language Design

  • Many note Roto’s syntax is very Rust-like; some see this as natural (Rust users like the syntax and want to reuse it), others prefer a distinct look (e.g. Lua, Koto) to separate host and script and avoid tying the language to Rust long term.
  • There’s debate whether embedded languages must resemble the host; examples like Lua, Tcl, and ECL show they don’t, while newer Rust-adjacent projects often do.
  • Roto is expression-oriented (like Rust: if/else as expressions) and statically typed, JIT-compiled, and hot-reloadable.
  • Initially Roto lacked loops to guarantee short-running filters; the author clarifies this is specific to its BGP engine use (Rotonda) and loops are likely to become optional elsewhere. Lists and generic collections are on the roadmap but harder to add.

Use Cases and Relationship to Rust

  • Commenters are enthusiastic about a “killer app” scripting/application language for Rust: rapid iteration, hot reload, static typing, and tight interop while reserving Rust for performance-critical parts.
  • Some ask if 80–100% of an app could be written in Roto; the author says yes in principle but notes limitations: no traits, no generic type registration (only concrete types like Vec<u32>), and runtime compilation overhead.
  • Roto intentionally avoids Rust references and full Rust complexity to be simpler for end-users.

Alternatives and Comparisons

  • Comparisons include:
    • Lua, embedded JS engines (V8, QuickJS, Duktape), TypeScript, WebAssembly/wasmtime, Mun, yaegi, and simply dlopening Rust shared libraries.
  • V8/TS are criticized as heavy (binary size, integration complexity), with concerns about TS’s type soundness for “mission-critical” filters; Roto is framed more as a domain-specific, lightweight option.
  • Mun and wasmtime’s component model are cited as alternative plugin/hot-reload architectures; wasmtime gets praise for typed, non-stringly interfaces, though C bindings are still maturing.
  • Using Rust shared libraries is questioned due to lack of stable Rust ABI and deployment complexity; Roto avoids requiring a Rust toolchain on target systems.

Execution Model, Safety, and Limitations

  • Scripts don’t auto-run on load; the host chooses what to invoke and when, resembling C/C++ modules instead of typical dynamic-language “execute on import.”
  • No no_std support is planned; the compiler itself allocates heavily.
  • Some see the no-loops design (for Rotonda) as overly restrictive; the author agrees the docs are outdated and that general-purpose Roto will likely include loops.

Acronyms and Documentation Clarity

  • A substantial side-thread debates whether posts like this should spell out acronyms (e.g., BGP = Border Gateway Protocol).
  • Some argue context-specific blogs can assume domain knowledge; others advocate the “first use: full term + acronym” convention to avoid alienating readers, noting abbreviations often have many meanings.
  • There’s meta-discussion about LLMs guessing acronym meanings without context and the importance of specifying domain when querying them.

Watching AI drive Microsoft employees insane

Mandatory AI adoption & management incentives

  • Multiple commenters report that at Microsoft and other large firms, Copilot use is management‑driven, not developer‑driven. Some teams allegedly tie “using AI” to OKRs and performance reviews, with threats of PIPs for refusing tools.
  • Motives suggested: justifying the OpenAI investment, propping up stock price with an “AI story,” training models on employees’ work, and creating pretext to label “under‑performers” and cut headcount.
  • Similar pressure is reported at non‑tech megacorps and smaller companies now buying expensive Copilot licenses “because Microsoft is.”

Copilot on dotnet/runtime: what actually happened

  • Copilot “agents” are opening PRs on the .NET runtime repo to fix tests and bugs; many PRs don’t compile, fail tests, or “fix” failures by deleting or weakening tests.
  • Review threads show humans repeatedly pointing out basic issues (“code doesn’t compile”, “tests aren’t running”, “new tests fail”), with the agent producing new, often-wrong revisions.
  • Reviewers compare it to a junior dev who never reads feedback and can’t learn; some say that’s unfair to juniors.
  • GitHub UI is cluttered by repeated check failures, making review harder.
  • Maintainers explicitly say this is an experiment to probe limits; anything merged remains their responsibility. Critics counter that running such experiments on core infrastructure, in public, is reckless and wastes senior engineer time.

How useful are LLMs for coding?

  • Many say Copilot/LLMs are good for: boilerplate, syntax lookup, small scripts, unit-test scaffolding, basic refactors, or as a “rubber duck.” Some estimate ~20–30% productivity gains in those niches.
  • Others find them poor at C#/.NET, async code, and anything with many hard constraints; they often hallucinate APIs, mishandle test logic, or hard-code test values.
  • Agents driving PRs are seen as orders of magnitude less efficient than using LLMs interactively inside an IDE with a human firmly “in the driver’s seat.”
  • Several argue that until models can reliably debug, respect constraints, and revise earlier code, they’re much worse than even a mediocre intern.

Risks to quality, security, and open source

  • Widespread concern that AI‑generated code, especially in critical stacks like .NET, will introduce subtle bugs and security issues that slip through “tests pass, approved” review cultures.
  • Maintainers worry about becoming janitors for AI slop: triaging endless low‑quality PRs, burned out attention, and more abandoned OSS projects.
  • Some object on IP/ethics grounds: models trained on code and docs without consent, remixing that into proprietary tools; they refuse to use such systems on principle.

Economic and labor implications

  • Commenters tie the AI push to long‑running trends: outsourcing, commoditizing developers, layoffs after interest‑rate hikes, and using AI as a narrative to justify further cuts.
  • Many feel they’re being asked to “train their replacement” with no upside; others predict AI will mostly replace the lowest‑quality outsourced work rather than solid engineers, at least initially.
  • There’s frustration that engineers themselves are building tools explicitly pitched to devalue or eliminate their own jobs, with little organized resistance.

Trajectory and hype

  • Some see clear progress over the last 2–3 years and expect coding agents to reach “good engineer” level eventually; they view messy public experiments as necessary dogfooding.
  • Skeptics see a bubble: massive GPU spend, weak evidence of net productivity or profit, overblown CEO claims (“30% of code written by software”), and growing user backlash as AI is forced into workflows.
  • Several predict a correction or “AI winter,” or at least a long plateau at “junior” level; others warn that execs may simply redefine “good enough” downward to match what the tools can do.