Show HN: LocalGPT – A local-first AI assistant in Rust with persistent memory

Use of LLMs for Docs and Perceived “Slop”

  • Many comments criticize that the README/post are clearly LLM-generated, interpreting this as low-effort “AI slop” and symptomatic of a broader decline in craftsmanship.
  • Counterpoint: some argue LLMs are excellent for documentation, especially for people who otherwise wouldn’t write any; if the docs are correct, they don’t care that an LLM wrote them.
  • Skeptics question whether authors rigorously verify LLM-written docs, noting that LLM docs are often generic, partially wrong, or too close to implementation details.
  • The “local-first” branding itself is cited as an example of misleading or un-proofread LLM copy (“No python required”, “local-first” while defaulting to cloud APIs).

“Local-First” Naming and Architecture

  • There is persistent confusion and criticism over the name: users see ANTHROPIC_API_KEY and conclude it isn’t truly local.
  • Others point out the code can target any OpenAI/Anthropic-compatible endpoint, including local Ollama/llama/ONNX servers; cloud is a fallback when local isn’t configured.
  • Some commenters object that calling a cloud client “local” (even if data files like MEMORY.md stay local) dilutes the term; if internet/API keys are required, they don’t consider it local-first.
  • A few users are pleased that proper local models are at least supported and that the memory format (Markdown files) reduces data lock-in.

Comparison to OpenClaw and the Agent Ecosystem

  • Several see this as essentially an OpenClaw clone (same MEMORY/SOUL/HEARTBEAT pattern) with fewer features; they ask what the unique value is beyond “written in Rust”.
  • Others are glad to have a Rust, single-binary alternative and complain OpenClaw is a “vibe-coded” TypeScript hot mess: race conditions, slow CLI, broken TUIs, complex cron, poor errors.
  • There’s interest in whether LocalGPT can safely reuse OpenClaw workspaces and how it handles embeddings + FTS5 for mixed code/prose.

Security, Capabilities, and Autonomy

  • Multiple threads highlight the “lethal trifecta”: private data + external communication + untrusted inputs, e.g., an email tricking the agent into exfiltrating data.
  • Proposed mitigations:
    • Manual gating of sensitive actions (OTP/confirmation), with concerns about fatigue.
    • Architecting agents to only ever have two of the three “legs”.
    • Object-capability and information-flow–style systems: provenance/taints on data, fine-grained policies at communication sinks, dynamic restriction of who can be contacted.
    • Credential-less proxies like Wardgate that hide API keys and restrict which endpoints/operations are allowed.
  • Users also worry about agents autonomously hitting production APIs or modifying files outside a sandbox.

Local vs Cloud Models and Cost

  • Discussion on what local models are feasible (e.g., 3B–30B open models, Devstral, gpt-oss-20B), with trade-offs in speed and especially context length versus frontier models like Claude Opus.
  • Some say frontier cloud models are still unmatched; others argue many tasks don’t need that level, but manually deciding when to use which model is burdensome.
  • Debate over economics: local GPUs require high upfront cost; cloud subscriptions/APIs are cheap today but may rise; competition (Mistral, DeepSeek, etc.) might keep prices low.
  • Observations that current $20/month tiers are already usage-limited and that tools like OpenClaw can burn through API credits quickly.

Implementation, Tooling, and UX

  • Rust is defended as a good fit: high-level enough, strong types/ownership for correctness, and easy single-binary distribution.
  • Several users hit build issues on Linux/macOS (OpenSSL, eframe needing x11, ORT warnings); some workarounds are shared.
  • SQLite + FTS5 + sqlite-vec for local semantic search are praised.
  • Some lament lack of observability in agents (no clear “what is it doing/thinking?” or audit logs) and suggest runtimes like Elixir/BEAM for better supervision.
  • There’s disagreement over target users: some argue normal users need turnkey local setups without API keys or Docker; others note that a CLI + Rust toolchain clearly targets technical users.