Hacker plants false memories in ChatGPT to steal user data in perpetuity
Local vs Cloud LLMs
- Many advocate running models locally for privacy, control, and “uncensored” use cases.
- Recommended stacks: Llama 3.1 (esp. 8B), InternLM2 7B, Gemma/Gemma 2, Codestral, Aya, via tools like Ollama, GPT4All, LM Studio, Msty, OpenWebUI.
- Hardware views:
- For 20–70B models with Q4 quantization, high‑VRAM GPUs (e.g., 16–24 GB) are preferred.
- Smaller (≤13B) models can run on CPU with enough RAM but are slower.
- Quantization quality differs: Q4 often degrades output noticeably vs Q8/fp16.
- Consensus: small local models are decent for translation, summaries, and as memory aids, but unreliable for precise or obscure topics.
Prompt Injection, Memory, and Data Exfiltration
- Several point out that running locally does not inherently fix prompt injection or exfiltration: if the model reads untrusted content, it can be “hacked” regardless of location.
- Core limitation: LLMs do not reliably distinguish “instructions” from “data” or separate channels (user text, web pages, their own past output).
- The discussed attack uses hidden instructions (e.g., in an image or document) to:
- Force the model to call URLs that encode user inputs/outputs.
- Store this behavior in persistent memory so future chats keep exfiltrating.
- Some compare this to remote code execution / running a random executable from the internet; others stress that enabling memory by default makes it a notable new vector rather than user misconfiguration.
Mitigations and Red‑Team Dynamics
- Ideas: a separate model to detect/flag injection attempts; tools like Llama Guard; stronger observability; treating all prompts as untrusted input that must be sanitized.
- Multiple comments expect an “arms race” where attacking remains cheaper than defending; prior machine‑learning security experience suggests defense is harder and more expensive.
Usefulness vs. Risk of LLMs
- Strong divide:
- Enthusiasts report major productivity/learning boosts: rapid prototyping, code scaffolding and unit tests, complex text editing, finding libraries/tools, legal issue overviews, creative writing, studying complex topics interactively.
- Skeptics highlight frequent hallucinations, misleading confidence, and user over‑trust, especially for legal/financial advice and technical answers that reference non‑existent APIs or cases.
- Some call LLMs “idiot amplifiers” that let non‑experts produce authoritative‑looking nonsense, which then gets repeated, upvoted, and eventually scraped back into future training data.
- Ongoing debate: whether this is mainly a user‑education problem or a design/marketing failure, given LLMs are presented as general‑purpose “AI” rather than as fallible text generators.