Talkie: a 13B vintage language model from 1930
Local hardware & deployment
- Several commenters discuss VRAM needs. 20GB is borderline for 13B BF16 weights, though splitting layers across CPU/GPU via llama.cpp is possible but slower.
- Some compare high‑VRAM GPUs vs large shared‑RAM desktops; consensus: GPUs give more “usable” local LLMs, but you won’t “make your money back,” so buy what you’re happy to pay for.
- No GGUF is yet available; people note it should be convertible from the PyTorch checkpoint for use with tools like Ollama.
“Vintage” concept, data leakage & contamination
- The authors frame “vintage LMs” as trained solely on pre‑cutoff data to avoid benchmark contamination and post‑date knowledge.
- Commenters point out evidence of temporal leakage (e.g., anachronistic political facts, terminology, and future knowledge), arguing the model doesn’t fully meet its own “vintage” standard.
- Distinction is drawn between contamination by benchmark answers vs generic post‑cutoff text; some see them as nearly the same issue.
Behavior, style & capabilities
- Many are charmed by the 19th/early‑20th‑century prose: ornate, confident, discursive, and very different from modern LLM tone.
- Examples show it:
- Treats “computer” as a human job and distinguishes “digital” as “using fingers.”
- Gives period‑typical takes on India, empire, American Civil War causes, women, yoga, industrialization, etc.
- Produces speculative future visions (2025/2026, moon travel, computers) that feel like historical futurism.
- Users note a common pattern: first sentences may be accurate; then it drifts into plausible but wrong explanations, so it can “pollute your brain” if you don’t know the answer.
Historical bias, racism & ethics
- Commenters report explicitly racist, colonialist, and sexist outputs and stress that these reflect the surviving texts and power structures of the era.
- Some see this as historically honest and even desirable for future “uncensored” historical models; others find it troubling and question the value of partial moderation layered on top.
Epistemic snapshot & scientific testing
- Strong interest in using such models as “time capsules” or “epistemic snapshots” of a given era, comparable to other history‑only LLM projects.
- Several propose research uses: training models before key breakthroughs (e.g., relativity, nukes) to see whether they can rediscover them or predict events, though many doubt current LLMs could.
Speculation, simulations & future models
- People imagine combining era‑locked models with VR or personal archives to simulate past periods or one’s younger self, edging toward “time travel” or “simulation” experiences.
- Some are excited; others push back on simulation talk as philosophically dubious or psychologically risky.
Cost and practicality
- Back‑of‑the‑envelope FLOP and cloud‑pricing estimates suggest pretraining costs on the order of tens of thousands of dollars, seen as impressively affordable for bespoke models.