Show HN: Semantic Calculator (king-man+woman=?)

Overall impressions & comparisons

  • Many commenters find the tool fun, reminiscent of word games and “infinite craft”-style combinator systems.
  • The ranked list of candidate outputs makes it more engaging than a single answer.
  • Others argue that most outputs feel like gibberish with occasional hits, illustrating that the system has relational structure but no real “understanding.”

Behavior, UI, and dictionary quirks

  • Case sensitivity is critical: capitalized words often map to proper nouns (e.g., “King” → tennis player; “Man” → Isle of Man).
  • Red-circled words indicate missing entries; plurals, verbs, and some basic words (like “human”) often fail.
  • Proper nouns (countries, cities) must be capitalized to be recognized.
  • Mobile auto-capitalization and ad blockers can break interactions.

Amusing, odd, and failed equations

  • Users share many surprising or entertaining results (e.g., “wine – alcohol = grape juice,” “doctor – man + woman = medical practitioner,” “cheeseburger – giraffe + space – kidney – monkey = cheesecake”).
  • Simple arithmetic and chemistry are usually wrong (“three + two = four,” “salt – chlorine + potassium = sodium”).
  • Subtraction is widely seen as weaker and more random than addition.
  • Some directions in the space are “sticky,” e.g., “hammer – X” often yields something containing “gun.”

Biases and unsafe outputs

  • Several examples reveal gender stereotypes and offensive associations (“man – brain = woman,” “man – intelligence = woman,” biased race/crime relations).
  • Commenters stress that outputs reflect training data, not the author’s views, and suggest explicit disclaimers and/or filters.

Technical discussion: embeddings vs LLMs

  • The backend uses WordNet-based vocabulary with precomputed embeddings (mxbai-embed-large), excluding query words from results.
  • Commenters note that the classic “king – man + woman = queen” is heavily cherry-picked; often the closest vector is “king” itself unless excluded.
  • There’s debate about high-dimensional geometry, the “curse of dimensionality,” and how meaningful vector arithmetic really is.
  • Several compare this direct embedding math to LLM behavior: LLMs, with attention and context, often produce more intuitive analogies when asked to “pretend” to be a semantic calculator.
  • Others discuss nonlinearity of modern embedding spaces and why naive addition/subtraction works only sporadically.

Ideas and extensions

  • Suggestions include: decomposing a given word into a sum of others, using different embedding models, improving bias handling, and gamifying the system.