Search all text in New York City

Overall reception and uses

  • Many commenters find the project delightful and “exceedingly fun,” describing it as something they could spend hours exploring.
  • People immediately use it to find personal landmarks (e.g., childhood bagel shops) and local culture (graffiti writers, stickers, slogans, political posters).
  • Some note its value for OSINT and imagine that intelligence agencies likely have similar tools at global scale.

Playing with the search

  • Users test funny or crude words (“fart,” “pedo,” “sex,” “foo,” “fool”), getting amusing misreads and coining it as a kind of game.
  • Another game emerges: find real English words with the fewest hits; examples like “scintillating,” “calisthenics,” “perplexed,” “Buxom,” etc.
  • People search for graffiti tags, politicians’ names, slogans, and niche phrases to probe cultural traces across the city.
  • Food terms (“bagels,” “pizza,” “sushi,” “hotdog,” “massage”) reveal dense and uneven spatial distributions; one person notes sushi is heavily Manhattan‑centric.

OCR quality and quirks

  • Multiple comments say the idea is brilliant but current OCR accuracy is “pretty bad” for many queries.
  • Misreads of Google watermarks, cropped signs, and partial words generate large numbers of false positives.
  • Some searches work well for clear signage; others show systematic errors: “OPEN” → “OBEY,” “food” → “foo,” and numerous comical reinterpretations.

Technical and cost considerations

  • Commenters estimate OCR compute as manageable on consumer hardware, but highlight Google Maps / Street View API costs (tens of thousands of dollars at list prices) as the real barrier.
  • Discussion notes ~8 million panoramas processed; various back‑of‑the‑envelope calculations of image throughput and API fees appear.
  • A linked talk suggests the creator used publicly-available Street View imagery and macOS’s built-in OCR via Shortcuts, possibly without paid API access; it’s unclear how rate limits were handled.

Related projects and desired extensions

  • Links to similar efforts: earlier Brooklyn‑only and London versions, a New York traffic‑camera semantic search project, and a UK building‑safety use of Street View.
  • Several people want an API, deduplication of near-identical views, CLIP/semantic image embeddings, or a “text‑only Street View.”
  • Others imagine this as a Google Maps layer for discovering niche businesses by sign text.

Data freshness, filtering, and tangents

  • Some try to infer the capture timeframe from protest posters and political signs.
  • There’s curiosity about why some official notices or offensive words are hard to find and speculation around mild censoring in the write‑up’s links.
  • One tangent raises the lack of simple, accessible text‑to‑speech tools for blind users; replies point to cost and existing assistive tech rather than this project specifically.