Perplexity Deep Research
Performance vs Other Deep Research Tools
- Mixed comparisons to OpenAI’s Deep Research: some find Perplexity’s outputs shorter, shallower, or “worse,” others report equally strong results and better UX (e.g., CSV export working properly).
- Users note Perplexity struggles with tasks needing exhaustive coverage and data joining (e.g., “college majors of all Fortune 100 CEOs,” full 50-state tables), where OpenAI/Gemini sometimes do better.
- Several say Perplexity’s free Deep Research is “good enough” that they’ll drop paid OpenAI tiers; others find it clearly inferior or underwhelming on first try.
Use Cases, Strengths, and Failures
- Works well for:
- Summarizing single topics (e.g., Amiga 500 sound chip, niche political issues, GDP PPP ratios).
- Rapid initial exploration, surfacing relevant papers and references.
- Fails or under-delivers for:
- Complex recommender-system design, where it regurgitates boilerplate from common blogs.
- Domain-heavy or activist-skewed fields, where it uncritically amplifies utopian or unrealistic proposals.
- Trending topics or “how to combine X and Y” queries, where it rephrases the question without real implementation depth.
Speed, Thoroughness, and “Expertise”
- Perplexity’s research completes in seconds to ~1 minute; OpenAI’s often runs for several minutes.
- Some suspect OpenAI’s longer duration is partly artificial (traffic smoothing / marketing signal: “it took a long time, so it must be deep”).
- Multiple experts describe all current “deep research” systems as shallow: authoritative tone, neat structure, but weak insight and limited contextual understanding.
Commoditization, Moats, and Product-Market Fit
- Commenters see rapid feature copying: Gemini → OpenAI → Perplexity → open-source clones. “Deep Research” is widely seen as a generic term and now a “term of art.”
- Debate over Perplexity’s moat:
- Pro: early, polished LLM+search experience; multi-model access; credible Google replacement for some.
- Con: seen as a thin wrapper over foundation models, with no clear defensible edge, and “desperate” recent moves.
- Broader worry that foundation-model vendors are commoditizing entire app categories (RAG, vision, agents, research tools), threatening vertical AI startups.
Evaluation, Hallucination, and Ecosystem Effects
- No strong, agreed-on benchmark; GAIA and “test it on my expertise” are mentioned, with frequent hallucinations and mis-weighted sources.
- Some describe Perplexity+DeepSeek R1 as notably better at targeted sourcing, but still not academically rigorous.
- Concerns that AI search/deep research will drain traffic from publishers, likely accelerating paywalls.
- Ethical reservations surface around leadership behavior (e.g., strike-breaking offer) and lack of visible dogfooding inside AI companies.