New tools for building agents

What is an “agent”?

  • Thread spends substantial time arguing definitions.
  • One camp: “agent” is basically any LLM-backed program acting on a user’s behalf (almost all software).
  • Another: distinguishes workflows (predefined code paths orchestrating tools) from agents (LLMs dynamically choosing tools and steps at runtime).
  • Others reduce it to structured tool-calling/chatbots that turn natural language into backend function calls.
  • Some reference agent-oriented programming: autonomy, own event loop, reacting to environment, not just pure request/response.

Agents SDK, alternatives, and abstractions

  • Many like the new Agents SDK for being simpler and less “bloated” than earlier frameworks; some think it will kill weak agent startups.
  • Others prefer remaining framework-agnostic and recommend PydanticAI, smolagents, or other open-source agent frameworks.
  • Some view most “agent” needs today as simple workflows plus structured outputs, not needing high-level orchestration.

MCP vs OpenAI’s approach

  • Notable absence of first-class Model Context Protocol (MCP) support sparks criticism; seen by some as “anti-developer” and ecosystem-hostile.
  • Defenders say MCP can still be wired in via function calling and bridges; MCP is framed as a protocol for tools/context, not a full agent framework.
  • Some developers love MCP (especially for local tools, DBs, terminals); others find it heavy and over-engineered compared to minimal function-calling.

Pricing and economics (search, file, computer use)

  • Web search pricing (~$25–30 per 1K queries) widely called “absurd” and “not usable at scale,” especially versus Brave or Perplexity.
  • File search and computer-use pricing seen as expensive but less shocking; unclear yet which use cases can justify the costs.
  • Several note that both OpenAI and Google seem to price grounded search much higher than traditional web search.

Responses API, RAG, and Assistants deprecation

  • Responses API is viewed as a cleaner evolution with built‑in RAG, reranking, and “handoff” logic; some surprised by simplistic default chunking.
  • Assistants API is slated to sunset mid‑2026; OpenAI promises feature parity (threads, code interpreter, async) and a 12‑month migration window once ready.
  • Developers running significant traffic on Assistants wonder when to migrate; guidance is “no rush yet.”

Vendor lock-in, state management, and API churn

  • Strong concern that shifting state (history, tools, RAG, workflows) into OpenAI increases lock-in and switching costs.
  • Several experienced users say they keep using basic chat completions + structured outputs and manage all state themselves; they see frameworks and SDKs as unnecessary abstraction and fragile black boxes.
  • Comparisons are made to AWS: high-level managed services vs commodity primitives; many predict LLMs will commoditize, making lock‑in strategies fragile.

Other design and usability points

  • Some developers praise built-in tracing and integrations (e.g., AgentOps) but complain the OpenAI dashboard still lacks features like full cost tracing and exports.
  • Questions about structured outputs: JSON escaping issues for LaTeX/code, preference for YAML/Markdown, and confirmation that Pydantic-style typed outputs still work.
  • Requests for TypeScript SDK and Realtime/audio support in Agents; both not ready yet.
  • Curiosity about using Computer Use for usability testing is met with skepticism: AI’s interaction patterns differ from humans.

Overall sentiment

  • Mixed but engaged: enthusiasm for simpler abstractions, integrated RAG and tools, and genuine developer experience improvements.
  • Counterbalanced by worries about price, lock‑in, rapid API churn, and whether agent frameworks truly solve real production problems versus simple, explicit LLM calls.