Artificial Intelligence November 4, 2025

Elad Gil on AI markets with real winners and the categories still open

Elad Gil’s read on the AI market is blunt and mostly right. Some categories already have leaders with real staying power. Others still look busy, funded, and vaguely promising, but nobody has earned the right to call them won. That distinction matter...

Elad Gil on AI markets with real winners and the categories still open

Elad Gil’s AI market map is useful for one reason: some categories are already closed

Elad Gil’s read on the AI market is blunt and mostly right. Some categories already have leaders with real staying power. Others still look busy, funded, and vaguely promising, but nobody has earned the right to call them won.

That distinction matters more now than it did a year ago. Enterprise AI buying has sped up, but a lot of that spend is still pilot money. Pilots create noise. Durable products create pull. Gil’s framing, shared at TechCrunch Disrupt, cuts through that better than most investor commentary does.

His rough split looks like this:

  • Already consolidating: foundation models, AI coding, medical transcription, customer support agents, legal AI
  • Still open: finance tooling, accounting, AI security

That sounds obvious at first. It isn’t. The interesting part is why some categories tighten around a few players so quickly while others stay loose.

Where the winners are getting harder to dislodge

Start with foundation models. This market was never going to support hundreds of serious general-purpose players. Training frontier models now demands absurd capital, GPU access, and infrastructure maturity across the whole stack. Not just pretraining. Everything after it too.

A credible frontier model vendor needs:

  • massive mixed datasets, including strong code data and curated instruction corpora
  • reinforcement pipelines such as RLHF or RLAIF
  • alignment and policy systems that enterprises will actually trust
  • inference tricks like MoE routing, speculative decoding, KV cache reuse, and quantization like INT4 or INT8

Open weights help. They don’t erase the gap. Shipping a model is one thing. Running a full training-to-serving operation with strong evals for factuality, privacy, harmlessness, and policy compliance is another. That’s why the likely long-term leaders are still a short list: Google, Anthropic, OpenAI, maybe xAI, Meta, and Mistral.

The coding market is consolidating for a different reason. Distribution matters, but workflow fit matters more. The products pulling ahead, whether incumbent copilots or newer names like Cursor, Devin, and Windsurf, aren’t just wrapping an LLM around autocomplete. They’re building IDE-native systems with project context, file graph awareness, tool use, sandboxed execution, test generation, diff review, and continuous telemetry on what actually helps developers finish work.

That last part is underrated. A coding assistant gets better when it sees real usage: where suggestions are accepted, where they’re reverted, where latency causes abandonment, where pass@k looks good in a benchmark but falls apart in a sprawling production repo. Those feedback loops compound fast. New entrants can raise money and ship demos. Catching up on evals, UX, and integration depth is much harder.

Legal AI and medical transcription have another kind of moat: domain pain plus compliance.

Medical transcription is a good example of a category that sounds deceptively simple from the outside. It’s not just speech-to-text. A real clinical pipeline needs ASR tuned for messy medical audio, speaker diarization, medical entity extraction, note structuring, coding support for ICD or CPT, and clean hooks into EHR systems. Then come the non-negotiables: HIPAA, SOC 2, audit trails, retention policies, BAAs. Companies like Abridge and Ambience aren’t winning because they have a nicer prompt. They’re winning because they built a healthcare-grade system.

Legal AI has similar dynamics. Harvey’s traction says something important: firms will adopt AI products when citation fidelity, provenance, and workflow fit are good enough to trust. Legal users care about hallucinations more than almost any other buyer group, for obvious reasons. A system that can retrieve case law, filings, contracts, and discovery materials under tight constraints, then draft with citations that survive review, has a real product. A chatbot with legal vibes does not.

Customer support sits somewhere in the middle. The category has visible leaders such as Decagon and Sierra, while incumbents like Salesforce and HubSpot are bolting AI into existing stacks. The technical pattern here is pretty clear now: policy-grounded RAG, schema-constrained outputs, multi-turn orchestration, and heavy observability. Production support agents need escalation logic, guardrails, and action safety. If they can issue refunds, update accounts, or touch order systems, loose text generation is not enough.

Why finance, accounting, and AI security are still open

The open categories Gil points to are telling. They’re not weak markets. They’re attractive markets where demand exists but the shape of the product is still unsettled.

Finance and accounting look ripe for AI because the work is repetitive, document-heavy, and expensive. But those domains are full of edge cases, hard controls, and ugly integrations. General-purpose copilots can summarize statements or extract entities. That’s not the same as owning a workflow people will trust at close, during audit prep, or in reconciliation.

The same goes for AI security. It sounds like an obvious winner category because every enterprise now worries about prompt injection, data exfiltration, unsafe tool use, model abuse, and tenant isolation. But “AI security” is still a bundle of half-connected needs: model gateway controls, LLM firewalls, redaction, policy enforcement, access control, output scanning, audit logs, evals, and incident response. Some of that will live in standalone vendors. Some of it will collapse into cloud platforms and model providers. Some of it will end up in existing security tooling. That uncertainty is exactly why the field still feels open.

A lot of startups in these categories are selling a broad anxiety rather than a nailed-down product. That can work for a while. It rarely holds.

The technical pattern behind durable AI products

The categories that are consolidating tend to share three traits.

First, they have tight workflow integration. The model is embedded in the work, not hovering beside it. Coding tools live in the IDE. Medical transcription sits inside clinical documentation flow. Legal AI maps to draft-review-revise loops. Support agents connect directly to SOPs and backend systems.

Second, they have domain-specific evals. Generic model benchmarks are cheap theater at this point. Durable products track what matters in production: citation precision, bug-fix success, latency under long context, grounded answer rates, escalation quality, coding recall, clinical accuracy. If a team can’t tell you how it measures failure in its actual domain, you should assume it’s still doing demos.

Third, they have compliance and operational scaffolding built in. Prompt lineage, immutable logs, tool-call tracking, tenant isolation, PII redaction, retrieval audits, rollback paths. This stuff is boring until it isn’t. Enterprises buy trust as much as raw capability.

That has a direct implication for engineering teams deciding whether to build or buy.

If your company has a real edge in proprietary data, workflow ownership, or compliance posture, there’s still room to build. Fine-tune with LoRA if domain language needs help. Use SFT plus DPO when behavior really has to shift. Add RAG with hybrid retrieval like BM25 plus embeddings. Enforce JSON Schema outputs and validate everything. Keep deterministic steps around any tool use that can mutate data or trigger actions.

If you don’t have those advantages, platform gravity is real. Foundation model vendors already package structured outputs, function calling, and increasingly opinionated agent primitives. It gets harder to justify rolling your own stack when the only differentiation is prompt wording and a thin UI.

What technical buyers should pay attention to now

For senior engineers and tech leads, the useful question isn’t “which AI category is hot?” It’s narrower.

Ask:

  • Does this product own a workflow that users repeat every day?
  • Are the evals domain-specific and hard to fake?
  • Can the vendor explain latency, TTFT, and cost under realistic context sizes?
  • Is there a clear story for prompt injection, data leakage, and unsafe tool execution?
  • What happens when the model is wrong? Is there escalation, provenance, and a log you can audit?
  • If the model vendor improves overnight, does this product still matter?

That last one is a good filter. A lot of AI startups still look fragile when you imagine the base models getting 20 percent better and cheaper. The stronger ones survive that thought experiment because their value sits in data, workflow, compliance, and product integration.

Gil’s broader point lands because it matches what technical teams are seeing on the ground. Some AI markets already have leaders because the hard work has moved from flashy model demos to operational depth. Others are still early because nobody has packaged the pain into a product that holds up under real constraints.

That’s the part worth watching. Not who can generate the slickest output, but who can survive contact with production.

What to watch

The harder part is not the headline capacity number. It is whether the economics, supply chain, power availability, and operational reliability hold up once teams try to use this at production scale. Buyers should treat the announcement as a signal of direction, not proof that cost, latency, or availability problems are solved.

Keep going from here

Useful next reads and implementation paths

If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.

Relevant service
AI agents development

Design agentic workflows with tools, guardrails, approvals, and rollout controls.

Related proof
AI support triage automation

How AI-assisted routing cut manual support triage time by 47%.

Related article
Why VCs still think enterprise AI adoption finally starts next year

Venture investors are making the same call again: next year is when enterprise AI starts paying off. This time, the pitch is less gullible. TechCrunch surveyed 24 enterprise-focused VCs, and the themes were pretty clear. Less talk about bigger chatbo...

Related article
Why Benchmark is putting another $225M into Cerebras

Benchmark raising $225 million in special funds to buy more of Cerebras says two things pretty clearly. Cerebras is no longer priced like an oddball chip startup. It just raised $1 billion at a $23 billion valuation, with Tiger Global leading the rou...

Related article
Perplexity brings its Personal Computer agent to all Mac users

Perplexity has made Personal Computer available to all Mac users through its desktop app. The pitch is straightforward: give an AI agent access to local files, native Mac apps, web tools, and a large set of connectors so it can handle multi-step ...