Salesforce's Agentforce 360 puts governance at the center of enterprise AI
Salesforce has a new pitch for enterprise AI, and it's more grounded than the usual chatbot talk. With Agentforce 360, announced ahead of Dreamforce, Salesforce is bundling three things: a policy layer for AI agents, a build-and-test environment to s...
Salesforce wants Slack to be the control plane for enterprise AI agents
Salesforce has a new pitch for enterprise AI, and it's more grounded than the usual chatbot talk.
With Agentforce 360, announced ahead of Dreamforce, Salesforce is bundling three things: a policy layer for AI agents, a build-and-test environment to ship them safely, and Slack as the place those agents actually show up. The pitch is straightforward. Enterprise AI keeps stalling in pilot mode, so Salesforce is packaging the missing pieces around control, governance, and distribution.
That matters. Model quality still matters too, but most large companies aren't stuck because Claude or Gemini can't summarize an email. They're stuck because nobody trusts these systems to take actions, handle sensitive data, or behave consistently across teams. Salesforce seems clearer on that than a lot of competitors.
The parts that matter
The headline features are Agent Script, Agentforce Builder, and deeper Slack integration.
Agent Script, entering beta in November, is the most interesting piece technically. Salesforce describes it as a way to define agent behavior with explicit logic and policy controls instead of burying everything in natural-language prompts. Think if/then rules, tool bindings, routing logic, and guardrails around what an agent can say or do.
That sounds dry until you've had to maintain one of these systems. A lot of agent behavior today lives inside giant prompt blobs that nobody wants to review, version, or audit.
A system that says:
- if an account is delinquent, follow collections policy
- if confidence is low, escalate to a reasoning model
- if PII appears, redact and log it before output
is a lot easier to reason about than "please be helpful, compliant, and concise."
For teams trying to move agents into production, that's a real shift. Prompts are flexible, but they're brittle. Policy graphs and structured tool schemas are less magical and much easier to control.
Agent delivery needs a CI/CD spine
The second piece is Agentforce Builder, also due in beta in November. Salesforce is positioning it as a centralized environment to build, test, and deploy agents, with environment separation and versioning for prompts, policies, tools, and what it calls Agentforce Vibes.
The "vibe" branding is dumb. The underlying idea is solid.
Most enterprise AI rollouts get messy fast because every team creates its own prompt variants, tone rules, escalation instructions, and edge-case handling. A few months later, support agents speak one way, sales copilots another, and nobody can explain why similar workflows produce different answers. If Builder gives teams golden test sets, staging environments, policy versioning, and repeatable deployment, that's useful. If Vibes becomes a way to standardize role behavior and domain-specific defaults across apps, it addresses a real governance problem even if the name deserves some ridicule.
There's also a basic engineering point here. Enterprise agent systems need the same disciplines as application delivery:
- version control for behavior
- scenario testing before release
- rollback paths
- separation between dev, staging, and prod
- auditability when things go wrong
A lot of AI products still treat that as optional. It isn't.
Slack is carrying a lot of this strategy
Salesforce's third move is the one executives will talk about most: turning Slack into the front end for enterprise agents.
Sales, service, IT, and HR agents surface directly in Slack channels and DMs. Slackbot becomes a personalized assistant. Salesforce also says Slack will act as an enterprise search layer, with connectors to Gmail, Outlook, and Dropbox starting in early 2026.
Part of this is product design. Part of it is distribution. Distribution matters. An AI tool buried in a separate enterprise console usually gets ignored. One that shows up where people already ask for help has a better shot.
It also creates hard engineering problems that Salesforce still has to solve.
If Slack becomes the agent hub, the system has to respect Slack permissions, private channels, user identity, SSO mappings, and downstream app entitlements. It has to index messages and files without leaking content across teams. It has to answer fast enough that users don't feel like every request turned into a background batch job.
For everyday Slack workflows, two seconds is already on the edge. Any serious implementation will need caching, aggressive tool selection, access-aware retrieval, and careful gating of slower reasoning models. A generic "send everything to the smartest model" setup will break on both latency and cost.
Reasoning models help, and they also make things harder
Salesforce says Agentforce 360 supports reasoning models from Anthropic, OpenAI, and Google Gemini. That lines up with where enterprise AI is heading. Companies want fast, cheap models for routine classification and retrieval, then stronger models for ambiguous or high-risk decisions.
That architecture makes sense. Use a lighter model to triage the request, fetch relevant records, and decide whether a workflow is simple or messy. Escalate to a heavier reasoning model when the task involves compliance, planning, exception handling, or conflicting data.
The upside is better accuracy where it matters. The downside is cost, latency, and operational complexity.
If Salesforce gets the routing right, customers will care. If it doesn't, these agents turn into expensive assistants that still need humans hovering nearby.
Trust is still the hard part
Salesforce has spent years pushing its Einstein Trust Layer, including data masking, audit logs, and limits on provider data retention. This announcement doesn't spell out every control, but it clearly follows the same line.
That matters because the current enterprise AI problem is largely about trust boundaries.
Technical teams want answers to boring questions:
- Can this system redact PII before it leaves our tenant?
- Can it enforce field-level security from Salesforce and permission rules from Slack?
- Can we trace an action back to a user, policy version, and data snapshot?
- Can legal and security teams review logs after the fact?
- Can we keep retention and eDiscovery aligned with existing Slack Enterprise controls?
If those answers are fuzzy, deployment stalls. That's part of why Salesforce citing 12,000 Agentforce customers deserves some skepticism. Customer count isn't the same as deep production usage. The more revealing stat in the background is the MIT finding that roughly 95% of enterprise AI pilots fail to reach production. That number explains the whole product direction.
Strong position, familiar trade-off
Salesforce has a real advantage here. It already owns a lot of the business system context, especially in CRM and service, and it owns Slack. That gives it a better chance of grounding agent actions in actual enterprise workflows instead of leaving them as chat demos.
Microsoft has a similar story with Teams, Copilot, and the rest of the Microsoft stack. Google is still trying to turn Gemini's model strength into the same kind of enterprise operating surface. Anthropic has momentum through partnerships, including Deloitte and IBM, but it doesn't own a day-to-day work surface like Slack or Teams.
Salesforce's stronger argument isn't frontier model performance. It's packaging and operations.
There is an obvious trade-off. The deeper a company goes into Salesforce's policy tooling, vibe system, Slack-native workflows, and app integrations, the harder it gets to move later. You're not just wiring up APIs. You're encoding business process inside a vendor-specific control plane.
Some companies will take that deal. Shipping matters. Getting out of pilot hell matters. But the lock-in risk is real.
What technical teams should watch
If you're evaluating this seriously, ignore the agent branding for a minute and look at the implementation details.
Data and retrieval
If Slack, Gmail, Outlook, and Dropbox all feed the same search and action layer, retrieval quality becomes a security issue. Indexing has to be permission-aware. Documents need freshness signals so the agent doesn't cite an old HR policy or expired contract. Chunking and metadata tagging matter because bad retrieval quietly poisons good models.
Tool use
Any agent that can take action should have explicit tool schemas with validated inputs and outputs. Loose schemas lead to hallucinated parameters, malformed actions, or workflows that fail halfway through because the agent guessed.
Testing
Scenario testing isn't optional. Teams need golden datasets for known edge cases, along with policy regression checks when prompts or routing logic change. If Builder handles that well, it may end up being one of the most useful parts of the release.
Performance
Slack users won't wait around for a long reasoning chain. Fast-path the common cases. Cache frequent answers. Keep reasoning models off the hot path unless confidence drops or policy requires review.
The bet
Salesforce is betting that the next phase of enterprise AI will go to whoever makes agents controllable, testable, and close to where work already happens.
That's a better bet than shipping another pile of chatbot wrappers. Execution is the hard part. If Agent Script gives teams a reliable control layer, and if Slack can carry agent workflows without turning into a latency swamp or permission mess, Salesforce will have something stronger than another enterprise AI demo.
A lot of vendors are still selling intelligence. Salesforce is selling operability. For enterprise teams, that's usually what decides whether a system ships.
Useful next reads and implementation paths
If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.
Design AI workflows with review, permissions, logging, and policy controls.
How risk scoring helped prioritize suspicious marketplace activity.
Venture investors are making the same call again: next year is when enterprise AI starts paying off. This time, the pitch is less gullible. TechCrunch surveyed 24 enterprise-focused VCs, and the themes were pretty clear. Less talk about bigger chatbo...
Aaron Levie made a useful point at TechCrunch Disrupt. Enterprise SaaS apps are not about to vanish under a swarm of autonomous agents. They’re becoming the structured layer agents sit on top of. That matters because a lot of enterprise AI is still s...
May Habib is taking the AI stage at TechCrunch Disrupt 2025 to talk about a problem plenty of enterprise teams still haven't solved: getting AI agents out of demos and into systems that actually matter. A lot of enterprise AI still looks like a chat ...