01/22/2026 | Press release | Distributed by Public on 01/22/2026 13:19
Participants:
The conversation around generative AI is no longer about whether the technology can work but how, and how quickly, enterprises can scale it securely and effectively to see true value - how they can span the proverbial last mile. But the road from pilot to production has significant challenges. Too often, the narrative is that AI is easy and seamless, which sets up customers for missed expectations.
This conversation with three Salesforce executives helps surface the fuller, more authentic picture.
Q: You're building the roadmap, helping deploy the agents, and working through the realities of the transformation into being an Agentic Enterprise. Madhav, you've distilled this journey into three distinct stages: Stage 1, Stage 2, and Stage 3. Could you introduce that framework for us?
Madhav Thattai: This framework is a direct result of our learnings over the past 12 months. We recognized that adoption happens in structured stages. Stage 1: Trust is all about getting the foundation right, ensuring executive buy-in and security. Stage 2: Design is where you actually build the agent, focusing on context and control. Stage 3: Scale is where you move from a successful pilot to full-enterprise implementation. We use this to help our customers navigate the journey, and we use it to inform our product roadmap based on what they are learning.
Stage 1: Trust and Alignment
Q: Stage 1 is about establishing trust. Mark, your team of forward deployed engineers puts you on the front lines. What does trust entail at the beginning of a deployment, and what do companies need to do to get the process started correctly?
Mark Wakelin: You need an executive sponsor who is going to put themselves on the line to be accountable for the outcome of the program. They must have clarity on the business use case and the specific CEO-level KPIs tied to ROI that they are tracking.
The other critical element, before you even look at technology, is legal and security compliance. Is the legal team okay with AI being used? Is the industry regulator okay with it? Are there any ethical or humane use issues we should be considering? If you can't answer those questions, you can't get to production.
Adam Evans: Trust isn't just about safety; it's about permissions. A public-facing agent needs different capabilities before and after a user authenticates. It must dynamically unlock skills without ever exposing data to the wrong person. There should be no possibility for it to go off the rails, no matter how nicely you ask it. So people need to build agents in a defensive way, just like all apps. And that's a learning process to get to those best practices.
Thattai: We're going from a world where everybody has thrown a bunch of stuff at the wall and maybe has a bunch of agents that might not even create that much value. They've let 1,000 flowers bloom. Without that business KPI, you're doing AI for the sake of AI. And that will not lead to value.
What we're finding is that to build an agent that actually does something at scale that creates real ROI is pretty challenging. The true business value for enterprises doesn't reside in the AI model alone - it's in the "last mile." That is the software layer that translates raw technology into trusted, autonomous execution. To traverse this last mile, agents must be able to reason through complexity and operate on trusted business data, which is exactly where we are focusing.
Stage 2: Design and Context
Q: Once those foundational elements are locked down in Stage 1, we move to Stage 2: Design, where the focus should be on context, control, and the UX. Mark, your team sees customers wrestling with their data infrastructure at this stage. What are some common pitfalls companies encounter?
Wakelin: The biggest issue is almost always data. We see that the data the customer needs to hydrate an agent and make it useful is either not available or not in the right form to be ingested. For example, an automotive customer wanted an agent to answer questions about vehicle specs. But that data was locked in complex tables, not prose. Our old ingestion engine couldn't read it, leading to hallucinations.
We're solving this with Data 360 and our acquisition of Informatica, which allows us to unify and harmonize metadata across the enterprise. It provides a live context window so agents don't guess or hallucinate - they act on a single source of truth.
Thattai: Probably our biggest learning this year is that agents will end up in a kind of "prompt doom loop" if all you're doing is trying to give them natural language instructions. It works for simple things. Got an agent answering questions on a website? Great. But as you move into more complex business processes, you need more than natural-language instructions. These core workflows benefit from combining LLM flexibility with deterministic execution. The LLM provides the engine, but to reach the destination safely, you need the maps and traffic rules - that's the deterministic layer. That's what has driven our hybrid reasoning engine with Agent Script.
Wakelin: For example, companies don't want their refund policy to be something that gets creatively interpreted by an agent. If you return the item within 60 days, you get your money back, and if you don't, you don't.
In Deloitte's case, they wanted to capture and create new customer prospects but at the same time be able to augment those customer records with all kinds of industry data and observations.
That called for a deterministic approach while elaborating that data object with learnings from the "joys" that a large language model can bring to the table. We were blending these two technologies for control, accuracy, and business value.
Evans: Then there's the UX. For an agent to work, it has to be integrated into the flow of work. In our case, Slackbot is becoming the front door to the Agentic Enterprise. It already knows your context, your teams, and your files. Instead of just being a place to chat, Slackbot acts as a personal work agent that you can access in the flow of work without the "toggle tax."
Thattai: UX for the experience tends to be an afterthought. All the time goes into the construction of the agent, but you really want to be design-led. You want to understand what experience you're trying to build for your employees, for your customers, and then have that drive what kind of tasks the agent is going to do. What things are going to be valuable? How is that going to materialize in the experience? When does the agent escalate and hand off to a human? That's really important.
Stage 3: Scale and Orchestration
Q: We've established the foundation of trust and the design principles. Now we move to Stage 3: Scale, the ultimate goal. Adam, you mentioned that customers want multi-agent orchestration but often don't have enough successful singular agents yet. What is the technical and operational reality of achieving scale?
Evans: The technical reality of agent-to-agent (A2A) is that the performance is not there yet. The way everyone is doing A2A now is they're "chatty," and as soon as you have two or three agents talking, the performance for a live user situation - like a voice interaction - is going to be unacceptable. You need subsecond responses. The solution isn't to connect up a thousand agents and hope for the best; it's to focus on context engineering and separating the functionality to be testable. We are investing in roadmap features that allow multiple agents to share memory and execute in a shared memory execution space that's faster for high-performance situations.
Customers that have successfully moved past pilots are focused on killer use cases that are driving ROI, rather than having a thousand agents that don't create value.
Thattai: At this third stage, as customers mature and they've built an agent and are starting to get value from it, they start to ask different questions. "Okay, I built this agent that is executing a process, but I want to do that for 10,000 employees." Or "I've got an agent that's creating value for a customer who's engaging on my website; now I want to do that for 10,000 customers proactively."
We are seeing the shift happen now. We have approximately $540 million in agentic AI ARR and 18,500 deals in under a year. Customers like Grupo Falabella are seeing Agentforce now resolve 60% of inquiries without human help. Just three weeks after launching their pilot agent in Colombia, they were already seeing support levels on WhatsApp of 70% - a sign that more customers were getting their questions answered without needing to call.
Q: How do you start to build out these experiences?
Thattai: Customers think about orchestration. How do they manage all of these complex connections and relationships across these agents? Most customers are realizing that there's an interesting organization that is emerging that is part line of business, part IT, part technology that all kind of come together to build these agents.
There's also an emerging class of agent managers. Eventually their jobs will be to manage these agents. It might be management of humans and agents together, but those people need tools. Are the agents performing? How do they become better? How do I know that the agent is delivering value? I need to monitor these things at scale, and I need to understand how they need to work.
Wakelin: The challenges might be the pace and scale at which this is happening. But the conversations our forward deployed engineers have had with customers have generated more than 100 product recommendations that are really shaping and driving the roadmap.
If you think the value is just in the model, you've missed the point. The hyperscalers provide the car, and the LLMs provide the engine. But Salesforce provides the last mile - the traffic signs, the maps, and the security systems that allow the car to actually drive itself safely to the destination.
Next year, what will really start to take hold is the blending of human and digital labor into a single workforce.
Go deeper: