thomas.wieberneit@aheadcrm.co.nz

The Illusion of Value: Why Salesforce’s Agentic Work Unit is the New “Bad Query” of the AI Era

The Illusion of Value: Why Salesforce’s Agentic Work Unit is the New “Bad Query” of the AI Era

The News

On February. 25, 2026, Salesforce announced a pricing and metrics update. During the company’s Q4 FY2026 earnings call, CEO Marc Benioff, together with CMO Patrick Stokes, unveiled the Agentic Work Unit (AWU). Positioned as a metric to quantify the labor performed by autonomous digital systems, Salesforce defines an AWU as one discrete task accomplished by an AI agent.

According to Salesforce, this discrete task represents the exact moment “raw intelligence is converted into real work“. It is not a fixed unit but measured as a processed prompt, a completed reasoning chain, or an invoked tool. Salesforce explicitly designed the AWU to move the industry conversation away from the raw consumption of Large Language Model (LLM) tokens. As Benioff noted, tokens only measure “how much an AI talks,” whereas the AWU is intended to measure actual business execution.

The scale of this rollout is massive. Salesforce reported that its platform has already processed over 19 trillion AI tokens, translating them into 2.4 billion Agentic Work Units, with 771 million AWUs delivered in the fourth quarter alone. This new metric serves as the underlying foundation for Salesforce’s evolving Agentforce monetization strategy.

The bigger picture

Following a nearly 18-month period of pricing triangulation, which included a $2.00 per conversation model and a $0.10 per action “Flex Credit” model, Salesforce is leveraging the AWU to track system utilization, even as it wraps enterprise purchasing in familiar, unmetered per-user license agreements starting at $125 per user per month.  

To understand the significance of the Agentic Work Unit, one must view it through the lens of a broader industry crisis: the so-called “SaaSpocalypse” and the looming threat of the seat cannibalization trap. For two decades, the Software-as-a-Service business model has been dominated by seat-based licensing. However, as agentic AI systems mature and promise to be capable of executing multi-step workflows autonomously, they inherently reduce the need for human software operators. If an AI agent resolves 84% of tier-one support tickets without human intervention, the enterprise requires fewer human support seats. I have repeatedly written about pricing models, e.g., here, as part of my CustomerThink column.

This dynamic has forced the software industry into a frantic transition toward usage-based and attempts at outcome-based pricing models. Usage-based pricing, popularized by cloud infrastructure providers like AWS and data platforms like Snowflake, charges customers based on system consumption, e.g., compute seconds or data processed, or, these days, tokens consumed. While this protects the vendor’s margins and aligns with their variable cloud GPU costs, it shifts the financial risk of system inefficiency entirely onto the buyer. The vendors essentially play the role of Pontius Pilate and wash their hands in innocence.

Conversely, agile AI disruptors and customer service incumbents are aggressively pioneering true outcome-based pricing, where the billable event is delayed until a verified business success is achieved. For instance, Intercom’s Fin AI agent charges a strict $0.99 per successful resolution, while Zendesk recently started a $1.50 per automated resolution model. In these models, if the AI fails to resolve the customer’s issue, the customer pays nothing. Correspondingly, the vendor has skin in the game and needs to be interested in its software actually delivering value.

Salesforce’s introduction of the AWU represents a kind of a middle ground. Industry analysts like Constellation Research’s Liz Miller observe that the AWU acts as a placeholder for the agentic era, much like clicks and likes functioned in the early days of online and social media. It is still a usage-based consumption metric masquerading as an outcome metric. The industry is currently witnessing a tug-of-war: legacy giants are deploying metrics like the AWU to track utilization and justify high enterprise license costs, while pure-play AI vendors intend to leverage outcome-based pricing as a competitive weapon to steal market share by guaranteeing and demonstrating return on investment.

My point of view and analysis

As someone who has spent years helping organizations unlock their potential through digital transformation initiatives, I look at the Agentic Work Unit highly skeptical. When evaluating generative and agentic AI investments, the critical measure is the ability to deliver measurable business results, not just technological activity. In this context, the AWU represents a fundamental conflation: it equates doing work with achieving outcomes, which simply is not true.

By defining an AWU as a discrete task, such as invoking an API or triggering a workflow, Salesforce has created a metric that measures machine exertion rather than enterprise value. In the realm of autonomous systems, an AI agent can execute thousands of discrete tasks, burn through immense computational resources, and work incredibly hard while achieving absolutely nothing of commercial consequence.

Working hard on the wrong thing still doesn’t deliver results.

To fully grasp why measuring discrete AI tasks is a poor proxy for value, consider the analogy with an unoptimized database query vs. an optimized one in cloud data warehouses. A highly optimized SQL query returns a vital dataset in seconds for pennies. Conversely, a poorly written query forces the database engine into massive data scans. It might be running for hours and consuming plenty of CPU and memory resources. From the vendor’s billing perspective, the system successfully performed the discrete scanning tasks it was instructed to execute. However, it results in a massive consumption bill for the customer. The business gains little value, as much of it is harvested by the vendor; even worse, if the result is wrong. Yet the financial penalty is severe.  

The Agentic Work Unit operates exactly on this flawed economic principle. Autonomous AI agents are still highly susceptible to unique failure modes, e.g., the infinite reasoning loop. If an agent encounters an ambiguous prompt or lacks solid memory tracking, it may repeatedly call the same tool or query the same database in an endless cycle due to perfection bias. While engineers desperately build so-called Meta-Reasoners to halt this wasted computation, the AWU metric actively monetizes it. If a confused agent loops fifty times before timing out, it has successfully generated fifty AWUs delivering zero result. The customer is actively billed for the machine’s confusion.

Furthermore, agentic workflows suffer from compounding hallucinations, or epistemic debt. If an agent hallucinates a false premise in step one, it will still confidently execute subsequent tools based on that fabrication. By the time the workflow concludes, the agent may have triggered dozens of AWUs across multiple enterprise systems, corrupting data and requiring costly human remediation.

Ultimately, a dashboard celebrating 2.4 billion AWUs gives the illusion of massive productivity, but it is a vanity metric. If those tasks were merely redundant internal data reshuffling or failed reasoning loops, the actual profit multiplier of the organization remains unchanged. An Agentic Work Unit quantifies motion, but motion is not progress. Until AI pricing models mature to align the cost of digital labor with the verified delivery of business outcomes, enterprises must continue to treat effort-based metrics like the AWU with extreme skepticism.

Just my $.02. What do you think?