The Hidden Cost of AI Agents: What Every Vendor Pitch Deck Leaves Out
Jason Lemkin runs SaaStr, a B2B software community and events business that generates eight figures in revenue with three humans and a fleet of AI agents. In a LinkedIn post published in late March, Lemkin published what amounts to a real-world P&L for an AI agent stack, but the numbers that appear in vendor pitch decks are only half the story.
According to Lemkin's post, SaaStr's AI handles 70 percent of the company's closed-won deals. It has sent 60,000 hyper-personalized outbound emails through AI agents and held 2.75 million conversations through its AI advisor. An AI-powered sales development representative, known as an AI SDR, outperforms human SDRs 11 to 40 times on volume with better response rates, according to Lemkin. These are the numbers that make it into presentations.
The numbers that don't are harder to budget for.
The 30-Day Training Burden
Every agent SaaStr deployed required at least 30 days of intensive daily training. Lemkin estimates one to two hours per day of correction and fine-tuning per agent, a labor cost that doesn't appear on any agent pricing page. One of SaaStr's AI SDRs needed 47 separate corrections before it could handle pricing discussions correctly.
Lemkin's team found it could absorb roughly 1.5 new agents per month before quality started slipping. That deployment velocity ceiling is set by human bandwidth, not by agent capability or cost. Vendor presentations typically show the agent after the 30-day training period is complete, not the human hours that preceded it.
Silent Degradation: The Ops Risk Nobody Highlights
One of SaaStr's production agents quietly stopped ingesting new training data. No error message appeared. No alert fired. The agent kept running on an increasingly stale knowledge base for four months.
This is the failure mode that vendor documentation does not cover. AI agents do not fail loudly. They produce plausible outputs while their knowledge base quietly goes out of date. Standard monitoring tools do not catch it, including uptime checks and error rate dashboards. The only reliable detection is human review of agent outputs, which most teams do not have the capacity to maintain consistently.
Lemkin's summary: "AI agents are amplifiers. They take what works and multiply it. They take what's broken and multiply that too."
The Actual Moat
The chief revenue officer at Personio, a European HR software company, told Lemkin that the context you feed your agents is the competitive moat, not the agent technology itself. Any company can purchase agent infrastructure. What makes agents useful rather than generically competent is proprietary context: your product details, your customer history, your internal workflows.
This inverts how most agent vendors position themselves. If the moat is context and training, then competitive advantage sits in the ops and data infrastructure underneath, not in the agent layer. That has concrete implications: for buyers evaluating agents claiming proprietary technology as their differentiator, and for builders deciding where to invest.
Deploying agents also forces organizations to confront data quality problems that were previously invisible. About a third of SaaStr's Salesforce records turned out to be duplicates, discovered only because an agent read the full dataset end to end. The agent does not just execute; it audits.
What the Market Is Actually Buying
Agent infrastructure from companies like Salesforce (whose agent platform is called AgentForce), Artisan, Qualified, and Monaco can now be purchased and deployed in days, alongside tools built with AI coding assistants, a category of tooling known colloquially as vibe-coded tools. Lemkin's post suggests the constraint is no longer access to agent technology. It is the ops capacity to train, monitor, and maintain agents in production.
SaaStr found no single product that integrates AgentForce, Artisan, Qualified, Monaco, and vibe-coded tools into a unified management layer. This gap is itself a market signal: the agent orchestration layer is still fragmented.
For buyers, the durable lesson is that agent technology and agent ops capability are different purchases with different cost structures. Vendors quote per-agent pricing. The human hours required to train and monitor agents are not in the quote.
For investors evaluating agent startups, the question is whether a company is selling agent technology or agent ops infrastructure. Agent technology faces commoditization pressure as the underlying models improve. Ops infrastructure including training data pipelines, review processes, and monitoring is harder to replicate and defensible in a different way.
The agent market is moving from "can you build it?" to "can you operate it at scale?" Lemkin's post is an answer to that question, in numbers.