Ceramic.ai is making a simple bet about how AI products get built next: if search becomes cheap and fast enough, the bot stops guessing and starts checking itself as it goes.
That is the real claim hiding inside the company’s new search pitch. On the latest Cognitive Revolution podcast, Ceramic founder Anna Patterson said the company can deliver search at $0.05 per 1,000 queries with 50 millisecond latency, cheap enough to make repeated background lookups part of the normal answer path instead of a premium add-on. If those numbers hold outside the demo, the product design consequence is bigger than one startup launch. AI systems could start verifying claims, pulling fresher information, and updating an answer mid-response without turning every grounded interaction into a budget problem. Cognitive Revolution Ceramic
Patterson’s argument is that inference got cheap first, while search stayed expensive. On the podcast, she said search has remained around $5 to $15 per 1,000 queries, making it “the most expensive part of the stack” even as model inference got faster and cheaper. Brave’s public API pricing supports the low end of that claim, listing search at $5 per 1,000 requests. The upper end should be read as Patterson’s market characterization, not a settled industry benchmark. Cognitive Revolution Brave Search API
Why that matters is less about a prettier benchmark chart than about what developers can afford to ship. Patterson said Ceramic’s “Supervised Generation” setup runs between 12 and 35 searches while composing a single response, and that the whole loop still costs about one-third of one Brave search. Her broader point is that retrieval stops being an occasional grounding step and starts becoming ambient infrastructure inside the answer itself. Cognitive Revolution
Ceramic’s architecture is built around that idea. Patterson said the system searches at the start of generation, then continues launching new searches as the model writes and discovers new subtopics it needs to check. At GTC, she said, Ceramic used NVIDIA’s Nemotron 3 Nano as a small verification or “introspection” model sitting alongside a larger frontier model that writes the final answer. Ceramic and NVIDIA’s March announcement similarly described Nemotron 3 Nano as the featured verification engine inside Supervised Generation. Cognitive Revolution National Law Review
The company did not start here. Ceramic launched in March 2025 as a training infrastructure company with $12 million in seed funding led by NEA, with IBM, Samsung Next, and Earthshot Ventures participating. On the podcast, Patterson explained the move toward search as a response to a practical customer problem: companies wanted fresh data in their models, but retraining is expensive and still leaves a model stale by the time it ships. Search, in her telling, is the cheaper way to keep answers current. Business Wire Cognitive Revolution
That framing also helps explain why Ceramic is trying to turn search into a product-layer feature rather than a back-end utility. Patterson argued that cheap, fast retrieval could matter most in products where latency and trust are visible to users: voice systems, robots, assistive devices, and any workflow where the model should be checking facts instead of confidently freelancing. That is a more interesting story than another startup claiming its endpoint is cheaper.
It is also still a claim in search of independent proof. Ceramic has not published a third-party benchmark showing that the advertised $0.05 pricing and 50 millisecond latency hold under real production workloads, especially for the kind of multi-search loops Patterson is describing. The architectural thesis is plausible. The evidence is still mostly Ceramic talking about Ceramic.
Still, if Patterson is even directionally right, this is the part worth watching: not whether one search API got cheaper, but whether grounded generation becomes cheap enough to feel normal. Cheap inference changed what developers were willing to ask models to do. Cheap retrieval could change what they are willing to trust them to say.