Latest AI news, expert analysis, bold opinions, and key trends — delivered to your inbox.
On Monday, Stripe previewed a new billing feature designed specifically for AI companies struggling with one core issue: unpredictable model costs.
Here’s the play.
Instead of manually absorbing or awkwardly passing through LLM token costs, startups can now automatically add a markup on top of raw model usage. Want a 30% margin across providers? Stripe tracks API pricing, measures customer token consumption, and applies the markup in real time.
As Stripe put it:
“Say you’re building an AI app: you want a consistent 30% margin over raw LLM token costs across providers. Billing automates the process.”
This matters because AI pricing is messy.
Many startups rely on subscription tiers with usage caps. Once users exceed those limits, overage fees kick in. Last year, Cursor shifted from “unlimited” usage to rate-limited plans with additional charges — a move that reflected just how dangerous uncontrolled token consumption can be.
The risk is even sharper for agentic startups. The more users interact with AI agents, the more tokens they burn from providers like OpenAI, Anthropic, or Google (via Gemini). If pricing isn’t tightly structured, heavy usage can push a company into the red fast.
Stripe’s bet?
Make AI infrastructure costs programmable — and profitable.
Instead of treating model expenses as a volatile liability, startups can now lock in margins at the billing layer itself.
In the AI era, the real moat may not just be the model.
It’s who controls the margin.