Why AI Cost Attribution Is Broken (And How to Fix It)
AI adoption is accelerating across SaaS companies, ISVs, and enterprise software teams.
But while usage is scaling rapidly, one critical capability has not kept up:
Cost attribution.
Most organisations cannot answer simple questions about their AI spend:
- Which feature is driving the most cost?
- Which customers are the most expensive to serve?
- How much AI usage is internal vs product?
- How much is development vs production?
- Which region or product line is driving spend?
This is not a reporting problem.
It is a structural problem.
The Illusion of Visibility
AI providers give you dashboards.
These dashboards show:
- Token usage
- Model usage
- Total cost
At first glance, this looks like visibility.
In reality, it is not.
Provider dashboards answer:
"How much did you spend?"
They do not answer:
"Why did you spend it?"
The Root Problem: AI Lacks Context
An AI request is just an API call.
It contains:
- A prompt
- A model
- A response
What it does not contain is business context.
It does not tell you:
- Which feature triggered the request
- Which product it belongs to
- Whether it was dev or production
- Whether it served a customer or an internal user
- Which region it originated from
Without that context, cost attribution is impossible.
Why Traditional Approaches Fail
Most organisations try to solve this problem after the fact.
1. Spreadsheet Allocation
Finance teams take the provider invoice and attempt to allocate cost manually.
This is:
- Time-consuming
- Inaccurate
- Not scalable
2. Engineering Logs
Engineering teams attempt to reconstruct usage from logs.
This creates:
- Inconsistent tagging
- Missing data
- High operational overhead
3. Cost Per Token Thinking
Many teams rely on "cost per token" as a primary metric.
But tokens are not the problem.
Tokens are just the unit of measurement.
The real problem is:
What generated those tokens.
The Real Requirement: Dimensions
To understand AI cost, you need to understand it across dimensions.
Common dimensions include:
- Feature
- Product
- Customer or tenant
- Region
- Environment (dev vs production)
- Internal vs external usage
Without these dimensions, AI spend is a black box.
With them, it becomes measurable.
Example: When Attribution Fails
Imagine a SaaS company launches a new AI feature.
A month later, AI spend doubles.
Without attribution:
- Finance sees increased cost
- Engineering sees increased usage
- No one knows which feature caused it
With proper attribution:
- The increase is tied to a specific feature
- Token usage per request is visible
- Inefficient prompts are identified
- Cost can be optimised
This is the difference between reactive and proactive cost management.
The Dev vs Production Problem
One of the most common attribution failures is mixing development and production usage.
Engineering teams experiment heavily with AI.
Without clear separation:
- Development traffic inflates production costs
- Product margins become unclear
- Forecasting becomes unreliable
This is especially common in fast-moving ISVs and SaaS companies.
Internal AI vs Product AI
Another major blind spot is internal AI usage.
AI is used across:
- Support tools
- Sales automation
- Internal workflows
If this usage is not separated:
- Internal costs are mixed into product cost
- Product profitability is distorted
- Operational efficiency is hidden
Again, this is not a cost problem.
It is an attribution problem.
The Fix: Capture Context at the Source
AI cost attribution cannot be solved after the fact.
It must be solved at the point of request.
This means:
Every AI call should include metadata such as:
- Feature
- Product
- Environment
- Customer (where applicable)
- Internal vs external usage
This transforms AI usage from:
Raw API calls
into:
Structured, analysable data
The Role of an AI Gateway or LLM Gateway
An AI Gateway or LLM Gateway provides the ideal control point.
By routing all AI requests through a central layer, organisations can:
- Enforce required metadata
- Standardise provider access
- Capture consistent telemetry
- Analyse usage across dimensions
Instead of relying on fragmented logs, the gateway becomes a system of record.
From Cost Tracking to AI FinOps
When cost attribution is solved, new capabilities emerge:
- Feature-level cost analysis
- Customer-level profitability
- Region-based cost insights
- Token efficiency optimisation
- Budget control and forecasting
This is where AI FinOps begins.
Not with dashboards.
But with structured attribution.
The Bottom Line
AI cost attribution is broken because AI usage lacks context.
Provider dashboards show totals, not drivers.
Spreadsheets and logs cannot reconstruct intent reliably.
To fix this, organisations must move from:
- Token-level metrics
to:
- Dimension-based attribution
By capturing context at the point of request and analysing AI usage across features, products, environments, and customers, AI spend becomes understandable — and controllable.
In the era of AI-powered software, cost attribution is not optional. It is foundational.