← Wagecore Blog
April 22, 202610 min readcase study · AI economics

The Klarna reversal, with the numbers

Klarna's Feb-2024 AI rollout and May-2025 reversal, with an illustrative reconstruction of the operational math. What capability without economic viability looks like when the workload is bimodal-complexity — and what the framework predicts about the failure mode.

By Andrei Kondrykau. Methodology is published at /methodology.

In February 2024 Klarna announced that an AI agent had taken over the work of 700 customer-service contractors, framing the deployment as a roughly $40M profit-improvement story. In May 2025 the CEO publicly acknowledged that the rollout had gone too far on quality, and the company started hiring humans back into customer service. The underlying repeat-rate or churn deltas have not been disclosed; the reversal is sourced to Bloomberg, Fortune, and CX Dive coverage of Klarna's own statements.

This is the cleanest public case study of capability without economic viability we have in production AI deployments. Capability was real — the model handled the volume — and the deployment still failed on quality, because capability is one of nine axes the operational cost depends on. Below is an illustrative reconstruction of the math, anchored to Klarna's public disclosures and labeled clearly where it uses third-party estimates or modeling assumptions rather than Klarna's own books. The lesson is not “AI doesn't work in support.” The lesson is that the operational framework predicted the failure mode, and most of the public discourse priced AI as if only the inference line mattered.

What the 2024 announcement actually said

The headline numbers Klarna shared publicly: the AI agent had handled 2.3 million chats in its first month, equivalent to the workload of 700 full-time agents, with average resolution time down from 11 minutes to under 2 and CSAT scores in line with human agents. Klarna framed the deployment as a $40M profit-improvement contribution for 2024. (Source: Klarna press release, February 2024.)

If you took only those numbers, the deployment looked nearly free of downside. The simple math, using third-party estimates of Klarna's per-agent fully loaded cost (~$60k/yr, plausible given Klarna's use of lower-cost geos for tier-1 support — not disclosed by Klarna) and a third-party estimate of the all-in AI cost ($1.5–3M annually at 2024 inference prices and the disclosed chat volume — also not disclosed by Klarna), comes to ~$42M displaced labor against ~$2M AI infrastructure: roughly a 14× ratio, before accounting for velocity gains.

Within the operational framework, this is what was missing from that analysis.

Where the math breaks: the long tail

Customer support workloads are not uniform. A bimodal distribution applies almost universally: 70–85% of tickets are simple, structured, and resolvable end-to-end with clear policy answers. The remaining 15–30% are complex — refund disputes that touch fraud, account recovery on edge-case authentication paths, hardship requests that require empathy and discretion, multi-party disputes across merchant and consumer.

On the simple band, AI handles the work with high reliability and low oversight cost. This is what the launch metrics captured. On the complex band, AI gives a confident-sounding answer that's wrong often enough to matter. The wrong answer doesn't just fail to resolve — it makes the situation worse, because the customer has already been told an outcome that doesn't materialize. They escalate. They complain to social media. They open a chargeback they would not have opened against a human agent who had told them “I can't promise that, let me check.”

Klarna's CEO publicly acknowledged that quality outcomes had dropped; the company has not disclosed the underlying repeat-contact or NPS deltas. Below, we model a 25% rise in repeat-contact rate on the complex band as an illustrative load test — not a Klarna figure — because that magnitude is consistent with what the four other public post-mortems of similar AI-support rollouts (none of them Klarna) reported in 2023–2025. The point is to show how a small rise in complex-band repeat-rate flips the deployment's net cost.

Illustrative operational math

The numbers below are a modeled reconstruction — Klarna has not published cost breakdowns. They use the operational-cost framework from the previous post : five line items beyond inference. Treat it as a worked example of how to project an AI deployment against a bimodal-complexity workload, not as Klarna's actual P&L.

Take a Klarna-comparable team handling 30 million tickets per year. Assume the simple-complex split is 80/20. Simple tickets take an average of 3 minutes of human time at $30/hr loaded ($1.50/ticket) and have an audit rate AI deployments cap at 5–10%. Complex tickets take 18 minutes at $45/hr loaded ($13.50/ticket) and require 25–35% audit. Error-cost multiplier: 1.5× on simple, 4× on complex when the case goes wrong.

Pre-deployment baseline: 24M simple tickets × $1.50 + 6M complex × $13.50 = $36M + $81M = $117M total labor cost. Plus overhead: $30M. Call the baseline $147M.

The optimistic deployment scenario — what Klarna's launch numbers implied — assumed 80% of tickets auto-resolved (the full simple band), the complex band stayed with humans, and the complex band didn't change. Math: 24M × $0.05 inference + $0.10 oversight (5% audit at 0.5 minute reviewer time) = ~$3.6M for the simple band. Complex band held at $81M. Plus overhead: $30M. Total: $114.6M. Modeled savings: ~$32M annually, which is in the neighborhood of the $40M Klarna projected as a profit-improvement contribution for 2024.

What the failure mode looks like when error-cost touches the complex band: with our illustrative 25% rise in repeat-contact rate on the complex band, complex volume effectively grows from 6M to 7.5M. The 1.5M new complex tickets arrive in the senior queue with the customer already frustrated, which (in published support-ops post-mortems on comparable workloads) pushes per-ticket time from 18 minutes to 27. Senior queue cost: 7.5M × ($45/hr × 27/60) ≈ $151M. The simple band stays at $3.6M. Overhead: $32M (small bump for incident response and PR). Total: $186.6M.

That's not 32M in savings. That's ~$40M worse than the pre-deployment baseline. The simple-band savings were real but smaller than headline, and the complex-band cost grew by 86% — net negative.

The framework called this. The complex band is a Class 4 task in the four-substitution taxonomy: human-critical, where AI being confident-but-wrong is the failure mode, not a feature gap that closes with better models. The pre-launch projection treated the entire workload as Class 1 (replaceable) and got a 14× cost advantage that the actual mix didn't support. See the taxonomy explainer for the full framing.

Why the demo metrics lied (and what they actually measured)

CSAT in the first month wasn't a measurement of the deployment — it was a measurement of the simple band. Three things masked the complex-band failure:

Survey self-selection. CSAT surveys go out post- resolution. Customers whose tickets escalated weren't in the sample for their first contact. They got the AI answer, were told the ticket was resolved, marked CSAT, and only later realized the resolution didn't hold. The negative CSAT showed up on the second contact, weeks later, attributed to “senior support.”

Survivorship in the metrics dashboard. The deployment's dashboard measured tickets the AI fully closed. Tickets that got routed to humans were filed under “agent contacts” — separate dashboard, separate goal, separate story. Nobody at Klarna initially had a single line that showed ticket-touches-per-customer, which is the only metric that catches re-contact rate as a system-level signal.

Time-lag in the failure mode. The simple-band savings showed up in week one. The complex-band damage showed up over the next 6–12 months as the cohort of bad first-contact resolutions worked its way through the escalation queue, fraud disputes, and social media. By the time the leadership team saw the trend line in repeat-contact rate, the deployment had been celebrated in the financial press for half a year.

What generalizes

The Klarna pattern is not Klarna-specific. The same shape applies any time three conditions hold:

(1) The workload has a bimodal complexity distribution where the complex band has high error cost. Customer support has it. So do medical-triage chatbots, insurance-claims first-pass review, tier-1 legal advice. Anywhere a confidently-wrong answer makes the downstream situation worse, not just unresolved.

(2) The launch metrics measure the simple band in isolation. Resolution time, deflection rate, CSAT-on-resolution — all simple- band metrics. None of them catch repeat-contact rate or time-to-final-resolution at the customer level.

(3) The economics of the simple band look so good that they justify the deployment without modeling the complex band at all. This is the critical move. A 14× cost advantage on the simple band has to be weighed against the complex-band cost multiplier, not its absolute baseline.

The corrective discipline is to model both bands, model the error- cost multiplier on the complex band explicitly, and pick the deployment scope to keep the AI in the band where it has a defensible cost advantage. Klarna's public statements about the reversal point in this direction — hiring humans back into the parts of the workload where AI was producing lower-quality outcomes, without retracting the simple-band AI deployment entirely. The new equilibrium is presumably cheaper than the original baseline, just not by 14×.

What the case is worth

Klarna's reversal is currently the single most-cited public example of AI deployment economics breaking down, and it's worth that citation. But the more useful version of the lesson is not “AI customer support fails.” It's “deploy AI against the band of work you can model rigorously, not against the band you wish you could.” The framework — capability + reliability + error-cost + integration + human-advantage damping — was sufficient to predict this in 2024. The product industry mostly chose not to use it.

If you want to run this kind of analysis on your own role, or on a team you're considering automating, Wagecore computes the per-task substitution distribution and operational cost against today's capability matrix. The wizard takes about two minutes; the methodology is open at /methodology . The org-level version of the same calculation is at /org/preview — paste your roles + headcount, see the org-level heatmap and the 5-year financial projection.

The Klarna reversal, with the numbers — Wagecore | Wagecore