A/B Test Bandwidth: The Currency of Innovation

A/B Test Bandwidth: The Currency of Innovation
Mårten Schultzberg, Staff Data Scientist
Mårten Schultzberg, Staff Data Scientist

If you want to experiment like Spotify, check out our experimentation platform Confidence and get a personalized demo.

Contact us

At Spotify — and increasingly everywhere — AI is transforming how we build products. Developers ship features faster than ever. AI agents brainstorm alongside humans and write production code. Iteration cycles that once took weeks now happen in days.

But here's the paradox: as our ability to build accelerates, a new bottleneck has emerged. It's not moving from idea to implementation anymore — it's moving from implementation to validation.

The question isn't "how quickly can we build things?" It's "how many experiments can we run to prove what things work?" A/B test bandwidth — the number and complexity of experiments a company can run, analyze, and trust — has become the true constraint on innovation.

This bandwidth determines how much of AI's productivity boost that companies can actually capture and convert into better products. The recognition of this shift is happening at the highest levels, and it's about time. At Spotify, we've seen firsthand how experiment bandwidth drives innovation, and AI has only amplified this advantage. We are not surprised that more companies are now seeing this value. Just recently, OpenAI acquired the experimentation vendor Statsig for $1.1 billion. And Lovable, a software company that helps people build software with AI, have declared they are planning to build experimentation and online optimization right into the core of their product. When companies building some of the most transformative AI technology of our time makes experimentation infrastructure a strategic priority through acquisition, it signals just how critical this capability has become.

AI's new abilities and emerging challenges in product development

AI brings remarkable new capabilities to product development. Its most immediate promise is the ability to generate endless product ideas, developers can prototype in minutes, AI agents suggest features continuously, and brainstorming sessions become idea factories. Beyond ideation, AI agents are becoming sophisticated QA partners that can review code for edge cases, flag metric instrumentation errors, simulate complex user flows, and spot inconsistencies between specifications and shipped features.

By elevating both ideas and implementation quality, AI enables us to increase experiment hit rates, making the first version we ship more likely to faithfully represent the actual hypothesis we want to test.

However, these new abilities create different kinds of challenges. More ideas create a different kind of problem than we've faced before. As Mark Zuckerberg told Dwarkesh Patel about Meta's situation:

We already have more good ideas to test than we actually have compute or cohorts of people to test them with.

The challenge isn't idea shortage — it's idea selection and execution quality.

Zuckerberg continued:

We need to get to the point where the average quality of the hypothesis that AI generates is better than the best humans.

This highlights a crucial distinction between two fundamental product questions that AI affects differently:

"Have we built the right product?" (the what question)

"Have we built the product right?" (the how question)

The implementation quality challenge matters more than you might think. In our experience, both from Spotify and with Confidence customers, poor implementation quality silently kills experiments. When ideas fail, it's often because the variant didn't work as intended, not because the hypothesis was fundamentally wrong. Teams waste precious experiment bandwidth debugging implementation issues disguised as product learnings.

The path forward isn't just about generating more ideas faster, but about using AI's analytical capabilities to improve both the quality of our hypotheses and the reliability of our execution.

Orchestrating experimentation at scale

Expanding A/B test bandwidth isn't just about running more tests — it's about running the right experiments at the right time with proper safeguards across complex, high-velocity product environments.

Sophisticated experimentation requires infrastructure that helps teams:

  • Plan strategically: Understand which experiments can run in parallel without interference, which need isolation, and how to sequence tests for maximum learning.

  • Monitor rigorously: Track guardrail metrics, detect unintended side effects, and identify underpowered tests before they consume valuable traffic.

  • Execute confidently: Manage overlapping experiments, variants, and rollouts across different surfaces and teams without conflicts.

  • Leverage offline evaluation appropriately: Offline metrics can play a vital role in this ecosystem, particularly for recommender systems and algorithmic changes. They offer fast, safe iteration cycles that let teams verify code changes don't introduce bugs or regressions. As discussed in Schultzberg and Ottens' work, offline evaluation is excellent for catching obvious problems and technical regressions. However, for final product decisions, offline validation is rarely enough. The key is using offline evaluation as a necessary first step in your evaluation funnel, not a substitute for real-world experimentation.

This level of coordination can't be managed with spreadsheets or scattered dashboards. As teams test more ideas across overlapping user bases and shared infrastructure, interference, and collision become serious bottlenecks. Leading experimentation platforms treat testing not as a one-off tool but as a core organizational capability. They don't just help you launch experiments — they help you scale thoughtful experimentation programs that maintain statistical rigor and product safety even under high velocity.

How to invest in A/B test bandwidth

In a world where AI accelerates everything, experimentation bandwidth becomes the ultimate constraint on progress. Companies that expand this bandwidth thoughtfully will capture the most value from AI-driven development. Those that don't will drown in unverified changes and unclear outcomes. Here's where to focus your investment:

  • Prioritize idea quality over volume: Use AI to improve hypothesis generation and selection, not to flood your test queues with mediocre ideas.
  • Invest in implementation quality: Use AI-assisted QA to ensure experiments test what they're supposed to test, not broken approximations of your hypotheses.
  • Invest in tooling that sets you up for velocity: Make sure the experiment tool you are using is enabling experimentation at scale rather than getting in the way of it.
  • Build systematic experimentation capabilities: Invest in experimentation tooling that scales your speed of development and innovation while balancing safety in a way that suits your business.

How Spotify has optimized its experimentation bandwidth

Spotify has long seen the value of experiment velocity. Maximizing the experiment throughput and bandwidth in strategic parts of our apps has been a focus since 2019. By building our platform Confidence with velocity in mind from the beginning we have managed to reach quite extraordinary throughput with hundreds of teams experimenting in parallel with very few hiccups. A key principle for us is that adding more experiments and teams shouldn't make it harder for any individual team to run their tests.

A key to scaling our velocity is the surface concept in Confidence. Surfaces let us group related experiments and enforce standardized settings. This makes it easy for teams to overview experiments that are affecting the same part of the user experience that they are working on. It also makes it possible to standardize experiments in the same part of the app with required metrics and reviews. This short video introduces Surfaces

Precise sample size calculation is another critical factor in maximizing bandwidth. Confidence's calculator goes beyond simple online tools by considering the full experimental context: decision rules, multiple metrics, analysis methods, treatment group structures, and more. This precision reduces a major bandwidth killer — underpowered experiments that consume traffic without delivering clear answers. In a world where experimentation capacity is a major constraint on the speed of innovation, only launching tests with a reasonable chance of producing actionable results is an important part of not wasting bandwidth.

As just one example, 58 teams ran 520 experiments last year on Spotify's mobile home screen alone (on the Home surface in Confidence). Without too much friction for the experimenting teams, we managed to start on average 10 new experiments every week, across 58 more and less independent teams.

In a world where building is getting faster, learning is becoming the constraint. Make sure your organization is ready. The companies that excel at A/B test bandwidth will have a significant competitive advantage in the AI era.

To see how your product teams can build with AI and experiment like Spotify, book a demo of our experimentation platform, Confidence, now also available externally.