A/B Testing & Experimentation: Industry lessons¶
This document synthesizes insights from numerous articles published by leading tech companies like Netflix, Airbnb, Google, Uber, LinkedIn, Spotify, DoorDash, Facebook, Pinterest, Twitter, Shopify, Booking.com, Stitch Fix, Zalando, Traveloka, Dropbox, Better, Intuit, Grab and Lyft, detailing their A/B testing and experimentation methodologies, platforms, challenges, and innovations.
I. Evolution and Philosophy of Experimentation Platforms
Companies universally acknowledge A/B testing (online controlled experimentation) as the gold standard for data-driven product development and decision-making. The journey often starts with ad-hoc, manual A/B tests and evolves towards sophisticated, centralized platforms.
Early Stages (The “Stone Age” or “Crawl” Phase):
LinkedIn (Pre T-REX): Initial A/B tests involved manual bucket allocation (e.g.,
member_id mod 1000
) with no centralized tracking, leading to scaling issues and unreliable results. Test definitions were scattered in codebases, and deployment was tied to service releases.Zalando (Octopus - Crawl Phase < 2016): Teams set up tests individually. The first centralized platform, Octopus, focused on execution due to few initial customers, with manual analysis as a fallback. A key challenge was the lack of cross-functional knowledge between engineers and data scientists.
Traveloka (Pre-EXP < 2018): The first Experiments API had predefined filters and returned a single treatment parameter. It used a “look-up” caching mechanism for user assignment.
Maturing Platforms (The “Walk” and “Run” Phases):
Airbnb (ERF, 2014): Started as a tool to define metrics and compute them for a small set of experiments. Scaled to handle ~500 concurrent experiments and ~2500 distinct metrics daily.
Netflix (XP Platform): Evolved to support extensive A/B testing for every product change, from UI redesigns and personalized homepages to adaptive streaming algorithms and even title artwork. Handles vast scale and diverse test types.
LinkedIn (T-REX): Introduced concepts of “test” (hypothesis/feature) and “experiment” (single stage of testing/release with specific allocation). Deployed Lix DSL for defining experiments, enabling decoupling from application code and fast updates. Utilizes multi-level caching for low-latency evaluation.
Uber (XP): Moved from an older system (Morpheus) that struggled with correctness and flexibility to a new platform built on “parameters” (decoupling code from experiments) and a unified configuration stack (Flipr integration). The new system focuses on correctness by design, ensuring comparable cohorts.
DoorDash (Curie & Dash-AB): Developed Curie as an experimentation analysis platform to standardize processes, improve accuracy with scientifically sound methodologies (e.g., CRSE, variance reduction), and centralize results. Dash-AB serves as the central statistical engine.
Spotify (New Experimentation Platform): Rebuilt their platform focusing on ease of use, coordination (domains, timelines, exclusivity), holdbacks for cumulative impact, and a “salt machine” for dynamic user reshuffling without stopping all experiments.
Zalando (Octopus - Walk & Run Phases 2016+): Focused on establishing experimentation culture, improving data tracking, A/B test design quality, and analysis methods. Rebuilt the analysis system in Spark for scalability. Key challenges included data quality (SRM) and long runtimes.
Traveloka (EXP Platform, 2018): Adopted Facebook’s Planout framework for scalability, product context specificity, and better treatment parameter hygiene. Components include Ops Frontend/Backend and a client library using hashing for deterministic assignment.
Pinterest (A/B Testing Platform): Focused on real-time config changes, lightweight setup process, client-agnostic APIs, and a scalable analytics dashboard. Uses a UI for config, QA workflow, and simplified APIs.
Twitter (Duck Duck Goose - DDG): Created in 2010, processes terabytes of data. Engineers define eligibility, treatment buckets, hypothesis, and metrics. Uses Scalding pipelines for offline processing.
Dropbox (ML XR): Leverages machine learning (Expected Revenue - XR) to create a short-term proxy metric for long-term user satisfaction (subscriptions), accelerating experiment analysis from months to days.
Lyft: Focuses on addressing network effects, real-world dynamism, diverse lines of business, and fostering experimentation culture.
II. Experimentation Platform Architecture & Core Components
Most mature platforms share common architectural elements:
Configuration Management:
UI-driven Configuration: Pinterest, LinkedIn (T-REX), Traveloka (EXP), DoorDash (Curie) provide web UIs for setting up experiments, defining targeting rules, allocating traffic, and managing the experiment lifecycle. This reduces errors and standardizes setup.
Real-time Updates: Pinterest, LinkedIn, Uber emphasize the ability to change experiment configurations (ramp up/down, shut down) in real-time without code deploys, often using systems that push serialized configs to services (e.g., LinkedIn’s Lix DSL, Traveloka’s compiled Planout config).
Parameters/Feature Flags: Uber’s new platform uses “parameters” as the core abstraction, decoupling client code from specific experiments. Experiments override parameter values. This is a common pattern, where experimentation is an overriding layer on top of a feature flagging or remote configuration system (e.g., Uber’s Flipr).
Domain-Specific Language (DSL): LinkedIn’s Lix DSL (Lisp-like) defines segmentation and randomized splits, offering flexibility and deterministic evaluation. Twitter uses a lightweight DSL for defining event-based metrics.
Assignment & Randomization Engine:
Randomization Unit: Platforms support various units: users (most common), sessions (Lyft, Netflix), devices, requests, listings, clusters (LinkedIn, Facebook for network effects), regions (Lyft, Netflix for quasi-experiments), time-splits (DoorDash, Lyft). DoorDash’s Dash-AB and Uber’s new platform are designed to be generic regarding randomization units.
Deterministic Hashing: Most platforms (e.g., Stitch Fix, Traveloka, LinkedIn, Uber) use deterministic hashing (e.g., SHA1 of experiment_id + randomization_unit_id + salt) to assign units to variants. This ensures consistency for users across sessions without needing to store every assignment.
Bucketing & Allocation: Users are hashed into a fixed number of buckets (e.g., 0-999). Treatment groups are then defined as ranges of these buckets. Uber uses a tree structure for treatment groups to allow complex splits and dynamic changes.
Stratified Sampling: Netflix uses stratified sampling to ensure homogeneity across key dimensions (country, device type) in experiment cells, mitigating bias from purely random sampling. LinkedIn also used stratification for cluster randomization.
“Salt Machine” (Spotify): A system using a tree of “salts” for hashing users into buckets, allowing reshuffling for new experiments without stopping all ongoing ones, critical for maintaining randomization integrity with high experiment velocity.
Client Libraries/SDKs:
Provide an interface for applications to fetch treatment assignments (e.g., Pinterest
get_group
,activate_experiment
; LinkedInexperimentationClient.getTreatment
).Uber has SDKs for all languages and clients, incorporating caching, fallback to default values for reliability, and prefetching capabilities.
Traveloka’s client library syncs compiled Planout configs and calculates treatment locally.
Logging & Data Pipeline:
Exposure Logging: Critical for analysis. Logs are generated when a user is exposed to an experiment or a feature controlled by an experiment is accessed (e.g., Pinterest
activate_experiment
, Uber’s SDKs log on parameter access).Data Ingestion: Events are typically sent to a streaming system like Kafka (Netflix, Pinterest) and then processed by batch or stream systems.
Data Processing:
Airbnb (ERF): Migrated from monolithic Hive queries to dynamically generated Airflow DAGs, computing by “Event Source” (query defining several metrics) rather than per experiment or per metric.
Twitter (DDG): Uses Scalding pipelines in three stages: 1. Aggregate raw sources to per-user, per-hour metrics. 2. Join with experiment impressions, calculate per-user aggregates for experiment duration. 3. Roll up all experiment data for dashboards.
Netflix: Data engineering systems join test metadata with core data sets (member interactions, streaming logs).
Data Storage: Processed data is stored in data warehouses (Hive, Spark) and often also in low-latency databases for serving results (e.g., Twitter uses Manhattan; Netflix uses Cassandra, EVCache, Elasticsearch; DoorDash uses PostgreSQL).
Analysis & Reporting:
Automated Analysis: Platforms like DoorDash’s Curie and Netflix’s XP automate the generation of experiment results, applying standardized statistical methods.
Metrics Management:
Airbnb: Metric hierarchy (Core, Target, Certified metrics).
Twitter: Three types of metrics: built-in, experimenter-defined (via DSL), imported (custom aggregates). Metrics organized into “metric groups.”
DoorDash (Curie): Allows data scientists to define metrics via SQL Jinja templates. Moving towards a standardized metrics repository (uMetric at Uber).
Netflix: Democratized and modular XP allows direct contribution of metrics, causal inference methods, and visualizations.
Dashboards & UI: Most platforms provide a web UI (e.g., Netflix’s ABlaze, Pinterest’s experiment web app, DoorDash’s Curie WebUI) for experiment setup, monitoring, and result visualization.
Statistical Engines:
DoorDash (Dash-AB): Centralized Python library for statistical analysis, supporting various metric types (continuous, proportional, ratio), variance calculation methods (regression, Delta method, bootstrapping), CRSE, and variance reduction (CUPED/CUPAC).
Uber (Statistics Engine): Handles different metric types, outlier detection, variance reduction (CUPED, DiD for pre-experiment bias), and various p-value calculation methods (Welch’s t-test, Mann-Whitney U, Chi-squared, Delta method, bootstrap).
Zalando (Octopus): Initially wrapped an open-source stats library, later rebuilt analysis in Spark. Uses two-sided t-tests.
III. Experiment Design & Methodologies
Standard A/B/N Testing: The foundational method, comparing one or more treatments against a control.
Randomization Units: As mentioned, user-level is common. Others include:
Time-Split (Switchback): Used when interference is high (e.g., logistics, pricing). All units in a region/time window get the same treatment, then switch. (DoorDash, Lyft, Uber). DoorDash found ORS (Operations Research Scientists) implementing their own switchback experiments in production to be most effective.
Cluster-Randomized: Groups of users (e.g., social graph clusters) are randomized. (LinkedIn, Facebook for network effects).
Geo/Region-Split: Randomization by geographic area, often using synthetic controls for analysis. (Lyft, Netflix for marketing campaigns).
Session-Split: Randomization at the session level. (Lyft).
Hardware-Split: For testing hardware like e-bikes. (Lyft).
Targeting and Segmentation:
Experiments are often targeted to specific user segments (e.g., by country, language, OS, user tenure).
Airbnb (ERF): Supports “dimensional cuts” (slicing metrics by user/event attributes), precomputing many dimensions. ERF Explorer allows interactive analysis.
Lix DSL (LinkedIn): Enables complex segmentation rules.
Holdouts:
Universal/Global Holdouts: A small percentage of users are held back from all (or a broad category of) product changes for an extended period (e.g., a quarter). Used to measure the cumulative, long-term impact of multiple shipped features. (Disney Streaming/Hulu, Spotify, Uber).
Disney/Hulu: Uses 3-month enrollment + 1-month evaluation. Found cumulative effects smaller than summed individual effects due to cannibalization and novelty wearing off.
Spotify: Uses quarterly holdbacks within “domains.”
Reusable Holdout (Google): A method to preserve statistical validity in adaptive data analysis, allowing multiple hypotheses to be tested on the same holdout data by carefully managing the “information budget.”
Overlapping Experiments & Layering (Google):
A system allowing multiple experiments to run on overlapping user populations by ensuring that parameters being modified by different experiments are orthogonal. Experiments are assigned to “layers,” and users can be in multiple experiments if they are in different layers. Traffic is divided among layers and then within layers among experiments.
Sequential Testing:
Allows continuous monitoring and early stopping if an effect (positive or negative) is detected, while controlling false positive rates.
Uber: Uses mixture Sequential Probability Ratio Test (mSPRT) for outage detection and monitoring key metrics, with jackknife/bootstrap for variance estimation under correlated data and FDR control for multiple metrics.
Spotify: Offers sequential testing in its platform; if chosen, results are available as the experiment runs.
Stitch Fix (Optimal Testing): Explores minimizing time-to-discovery in an “experiment-rich regime” using sequential testing and Bayesian methods, characterizing optimal rejection/continuation thresholds.
Interleaving (Airbnb):
For search ranking, interleaving blends results from two rankers (control and treatment) and presents them to the same user. User interactions (clicks, bookings) determine preference.
Uses “team drafting” to ensure fairness and “competitive pairs” to focus on real differences, improving sensitivity.
Complex attribution logic is needed due to multi-search-to-booking user journeys.
Achieved 50x speedup over A/B tests for ranking, with 82% consistency. Not suitable for rankers with set-level optimization.
Time-Lagged Conversions (Better - Convoys library):
Models conversions that don’t happen immediately using survival analysis (Kaplan-Meier for non-parametric estimation) and parametric models (Weibull, Gamma distributions) to predict eventual conversion rates and times.
IV. Statistical Analysis & Metrics
Core Statistical Concepts:
Netflix’s Blog Series: Provides excellent, intuitive explanations of:
False Positives (Type I error), Statistical Significance, p-values, Rejection Regions, Confidence Intervals. Conventionally, α = 0.05.
False Negatives (Type II error), Statistical Power (1-β). Conventionally, power = 80%. Power depends on effect size, sample size, and underlying metric variability.
These concepts are fundamental to experiment design (sample size calculation) and result interpretation.
Variance Reduction Techniques: Crucial for increasing experiment sensitivity (power) and reducing required sample sizes/durations.
CUPED (Controlled-experiment Using Pre-Experiment Data): Uses pre-experiment data (e.g., the same metric from a prior period) as a covariate to reduce variance in the outcome metric. Widely adopted (Microsoft, DoorDash, Netflix, LinkedIn, Booking.com, Spotify).
DoorDash (CUPAC - Control Using Predictions As Covariates): Extends CUPED by using ML model predictions (based on pre-experiment features uncorrelated with treatment) as the covariate. Showed significant power improvements (e.g., ~40% reduction in test length for ASAP metric).
Dropbox (ML XR): Uses ML-predicted Expected Revenue (XR) as a short-term proxy for long-term LTV, accelerating A/B test analysis. XR lift is used, and a ~3% systematic uncertainty from the model is accounted for.
Regression Adjustment: More general form of covariate control. Facebook uses a cluster-based regression adjustment for network experiments.
Trigger Analysis / Overtracking Correction (Booking.com, Airbnb):
Overtracking: Including non-treatable users in an experiment dilutes the treatment effect and increases variance. If an experiment overtracks by a factor
k
(k untreatable users for every 1 treatable), the required sample size to detect the same effect on the treatable population increases byk² * (σ²_o / σ²_t)
.Trigger Analysis: Analyzing only users who were actually exposed to the differentiated experience. This effectively reduces
k
.Bot traffic can be a form of overtracking.
Metric Types & Calculation:
Airbnb (ERF): Defines metrics in a common configuration language.
DoorDash (Curie/Dash-AB): Supports continuous, proportional, and ratio metrics. Uses linear models, bootstrapping, and Delta method for p-value/SE.
Uber (XP): Categorizes metrics as proportion, continuous, and ratio. Uses Welch’s t-test, Mann-Whitney U, Chi-squared, Delta method, bootstrap.
Netflix (XP): Supports high-performance bootstrap for compressed data to estimate quantile treatment effects.
Spotify: Developed a computationally efficient Poisson bootstrap method for comparing quantiles at scale, approximating the distribution of ordered indexes selected in bootstrap samples with a Binomial distribution.
Heterogeneous Treatment Effects (HTE) / Uplift Modeling:
DoorDash: Uses HTE models (S-learner, T-learner, X-learner meta-learners with LightGBM as base) to personalize promotions for churned customers, identifying subpopulations with positive reactions to reduce costs.
Uber: Emphasizes going “Beyond Average Treatment Effects” by calculating Quantile Treatment Effects (QTEs) using quantile regression. This helps understand impact on different parts of the outcome distribution (e.g., p95 ETA).
Handling Multiple Comparisons:
Lyft: Uses Benjamini-Hochberg for Multiple Hypothesis Testing (MHT) correction when multiple primary metrics are used.
Uber: Applies BH correction for FDR control when monitoring multiple metrics in sequential tests.
V. Addressing Specific Experimentation Challenges
Interference / Network Effects: When treatment of one unit affects outcomes of another.
LinkedIn: Developed a method to detect interference by running an “A/B test of A/B tests.” They cluster the social graph and split clusters into two arms: one with individual-level randomization, the other with cluster-level randomization. A significant difference in estimated effects (using a Hausman-like test) indicates interference. Used reLDG for clustering.
Facebook (Network Experimentation at Scale): Deploys cluster-randomized experiments, often using graph clustering (e.g., Louvain). Introduces cluster-based regression adjustment to improve precision, finding imbalanced clusters can be superior. Leverages “trigger logging” (logging exposure only when a unit calls the service and receives a treatment assignment) for further variance reduction.
Lyft: Uses time-split tests and region-split tests (with synthetic controls) to manage network effects.
DoorDash: Relies heavily on switchback tests (region/time window randomization) for logistics experiments due to strong network effects.
Resource Constraints / Two-Sided Marketplaces:
Stitch Fix (Virtual Warehouse): For inventory-constrained experiments, they virtually partition inventory between variants (e.g., 50 shirts for variant A, 50 for B out of 100 total). This prevents one variant from depleting resources available to the other, ensuring valid estimation of what would happen if a variant were rolled out. Also used to test different inventory strategies.
Uber/Lyft/DoorDash: Two-sided marketplaces inherently have interference. Switchback/time-split/geo-split designs are common.
Small Sample Sizes / Low Power:
Netflix (Quasi Experiments): For non-member experiments (e.g., TV ad impact) with small numbers of geo units, they use re-randomization for balance, Difference-in-Differences (DID), and Dynamic Linear Models (DLMs, similar to Google’s CausalImpact) for time-series analysis, especially with multiple on/off interventions. For member-focused tests without historical outcomes (e.g., new show promotion), they use rich member data with propensity score matching and regression adjustment.
Variance reduction techniques (CUPED/CUPAC) are key.
Scalability and Performance of Experimentation Engine:
LinkedIn (Lix Engine Rewrite): Rewrote their Clojure-based engine in Java. Optimized evaluation tree data structures (plain arrays vs. node objects) for lower memory and faster traversal. Used auto-generated stubs for type resolution instead of reflection. Implemented short-circuiting for remote calls. Resulted in ~20x speedup.
Airbnb (ERF Scaling): Moved from monolithic Hive queries to Airflow DAGs, computing by “Event Source.”
Twitter (DDG Pipeline Scaling): Employed Hadoop MapReduce optimizations like
RawComparator
for Thrift objects (OrderedSerialization
in Scalding) for ~30% compute time saving.
Experiment Velocity & Capacity:
DoorDash: Increased consumer-facing experiment capacity 4x by using “Dynamic Values” infrastructure for parallel, mutually exclusive experiments. Quadrupled experiment capacity again by improving sensitivity with CUPED. Increased logistics experiment capacity 1000% via a fail-fast culture, scalable processes (standardized metrics, two-zone system for interaction testing, weekly cadence), optimal statistical methods (CUPAC, switchback optimization), and platform automation (Curie).
Spotify: New platform aims to allow autonomous teams to start/stop experiments anytime. “Salt machine” helps manage allocations.
Zalando: Evolved platform to handle more concurrent tests, focusing on scalability and trustworthiness.
Long-Term Effects & Cumulative Impact:
Universal Holdouts: (Disney, Spotify, Uber) - A small user segment is held out from all (or a category of) new features for months/quarters. At the end, a single experiment compares this holdout group to users who received all shipped features, measuring the true cumulative impact. Disney found this crucial as summed individual experiment uplifts often overestimated actual cumulative impact due to cannibalization or novelty effects wearing off.
Airbnb (Future Incremental Value - FIV): Uses Propensity Score Matching (PSM) to estimate the long-term (1-year) causal effect of actions (e.g., a booking) on outcomes like revenue. This addresses the short duration of typical A/B tests and selection bias. The platform ingests client configs defining focal/complement groups and automates the PSM pipeline.
Experiment Guardrails (Airbnb):
A system to prevent negative impacts on key metrics. Defines three types of guardrails:
Impact Guardrail: Escalates if global average treatment effect is more negative than a threshold
t
.Power Guardrail: Ensures standard error is small enough (
StdErr < 0.8 * t
) for reasonable FPR and power. Thresholdt
adjusts with global coverage (t = T / sqrt(coverage)
).Stat Sig Negative Guardrail: Escalates for any statistically significant negative impact on critical metrics.
Flags ~25 experiments/month; 20% are stopped.
Quasi-Experiments & Causal Inference (when A/B is not feasible):
Netflix: Uses DID, synthetic controls, DLMs for geo-based marketing campaigns or when randomization is impossible. Uses propensity scores and doubly robust estimators for observational studies.
Airbnb (ACE - Artificial Counterfactual Estimation): ML-based causal inference for non-randomizable scenarios. Trains an ML model on a holdout (non-treated) group to predict counterfactual outcomes for the treated group. Addresses ML model bias (from regularization/overfitting) using A/A tests for debiasing and constructing empirical CIs.
Shopify: Mentions using quasi-experiments and counterfactuals. (Article link was 404, so details are missing).
Facebook (Constrained Bayesian Optimization): Optimizes system parameters (e.g., ranking, compiler flags) via noisy A/B tests where direct evaluation is expensive. Uses Gaussian Processes (GPs) and Expected Improvement (EI). Derives Noisy Expected Improvement (NEI) integrating over posterior of true function values, optimized with quasi-Monte Carlo (QMC) for efficiency. Handles noisy constraints.
Uber (Bayesian Optimal Experimental Design - OED with Pyro): Focuses on designing experiments to maximize information gain, especially when evaluations are costly. Uses variational inference within the Pyro probabilistic programming language.
Multi-Armed Bandits (MABs):
Stitch Fix: Integrated MABs into their platform. Data scientists implement reward models as microservices. Platform handles policies (ε-greedy, Thompson Sampling, custom proportional policies) and deterministic allocation. Supports contextual bandits.
Uber: Uses MABs for content optimization (e.g., email subject lines) and hyperparameter tuning for ML models (e.g., Uber Eats feed ranking) using contextual MABs and Bayesian optimization.
Traveloka & Spotify: Mention MABs as part of their advanced experimentation capabilities.
Experimentation on External Platforms (e.g., Google Ads - DoorDash):
Leverages Google Adwords’ “campaign drafts and experiments” for A/B testing bidding strategies.
Success metric: Cost-Per-Acquisition (CPA) at comparable spend levels.
Challenge: Calculating standard deviation for CPA (a ratio metric with unknown distribution) with limited weekly historical data. Solution: Bootstrapping (randomly sampling weekdays to form new weekly cohorts) to enlarge dataset and empirically estimate std dev.
Runs A/A tests for campaign warm-up/stabilization.
VI. Culture & Process
Democratization & Standardization:
Netflix: XP is democratized, allowing data scientists to contribute metrics, methods, and visualizations. Broad accessibility of results.
DoorDash (Curie & Dash-AB): Standardization of analysis and metrics improves quality and shareability.
Spotify (Search Team Journey): Emphasized a roadmap (individual experiment quality -> cross-experiment quality -> total business impact measurement) and “constant injection of energy” (support, templates, champions) to build culture.
Zalando: Company-wide training, embedded A/B test owners, peer reviews for causal inference research.
Decision Making & Governance:
Airbnb (Experiment Guardrails): Formal process for escalating experiments with potential negative impact.
Spotify: Experiment Planner requires specifying success/guardrail metrics, MDE, and test type (one/two-sided, superiority/non-inferiority) upfront.
Lyft: Guided hypothesis workflow to pre-register hypotheses and primary/guardrail metrics. Results and decision tracking to align on investment trade-offs.
Twitter (Experimenting to solve Cramming): For the 280-char limit, ran multiple experiments grouped by countries with similar “cramming” behavior to reduce within-group variance and improve power. Used A/A tests to check for bucketing bias.
Fail-Fast, Learn-Fast:
DoorDash: Explicitly fosters a culture of failing fast to learn fast, operating at the lowest level of detail, and intellectual honesty.
This is a common theme, as rapid iteration is a primary goal of experimentation.
Quality Assurance:
Pinterest: Implemented a review tool and helper group for experiment changes. PR links required for code changes. Test-only copies of experiments.
Spotify: Automated validity checks: Sample Ratio Mismatch (SRM), pre-exposure activity differences, crash increases, property collisions.
DoorDash (Dash-AB): Imbalance tests (Chi-square for SRM), flicker tests.
VII. Future Directions & Open Challenges (as highlighted by Netflix & Facebook)
Computational Causal Inference (CompCI - Netflix):
Developing performant, general, and robust software for causal inference.
Optimizing for sparse data, efficient counterfactual matrix creation, vectorization for multiple KPIs/treatments, efficient bootstrap (e.g., bag of little bootstraps).
Structuring software to enable detection of causal identification.
Handling cyclic causal graphs, autocorrelation in time-series data, marginal treatment effects with non-randomized subsequent treatments, and time-varying treatment availability.
Network Experimentation (Facebook):
Improving cluster design for bias-variance trade-off.
Developing model-based alternatives to agnostic design-based analysis for network experiments.
Adaptive Experimentation & OED (Uber, Facebook):
Moving towards more adaptive experiments (e.g., contextual bandits, reinforcement learning) to dynamically optimize and respond to changing conditions.
Bayesian Optimal Experimental Design to choose designs that maximize information gain, especially when evaluations are costly.
This synthesis shows a clear trend towards more sophisticated, scalable, and automated experimentation platforms. While A/B testing remains central, there’s significant work in variance reduction, handling interference, long-term/cumulative effects, causal inference beyond A/B, and integrating ML for both analysis acceleration (e.g., proxy metrics) and direct optimization (e.g., MABs). A strong data-driven culture, supported by robust processes and education, is equally vital for success.