66
NBER WORKING PAPER SERIES DESIGN AND ANALYSIS OF CLUSTER-RANDOMIZED FIELD EXPERIMENTS IN PANEL DATA SETTINGS Bharat K. Chandar Ali Hortaçsu John A. List Ian Muir Jeffrey M. Wooldridge Working Paper 26389 http://www.nber.org/papers/w26389 NATIONAL BUREAU OF ECONOMIC RESEARCH 1050 Massachusetts Avenue Cambridge, MA 02138 October 2019 The views expressed herein are those of the authors and do not necessarily reflect the views of the National Bureau of Economic Research. At least one co-author has disclosed a financial relationship of potential relevance for this research. Further information is available online at http://www.nber.org/papers/w26389.ack NBER working papers are circulated for discussion and comment purposes. They have not been peer-reviewed or been subject to the review by the NBER Board of Directors that accompanies official NBER publications. © 2019 by Bharat K. Chandar, Ali Hortaçsu, John A. List, Ian Muir, and Jeffrey M. Wooldridge. All rights reserved. Short sections of text, not to exceed two paragraphs, may be quoted without explicit permission provided that full credit, including © notice, is given to the source.

Design and Analysis of Cluster-Randomized Field

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Design and Analysis of Cluster-Randomized Field

NBER WORKING PAPER SERIES

DESIGN AND ANALYSIS OF CLUSTER-RANDOMIZED FIELD EXPERIMENTSIN PANEL DATA SETTINGS

Bharat K. ChandarAli HortaçsuJohn A. List

Ian MuirJeffrey M. Wooldridge

Working Paper 26389http://www.nber.org/papers/w26389

NATIONAL BUREAU OF ECONOMIC RESEARCH1050 Massachusetts Avenue

Cambridge, MA 02138October 2019

The views expressed herein are those of the authors and do not necessarily reflect the views of the National Bureau of Economic Research.

At least one co-author has disclosed a financial relationship of potential relevance for this research. Further information is available online at http://www.nber.org/papers/w26389.ack

NBER working papers are circulated for discussion and comment purposes. They have not been peer-reviewed or been subject to the review by the NBER Board of Directors that accompanies official NBER publications.

© 2019 by Bharat K. Chandar, Ali Hortaçsu, John A. List, Ian Muir, and Jeffrey M. Wooldridge. All rights reserved. Short sections of text, not to exceed two paragraphs, may be quoted without explicit permission provided that full credit, including © notice, is given to the source.

Page 2: Design and Analysis of Cluster-Randomized Field

Design and Analysis of Cluster-Randomized Field Experiments in Panel Data SettingsBharat K. Chandar, Ali Hortaçsu, John A. List, Ian Muir, and Jeffrey M. WooldridgeNBER Working Paper No. 26389October 2019JEL No. C23,C33,C5,C9,C91,C92,C93,D47

ABSTRACT

Field experiments conducted with the village, city, state, region, or even country as the unit of randomization are becoming commonplace in the social sciences. While convenient, subsequent data analysis may be complicated by the constraint on the number of clusters in treatment and control. Through a battery of Monte Carlo simulations, we examine best practices for estimating unit-level treatment effects in cluster-randomized ˝field experiments, particularly in settings that generate short panel data. In most settings we consider, unit-level estimation with unit fixed effects and cluster-level estimation weighted by the number of units per cluster tend to be robust to potentially problematic features in the data while giving greater statistical power. Using insights from our analysis, we evaluate the effect of a unique ˝field experiment: a nationwide tipping field experiment across markets on the Uber app. Beyond the import of showing how tipping affects aggregate market outcomes, we provide several insights on aspects of generating and analyzing cluster-randomized experimental data when there are constraints on the number of experimental units in treatment and control.

Bharat K. ChandarStanford University579 Serra MallStanford, [email protected]

Ali HortaçsuKenneth C. Griffin Department of EconomicsUniversity of Chicago1126 East 59th StreetChicago, IL 60637and [email protected]

John A. ListDepartment of EconomicsUniversity of Chicago1126 East 59thChicago, IL 60637and [email protected]

Ian [email protected]

Jeffrey M. WooldridgeDepartment of Economics, Michigan State [email protected]

Page 3: Design and Analysis of Cluster-Randomized Field

1 Introduction

In the past few decades field experiments in the social sciences, and in particular economics,have grown dramatically in their usage. Today, organizations as varied as firms, governments,and NGOs partner with academics to generate data via field experiments to solve theirproblems. As field experiments have evolved, the unit of randomization has expanded beyondthe individual level. For example, cities, villages, regions, or even countries have been thetarget for interventions of both scientists and practitioners. Understanding the benefits andcosts of design and analysis choices at such levels of randomization is the goal of our study.

This is not entirely original ground. There exists an extensive literature on the analysisof cluster-randomized field experiments and computation of cluster-robust standard errors.Abadie et al. (2017) present clustering as essentially a design problem. They show thatclustering is justified when either a subset of clusters are sampled randomly from a broaderpopulation of clusters or if assignment is correlated within clusters. We focus on the lattercase. In such instances Athey and Imbens (2016) recommend inference based on the clusteras the unit of analysis, arguing that, in practice, including unit-level characteristics gener-ally improves precision by a relatively modest amount compared to including cluster-levelaverages as covariates in a cluster-level analysis. If the unit-level population treatment effectis the main target of interest, they suggest cluster-level inference to supplement unit-levelestimation, arguing that in cases with very different cluster sizes, unit-level estimation maybe noisier.

Neither of these papers, however, explicitly considers panel data settings. A seminalpaper on cluster-randomized treatments in panel data is Bertrand et al. (2003), which showsthrough simulation that not clustering standard errors in panel data can lead to systematicoverrejection of significance tests in the presence of serial correlation. While Bertrand et al.(2003) demonstrate the importance of clustering standard errors or using randomization-based inference, they do not offer recommendations on unit-level versus cluster-level analysisof treatment effects. In big data settings, the number of units observed may greatly exceedthe number of clusters, so such issues merit consideration.

Brewer et al. (2017) argue, in response to Bertrand et al. (2003), that straightforwardmethods can be used to yield a correctly sized test in diff-in-diff settings. They proposeusing cluster-randomized standard errors (CRSE) and, instead of using the standard normalapproximation to compute p-values, using a t-distribution with degrees of freedom equal tothe number of clusters minus one. Through simulation they show that while these methodscan give tests that are correctly sized they lead to underpowered estimates. They recommendusing the bias-corrected FGLS method derived in Hansen (2007) or FGLS with CRSE toimprove power.

Yet, both Brewer et al. (2017) and Hansen (2007) focus on achieving higher power bycorrecting for serial correlation in the errors, an issue that is typically less important inpanels with few time periods. In Monte Carlo simulations, Hansen (2007) shows that serialcorrelation does not pose a serious problem to inference when there are six time periods inthe data. In randomized control trial settings, where measuring outcomes repeatedly mightbe infeasible or too expensive, having many time periods may not be possible. In such cases,

2

Page 4: Design and Analysis of Cluster-Randomized Field

there may be more effective ways to improve the efficiency of estimators than focusing onmodeling serial correlation.

In this study, we combine the panel data literature on cluster-robust standard errors(Moulton (1990), Liang and Zeger (1986), Cameron and Miller (2015)) with the finite pop-ulation causal standard errors literature (Abadie et al. (2017), Abadie, Athey, Imbens, andWooldridge (2014), Athey and Imbens (2016)), focusing on cluster-randomized experimen-tal settings with short panels. Our main contribution is to compare different methods ofestimation—most notably unit-level versus cluster-level estimation of treatment effects—interms of whether they provide correctly sized tests of the treatment effect.

We conduct extensive simulations on both synthetic data constructed under increasinglycomplex data generating processes and on novel experimental data described below. Amongthe factors we consider are the number of clusters, the number of units, the variation incluster size, the level of intracluster correlation, and the number of time periods.

One key insight is that unit-level estimation with unit fixed effects and cluster-levelestimates weighted by the number of observations per cluster can give higher power thanunweighted cluster-level estimates in most settings we consider. Importantly, this result holdswithout sacrificing proper coverage of the treatment effect. This is especially true when thereis treatment effect heterogeneity at the cluster level that is correlated with cluster size. Inshort panels, these simple changes can lead to much larger improvements in precision thanproperly modeling serial correlation.

Another key insight is that from a design perspective, we show that when there areheterogeneous treatment effects at the cluster level, there are important cases when theanalyst should include more treated clusters than control clusters. The intuition is that theadded treated sample can be valuable for averaging over the heterogeneity and recoveringthe population mean effect.

We also evaluate the robustness of computing stratified standard errors, as suggested byAbadie et al. (2014). We find that stratifying standard errors based on observable character-istics that are correlated with the treatment effect can lead to substantial improvements inpower, but when the data-generating process is complex—e.g. with differently sized clustersand complex spatial correlation—it may lead to underestimates of the true standard error.

We put our insights into action by evaluating the effect of introducing tipping on Uber’smarketplace. In July of 2017 we created a unique opportunity to explore this class of issueswhen we helped Uber launch optional in-app tipping. We designed a field experiment across209 operational cities in the United States and Canada where we placed half of the cities in atreatment group, in which tipping started on July 6, 2017, and half of the cities in a controlgroup, where tipping started 10 days later. We use this exogenous variation to evaluatemarket level changes in earnings, labor supply, demand, and other outcomes of interest (ina companion study, Chandar et al. (2019), we explore individual behaviors). Over 800,000drivers were included in the experiment. Units were cluster-randomized because of concernsabout interactions between drivers within a city and because of business concerns aboutrandomly excluding access across drivers within a city to a popular new feature.

While the main results are not statistically significant at conventional levels, we find im-

3

Page 5: Design and Analysis of Cluster-Randomized Field

provements in precision from adopting the methods described. Point estimates are imprecisebut suggest a short-run increase in labor supply, decrease in demand, fall in utilization, andfall in hourly earnings for drivers. We view our methodological work as the main contributionof our study, but exploring the tipping data at the market level does provide new insightsinto the tipping literature, which to date has not experimentally examined the questionsthat we explore.

The remainder of our paper proceeds as follows. Section 2 describes the simulation set-ting. Section 3 shows simulation results under different data generating processes increasingin complexity to mimic the Uber data. Section 4 considers simplified simulation settingsthat seek to isolate the effect of various potential features of the data. Section 5 presentsresults and placebo test simulations in the real data for labor supply. Section 6 summarizesempirical results for the full set of market outcomes. Section 7 concludes.

2 Simulation Set Up

We begin by describing the simulation framework used in the remainder of the paper tocompare different methods of estimation. The motivating example for our simulations is theUber tipping launch. In the Uber tipping case, which we expand on much more patiently inSection 5, the entire population of cities is observed. The uncertainty in the Uber context isnot from sampling from the population, but instead from the treatment assignment for eachcity, which is randomized. In the following simulations we model a scenario where the entirepopulation is observed using the potential outcomes framework.

Let N be the number of individuals in the population. These individuals are split betweena set of cities C, with the city that individual i belongs to denoted ci. Each city c hasNc individuals. Treatment assignment occurs at the city level, so we can separate C intoC = {Ctreated, Ccontrol}, where Ctreated is the set of treated cities and Ccontrol is the set ofcontrol cities. Let wi,t be the treatment status for individual i in time t, where t goes from1 to T . In the Uber example, wi,t is whether an individual is in a city with tipping at timet. Since the experiment is cluster-randomized,

wi,t =

{1 if ci ∈ Ctreated and t ≥ Ttreated

0 otherwise

where Ttreated is the time of the treatment event. We assume that the time of the treatmentevent does not vary by treated city.

Let Yi,t, the outcome for individual i in time t, be

Yi,t =

{Yi,t(1) if wi,t = 1

Yi,t(0) otherwise

Yi,t(1) is the outcome for individual i in time t if they are treated, and Yi,t(0) is the outcomeif they are in the control group. While in the pre-treatment period we observe Yi,t(0) for alli, when t ≥ Ttreated, we observe only Yi,t(0) for some individuals and only Yi,t(1) for others,

4

Page 6: Design and Analysis of Cluster-Randomized Field

depending on the treatment status for their city. If we take the population of interest to beactive drivers in cities across the United States and Canada, then our application is one withno sampling uncertainty of individuals (since we observe the entire population). However,there is assignment uncertainty since in the post-period we only observe one of their potentialoutcomes.

We run Monte Carlo simulations to evaluate different estimation strategies under thisframework. We first sample Yi,t(0) for all i, t and Yi,t(1) for t ≥ Ttreated from some chosendata generating process (DGP). These vectors of outcomes constitute the fixed population.Then, for each run of the simulation we randomize treatment assignment to the clusters andestimate the treatment effect using each estimation strategy. By computing the estimatedtreatment effect many times, we can estimate separately for each estimation strategy thedistribution of the treatment effect and standard error over the randomizations of treatmentstatus. Over these randomization distributions, we estimate coverage of the treatment effect(the proportion of times the true population treatment effect falls within the 95% confidenceinterval on a run) and the rejection frequency (the proportion of times the null hypothesisof no effect is rejected when the alternative is true) for each estimation method. Note thatthe source of variation for each run of the simulation comes only from the randomization oftreatment status; there is no variation in the potential outcomes of the underlying populationacross runs of the simulation.

The treatment effect in the population is

τp =

∑Ni=1

∑Tt=Ttreated

[Yi,t(1)− Yi,t(0)]

N(T − Ttreated + 1),

that is, the average treatment effect across individuals in the post-treatment period. Notethat the treatment effect depends on the population, which is drawn from some data gener-ating process. While in Section 3 we draw only one fixed population for each set of results,in Section 4 we present error bars that show how coverage and rejection frequencies varydepending on the population that is drawn.

Across the simulations we vary several dimensions: the number of time periods, whetherthere is cluster heterogeneity in outcomes and treatment effects, whether there are individual-level intercepts, and the complexity of the spatial correlation within clusters, in addition toother features of the data-generating process. We also consider cases where the treatmenteffect is correlated with observable characteristics. Exact parameter values for the data-generating process for each population are presented in Appendix A.

3 Simulation Results

Using the above potential outcomes framework we run Monte Carlo simulations to examinethe performance of various models for identifying treatment effects in settings with clusterrandomization. In the results that follow we iteratively add complexity to the data generatingprocess, starting with a simple one-period model. In all of the examples in this section thereare 20,000 individuals across 212 clusters.

5

Page 7: Design and Analysis of Cluster-Randomized Field

3.1 One-Period Model, Homogeneous Effect

We first consider one-period models with a homogeneous treatment effect. Since there isonly one period we drop the time subscript t. Treatment status for individual i is

wi =

{1 if ci ∈ Ctreated

0 otherwise

The data generating process is

[DGP ] : Yi = τwi + εi

τ is the average treatment effect, and εi ∼ N(0, σ2) is an error term. We first consider a casein which every city has the same number of observations. In these simulations there is noheterogeneity in the treatment effect across cities. We relax this assumption in subsequentmodels. Exact parameter values are in Table 27 in the appendix.

We estimate the regression model

Yi = τwi + εi

across individuals. In this case, standard errors give proper coverage under the null hy-pothesis that τ = 0, regardless of whether the regression is clustered at the city level usingLiang-Zeger (LZ) clustered standard errors (Liang and Zeger, 1986). Table 1 shows the re-sults of a Monte Carlo simulation under this data generating process when each city is thesame size. Specifically, for each city c the number of individuals is set to N

|C| , rounded to thenearest integer.

In order to compare the individual-level regressions to ones collapsed to the city-level,we also consider regressions of the form:

Yc = τwc + εc

with

Yc =1

Nc

∑i:ci=c

Yi, εc =1

Nc

∑i:ci=c

εi

and wc the treatment status for cluster c.Row 3 in Table 1 show the results from the city-level regression. In this case there is little

difference between the various methods on efficiency of the estimator. Results are similarunder the alternative hypothesis, as shown in Table 2.

3.2 One-Period Model: Homogeneous Treatment Effect With Het-erogeneous Cluster Sizes

When clusters vary in size, it becomes necessary to weight the cluster-level regression by thenumber of observations in each cluster to maintain the same power as in the individual-levelregression. For Tables 3 and 4 we vary the size of the clusters, with sizes roughly calibrated

6

Page 8: Design and Analysis of Cluster-Randomized Field

to the sizes of cities in the Uber field experiment. Let N∗c be the number of observations incity c in the real data. In the simulated data, we set Nc = round(N × N∗c∑

c′N∗

c′), where we set

N to 20,000.Under both the null and alternative hypotheses, the unweighted city-level regression has

larger standard errors. This is because errors are heteroskedastic at the city level since theyare averages across individuals. Running weighted least squares with weights equal to clustersizes gives increased efficiency and higher rejection rates when the alternative is true.

3.3 One-Period Model: Cluster Intercepts

We next consider a one-period model with cluster-specific intercepts. The data-generatingprocess is

[DGP ] : Yi = τwi + αci + εi

where αci ∼ N(0, σc) is a cluster-specific intercept. We again estimate the regression models

Yi = τwi + εi

andYc = τwc + εc

Tables 5 and 6 show results from simulation when clusters are equal sizes. Individual-level regressions that are not clustered underestimate the standard error. This is consistentwith the findings of Abadie et al. (2017), who show that if clusters vary in mean outcomes,assignment is clustered, and fixed effects are not included in estimation then standard errorsshould be clustered. The clustered individual-level model and the city-level model givecomparably small standard errors.

When clusters vary in size, as in Tables 7 and 8, the unweighted city-level model hashigher variance in the estimated coefficient than the clustered individual-level model. UsingEicker-Huber-White (White, 2014) standard errors in the city-level model does not havea considerable effect on the estimated standard errors. The city-level model weighted bysize reduces the variance of the estimator to be comparable to the individual-level models,but the standard error is too small, leading to poor coverage of the true treatment effect.Weighting the city-level model by the number of observations and also using EHW standarderrors leads to similar performance as the individual-level model with clustering. Weightedleast squares is insufficient unless also using robust standard errors because weighting bythe size of the cluster ignores the random cluster-specific intercept, which contributes to thevariance in the outcome. In settings where clusters vary in size, EHW estimation can leadto conservative standard errors and WLS can lead to standard errors that are too small.Combining the two leads to tests that have better size.

Notwithstanding, the unweighted city-level estimator maintains better coverage of thetreatment effect than the other models. This result echoes the recommendations of Atheyand Imbens (2016) for a simple one-period model with differing cluster sizes. In subsequent

7

Page 9: Design and Analysis of Cluster-Randomized Field

results, we show that the individual-level estimator and the weighted city-level estimator mayperform better when having multiple time periods because it allows us to include cluster-fixedeffects in the models.

3.4 Two-Period Model: City and Time Fixed Effects

In the Uber tipping context, our goal is to estimate the treatment effect across individualsin a multi-period setting where cities only become treated after the launch of the product.In order to simplify the model while still keeping it informative, we consider a two periodmodel.

Let Yi,t be the outcome for individual i in time period t. Treatment status for individuali at time t is now

wi,t =

{1 if ci ∈ Ct and t = 2

0 otherwise

Individuals are only treated if they are in a treated city and t = 2.The first DGP we consider is

[DGP ] : Yi,t = τiwi,t + αci + βt + εi,t

where τi ∼ N(τ, σ2τ ) is an individual-specific treatment effect, βt ∼ N(0, σ2

t ) is a time effect,and εi,t ∼ N(0, σ2

ε ) is an independent and identically distributed error term. Cluster sizesare calibrated to the Uber data. The treatment effect now varies by individual under thealternative hypothesis, but under the null τi = 0 for all i.

We consider individual-level regressions of the form

Yi,t = τwi,t + αci + βt + εi,t

and city-level regressions of the form

Yc,t = τwc,t + αc + βt + εc,t

Because we have repeated observations for each cluster over time we can now use fixed effectsestimation to account for cluster intercepts. Results are in Tables 9 and 10.

Clustering the standard errors without also including cluster fixed effects leads to un-derestimates of the standard errors for the individual-level models and the weighted citymodels. Clustering alone is insufficient to account for the within-cluster correlation betweenindividuals when there are only around 200 clusters. For these models, standard errors areboth smaller and more properly sized when fixed effects are included. This shows that in-cluding cluster fixed effects can improve both the coverage and power of the treatment effectestimator even when using Liang-Zeger clustered standard errors. We do not see the sameresult for the unweighted model. While for the individual-level models (and similarly forthe weighted city-level models) LZ clustered standard errors need to account for both cor-relation across individuals and across time, for the unweighted city-level models clustering

8

Page 10: Design and Analysis of Cluster-Randomized Field

only needs to correct for correlation across the two time periods. For this simple two-periodmodel, clustering the standard errors is adequate for accounting for such correlation. Thefixed effects drop too many degrees of freedom, leading to higher variance in the estimatedtreatment effect.

The individual-level model and the weighted city-level model with cluster effects havemuch smaller variance in the estimated treatment effect than the unweighted city-level model,consistent with what we found in the one-period case. However, in contrast to our earlieranalysis, the clustered weighted city-level model now also gives proper coverage of the treat-ment effect after we include cluster effects, meaning it gives higher power than the unweightedmodel without sacrificing the size of the test.

3.5 Two-Period Model: Individual Fixed Effects and Heteroge-neous Cluster Effects

Next we add individual effects and heterogeneous treatment effects at the cluster level. TheDGP is

[DGP ] : Yi,t = τiwi,t + αci + βt + hciwi,t + γi + εi,t

where hci ∼ Normal(0, σh) is cluster-level heterogeneity in the treatment effect and γi ∼Normal(0, σγ) is an individual-specific intercept term. Under the null we set hci = 0 for alli. We again consider individual-level regressions of the form

Yi,t = τwi,t + αci + βt + εi,t

and city-level regressions of the form

Yc,t = τwc,t + αc + βt + εc,t

The additional terms in the model increase the variance of the estimated treatment effect.In order to display reasonably large rejection frequencies under the alternative hypothesis,we increase τ to 0.1 from 0.035. Results are in Tables 11 and 12. Under the null for thismodel, in the individual-level model clustering by city does not improve estimation when cityfixed effects are included. Under the alternative hypothesis, when there are heterogeneoustreatment effects at the cluster level clustering becomes necessary. The mean standard errorfor the individual-level model without clustering in Table 12 underestimates the standarddeviation of the estimated treatment effect.

Including individual fixed effects instead of city fixed effects leads to a substantial in-crease in the mean standard error because of reduced degrees of freedom, making the testtoo conservative. Of all the models that give properly sized standard errors, the clusteredindividual-level model with city fixed effects has the highest rejection frequency.

9

Page 11: Design and Analysis of Cluster-Randomized Field

3.6 Two-Period Model: City and Time Fixed Effects With Com-plex Correlation

We now consider situations where the errors are additionally correlated within cities for eachtime period. Instead of an identity matrix, the covariance matrix of εi,t is now block-diagonal,with each block corresponding to a city in a given time period. The covariance matrix foreach city at time t, Σc,t, has the following form:

σ2 σc12 . . . σc1Nc

σc21 σ2 . . . σc2Nc

. . . . . . . . . . . . . . . . . . . . .

. . . . . . . . . . . . . . . . . . . . .σcNc1 σcNc2 . . . σ2

For simplicity we again assume that the diagonal elements σ2 of the error covariance

matrix are constant.The goal is to model a complex, a priori unknown, covariance structure in the errors.

For example, in the Uber example it might be that drivers in a city who work in the sameareas during the same times of the week have correlated outcomes. Simply including cityfixed effects would not correctly model this aspect in the data. To simulate this setting weneed a dense covariance matrix with the correlation between drivers increasing with their(geographic or temporal) proximity. We draw the vector of errors for each city in each timeperiod, ~εc,t, from a Markov chain to generate this sort of covariance structure. The samplingalgorithm, shown in Algorithm 1, resembles an AR(1) process.

Algorithm 1: Sampling algorithm to generate a dense covariance matrix for each cluster.

Initialize u(1);for d = 1, ..., D do

- Sample u(d+1) ∼ f(u(d+1)|u(d));end

Collapse u(1), ..., u(D) to a vector U ;Set ~εc,t to the last Nc elements of U

In Algorithm 1, u(d) and u(d+1) may represent errors for drivers who work in very similarareas. We assume the joint distribution of u(d) and u(d+1), f(u(d), u(d+1)), is bivariate Normalwith diagonal elements equal to σ2 and off diagonal elements equal to ρ. We set ~εc,t to onlythe last Nc elements of the vector U , dropping the first few draws from the chain, so thatthe variance stabilizes to a constant. As ρ approaches 1 all of the error terms within a cityfor a given time period approach equivalence, and if ρ = 0 then all the error terms will beindependent.

To summarize, the DGP is

[DGP ] : Yi,t = τiwi,t+αci + βt + hciwi,t + γi + εi,t

εi,t ∼ N(0,Σ) for block-diagonal Σ(1)

10

Page 12: Design and Analysis of Cluster-Randomized Field

The estimated models are the same as in our earlier analysis. Results in simulation whenρ is set to 0.5 are in Tables 13 and 14. Under the null it becomes necessary to clusterthe standard errors when there is unmodeled correlation in the errors. The individual-levelmodel with individual fixed effects again gives a conservative test relative to the model withcity fixed effects.

The standard errors for the weighted city-level model are too small in this case unlessclustered by city. Again, the unweighted city-level model has larger standard errors than theother models.

3.7 Heterogeneity from Observables

To model situations where the treatment effect varies with observable driver characteristics,we consider cases with an interaction between observable characteristics zi and the treatmentindicator:

[DGP ] : Yi,t =τiwi,t + ci + t+ hcwi,t + τzwi,t × zi + αi + εi,t

εi,t ∼ N(0,Σ) for block-diagonal Σ

This format is similar to the one used in Abadie et al. (2014). They model the potentialoutcomes as

Yi(x) = Yi(0) + (ξ0 + ξ1Zi + ξ2ηi)× x,

where ξ1 and ξ2 are varied across simulations and ξ0 is set to 0. ηi ∈ {−1, 1} is an unobservedsource of heterogeneity in the treatment effect, x ∈ {−1, 1} is the treatment status, andZ ∈ {−1, 1} is a binary attribute of the individual. When a non-negligible proportion ofobservations in the full population are observed and ξ2 6= 0, they show standard errorsstratified by Z are less conservative and closer to the true variance of θ than EHW standarderrors are when estimating the model

Yi = β0 + β1Zi + θXi + εi

The potential outcomes for the DGP in our setting can be expressed as

Yi,t(1) = Yi,t(0) + (τ + ηi + hci + τzzi)× wi,t

where ηi ≡ τi − τ is a source of unobserved heterogeneity at the individual level and hc is asource of unobserved heterogeneity at the cluster level. zi is the source of heterogeneity inthe treatment effect determined by observables.

In simulation, we set the observable attribute zi to the quintile of the cluster size, sub-tracting 3 so that z is centered around 0 across clusters. Ties are broken arbitrarily so thatan equal number of clusters fall in each quintile, even if clusters are identically sized. Since inour simulations zi is the cluster size quintile there is no intracluster variation in z. Collapsingto city-level means that the potential outcomes are

11

Page 13: Design and Analysis of Cluster-Randomized Field

Yc,t(1) =Yc,t(0) + (τ + ηc + hc + τzzc)× wc,t≈ Yc,t(1) = Yc,t(0) + (τ + hc + τzzc)× wc,t

which closely resemble the outcomes in Abadie et al. (2014). The models we fit are again

Yi,t = τwi,t + αci + βt + εi,t

for individuals and

Yc,t = τwc,t + αc + βt + εc,t

when collapsing to city means.Simulation results are summarized in Table 15 for a case where there is heterogeneity

in the treatment effect and clusters are equal sizes. Stratified standard errors are lower onaverage than unstratified standard errors across all of the models. The individual-level modelwith city fixed effects has the smallest reduction in standard errors. Table 16 shows resultsare similar when clusters vary in size. Because under the modeling process large clustershave a larger treatment effect, the mean treatment effect across individuals is greater in thecase with unequal clusters.

The city-level models that are not weighted are biased estimates of the individual-levelmean treatment effect when clusters are different sizes. Since they weight each cluster equally,they overweight small clusters, which have small treatment effects, and underweight largeclusters, which have large treatment effects. The unweighted model estimates a differentquantity, treating each city as the unit of interest instead of each individual. The individual-level models and the weighted city-level models give unbiased estimates of the individual-leveltreatment effect and have comparable standard errors.

12

Page 14: Design and Analysis of Cluster-Randomized Field

To isolate the effect of the heterogeneity from observables, we plot the ratio between thestratified and unstratified standard errors as we vary τz. Figure 1 shows the pattern whenclusters are of equal sizes. All estimates are clustered at the city level. The two plots in thetop panel show that both the stratified and unstratified models become too conservative asthe amount of heterogeneity increases, but the plot in the bottom-left panel shows that thestratified model does increasingly better than the unstratified model. Results are similar inFigure 2, where cluster sizes are roughly proportional to Uber cities.

4 Isolating Effects

What causes differences in standard errors across the different regression specifications?In the following simulation results we show the effect of varying the number of clusters,the number of observations per cluster, the amount of dispersion in cluster size, the levelof intracluster correlation, and the proportions of treatment and control cities. We alsoconsider models with more than two time periods and serial correlation across those timeperiods.

Based on the results in the previous section and our review of the literature, we onlyconsider cases where it is most evident that clustering is necessary to gain proper coverage.We examine cases with city heterogeneity in the treatment effect and complex, unmodeled,spatial correlation in the errors within a cluster. For the following figures, the DGP is

[DGP ] : Yi,t = τiwi,t + αci + βt + hciwi,t + γi + εi,t

εi,t ∼ N(0,Σ) for block-diagonal Σ

The models labeled ”Cit. Clustering” are collapsed to the city level and clustered by city.They include time and city fixed effects. The models labeled ”Cit. Clustering Weighted”are similar, but weighted by the number of individuals in each cluster. Models marked ”Ind.Clustering” are at the individual level, include city and time fixed effects and are clustered bycity. The models labeled ”Ind. Ind FE City Clus.” are at the individual level with individualand time fixed effects, and are clustered by city. These four were chosen because they werethe most robust specifications from the previous simulations.

We report coverage of the treatment effect and the probability of rejecting the null hy-pothesis in the following figures. Coverage and rejection rate results in the trend lines arethe average over estimates from 25 different populations drawn from the DGP. For eachpopulation, coverage and rejection rate are computed over 200 different re-randomizationsof treatment status. We also show error bars that span the minimum to maximum coverageand rejection rate estimates across the 25 different populations to show how much the resultsvary depending on the population.

4.1 Varying the Number of Observations and Clusters

We start with cases where all clusters are the same size and there are two time periods.First, we vary the number of observations, holding the number of clusters fixed at 200. The

13

Page 15: Design and Analysis of Cluster-Randomized Field

number of observations per cluster varies, but in each specification of the DGP the clustersare equally sized. Results are shown in Figure 3. The clustered city-level models and theindividual-level model with individual fixed effects are conservative. All have coverage above95% for each number of observations. The individual-level model with city fixed effects hashigher power and has coverage roughly centered around 95%.

In Figure 4 we see that results are similar when increasing the number of clusters, holdingfixed the number of total observations at 20,000. When there are few clusters, the individual-level model with city fixed effects has coverage of the treatment effect below 95%. It does notexplicitly model idiosyncrasies across individuals, relying on cluster robust standard errorsto account for this source of variation in the data-generating process. Without sufficientlymany clusters, clustered standard errors may not suffice. All of the other models maintaincoverage above 95%. Power increases with the number of clusters across the specifications.

4.2 Varying the Amount of Cluster Dispersion

Next we vary the amount of cluster dispersion. Let Nc be the mean cluster size when thereare 20,000 observations and 212 clusters, or each cluster’s size if observations were equallyspread across clusters. With N∗c set to the number of observations in city c in the real dataand N set to 20,000, again let Nc = round(N × N∗c∑

c′N∗

c′).

We set the number of observations in cluster c to

Nc,δ = ceiling((1− δ)Nc + δNc), (2)

varying δ from 0 to 1. When δ = 0 the clusters are all the same size, while when δ = 1the clusters match the calibrated Uber data case. All specifications except the individual-level model with city fixed effects maintain proper coverage of the treatment effect as thevariation in cluster size increases. The individual-level model with individual fixed effectsand the clustered and weighted city-level model have higher average rejection rates than theunweighted city-level model when the variation in cluster sizes is large. This suggests thatinformation about the number of individuals in each cluster can improve precision.

4.3 Varying the Amount of Spatial Correlation

In Figures 6 and 7 we vary the amount of spatial correlation in the errors. In Figure 6clusters are the same size. The amount of spatial correlation does not seem to have ameaningful effect on the city-level models or the individual-level model with individual fixedeffects when clusters are equal. The individual-level model with city fixed effects has reducedcoverage when the spatial correlation is very high, even when clusters are equal sizes. Powerdeclines as the correlation increases for all specifications. When clusters are different sizes,all models except the unweighted city-level model witness a small decline in coverage whenthe correlation parameter is close to 1. The unweighted city-level model is dominated inrejection rates by the other models, though rejection rates converge across the specificationsto a low number when the intra-cluster dependence parameter is close to 1. However, we

14

Page 16: Design and Analysis of Cluster-Randomized Field

also find that the outcome for the unweighted city-level model is less dependent on the thepopulation that is drawn. In Figure 7 the error bars show that the minimum coverage underthe null across the 25 drawn populations is close to 80% for the weighted city-level model andthe individual-level model with individual fixed effects. It is close to 70% for the individual-level model with city effects. In contrast, for the unweighted city-level model coverage is 98%or higher across every population that is drawn. In extreme cases, the unweighted city-levelmodel may be more robust to the characteristics of the population.

4.4 Varying the Number of Treated Clusters

In the previous results there were an equal number of treated and control clusters. In Figure8 we vary the number of treated clusters out of 212 clusters in total, with cluster sizescalibrated to the Uber data. All specifications aside from the unweighted city-level modelhave an inverted U-shape for coverage and power. This means that when there are too fewor too many treated clusters they underestimate the standard error.

Under the null, the inverted U-shape for coverage is symmetric, meaning having too manyor too few treated clusters is equally harmful. In contrast, under the alternative hypothesiscoverage and rejection rates are highest when there are slightly more treated clusters thancontrol clusters. This is likely because when there are heterogeneous treatment effects atthe cluster level, additional treated clusters can be valuable for averaging over them andrecovering the mean in the population. For this reason, it may be preferable in certaincases to have more treated clusters than control clusters when designing an experiment ifthe researcher expects a lot of cluster-level treatment effect heterogeneity.

4.5 Varying the Number of Periods

In Figures 9 and 10 we report results when increasing the number of periods before and afterthe treatment event, again with an equal number of treated and control clusters and clustersizes calibrated to the Uber data. We additionally show results when using the bias-correctedFGLS estimator for an AR(1) process in Hansen (2007), though we expect it to give littleimprovement in rejection rates in this case since there is no serial correlation in the responsebeyond individual- and city-level fixed effects. In Figure 9 the treatment event occurs in thefinal period, so in all preceding time periods all of the clusters are untreated. In Figure 10the treatment event occurs in the second time period.

Rejection rates are increasing across all specifications as the number of periods increases,though the improvement levels off after there are more than five periods. This means that thereturns to repeated observations in terms of statistical precision flatten out quickly in cluster-randomized experiments. As the number of time periods increases all of the specificationsshow a reduction in coverage of the treatment effect. We still find substantially higherrejection rates for the weighted models, regardless of the number of time periods.

In Figure 11 we add additional serial correlation in the errors at the individual level. Weassume this follows an AR(1) process. We draw error terms for each city using Algorithm 1

15

Page 17: Design and Analysis of Cluster-Randomized Field

to generate spatial correlation and update them to allow for serial correlation. Algorithm 2shows the precise process:

Algorithm 2: Algorithm to create serial correlation in the data generating process. φ isa scalar that controls the level of serial correlation.

Draw a vector v(1) from Algorithm 1;for d = 1, ..., D do

- Sample v(d+1) from Algorithm 1;

- Set v(d+1) = φv(d) + v(d+1);

end

Set (~εc,1, . . . ,~εc,T ) to equal the last T vectors (v(D−T+1) . . . v(D))

We discard the first four time periods of the above process so that the variance of the errorterms converge to be roughly constant over time. We vary the serial correlation parameterφ and examine its effect on coverage and rejection rates, only considering cases with 4additional periods before treatment takes place.

The serial correlation parameter seems to have little effect on the estimates, at least forthe moderate levels plotted in Figure 11. Coverage appears unchanged, while rejection ratesdecay slightly. When city or individual effects are included such correlation does not seemto have much effect on estimates. Even in this case, the FGLS estimators have little orno improvement in power. This is consistent with Monte Carlo evidence in Tables 3 and4 of Hansen (2007), which shows that for short panels potential gains in power from usingthe FGLS estimator can be quite small. A much larger increase in power comes from usingweighted least squares or individual-level estimation.

5 Simulations in Labor Supply Data

We now evaluate the aforementioned approaches on novel experimental data from Uber. Webegin by focusing on changes in labor supply resulting from the launch of in-app tipping.We start with a description of the data and results for a balance test in the period before theproduct launch. We then provide results for the effect of tipping on labor supply along withfindings from placebo test simulations similar to ones in Bertrand et al. (2003). We focuson labor supply as an exemplar case to show the complications in the data and the impactthis has on estimation of the treatment effect.

5.1 The Data

We worked with Uber to launch optional in-app tipping on its platform in June, 2017. Welaunched tipping in three waves. On June 20, 2017, tipping launched in Seattle, Houston,and Minneapolis for testing. We refer to this as the alpha launch. On July 6, 2017, tippinglaunched in half of the remaining operational cities in the U.S. and Canada, with the otherhalf serving as a control group to evaluate changes in market aggregates of interest: earnings,labor supply, demand, and other market outcomes. We refer to this as the beta launch.

16

Page 18: Design and Analysis of Cluster-Randomized Field

Finally, on July 17, tipping launched across the rest of the US and Canada. We refer tothis as the full launch. For the period between July 6 and July 17, we have experimentalvariation across cities to evaluate the impact of the introduction of tipping. While we wouldhave preferred a longer window, firm level constraints on desiring a fast roll out led to ourfield experimental design.

Cities with business or technological restrictions were excluded from our field experiment.This includes cities such as New York, which launched tipping on July 6 because of busi-ness purposes. The remaining 110 largest cities were randomized to treatment and controlaccording to a matched pairs design. First, cities were grouped by whether they had Uber-POOL and UberEATS. Then within these groupings, cities were matched with the city thatwas most similar across several dimensions, including city size, the current state of driverearnings, and the demographic composition of the city. Similarity was determined by takingthe Euclidean distance, weighting each of these components equally.

Within these pairs, cities were randomly assigned to treatment or control. Cities thatdid not have a pair within their group because the group had an odd number of citieswere randomly assigned to treatment or control. After the randomization, the assignmentof suburbs of large cities were overridden so that they had the same treatment as theirmain metro area. For instance, Orange County and Los Angeles have the same treatmentstatus. The remaining cities were simply randomly assigned to treatment and control withoutconstructing groupings or pairs. Data from the experiment has 215 cities in total, 109 intreatment and 106 in control. We drop six highly irregular cities from the analysis for reasonswe discuss below, leaving 104 cities in treatment and 105 in control.

We consider the four weeks preceding the beta launch and the week after the beta launchto determine whether tipping had an effect on drivers’ labor supply choices. Drivers whoworked on the platform on any day between June 8 and July 13 were included in the dataand were matched to the city in which they worked the most in the 5-week period. For eachdriver, we aggregated total minutes worked for each week, so the observations in the dataare at the driver-week level. The number of minutes worked was coded as 0 for weeks thedriver did not log onto the Uber app.

The data present several challenges for analysis. First, the dispersion in city size is largein the real data. Figure 5 suggests that when cluster sizes are very unequal power declines. Aplot of each city’s size, ordered from smallest to largest is given in Figure 12. The distributionis quite skewed, as more than half of drivers work in the nine largest cities: Los Angeles,Miami, Chicago, San Francisco, Washington D.C., Boston, Philadelphia, Dallas, and Tampa.

The distribution of outcomes in the real data also does not resemble the normally dis-tributed outcomes from the simulations. A driver’s labor supply in a given week can take onany integer value greater than or equal to 0 (up to 10,800 minutes in the week in principle atthe time of the experiment). 36% of the labor supply observations are 0 because drivers onUber often take weeks off. This means that the real data is semi-continuous at the individuallevel. There is also considerable heterogeneity across cities, as seen in Figure 13. Conditionalon working, the distribution of minutes worked for drivers is fat-tailed and varies by city.Densities for the number of minutes worked conditional on working are shown for various

17

Page 19: Design and Analysis of Cluster-Randomized Field

cities in Figure 14.While the data is irregular at the driver level, it is somewhat better behaved at the city

level. The density of city-level minutes worked per week, shown in Figure 15, has a heavyright tail, but is closer to Normal than the individual-level distribution. Figure 16 shows themean number of minutes worked in each week for a sample of cities in the data set. Citiesdiffer in week-to-week variation in labor supply, but generally changes are not substantial.Two exceptions are Corpus Christi and Anchorage, shown in Figure 17. Both cities wererandomized to the treatment group. Uber resumed operations in Corpus Christi on June29th when new state regulations were passed after a long hiatus. Similarly, Uber beganoperating again in Anchorage on June 15 following the passage of municipal legislation. Weremoved these two cities and four other cities that had weeks in which no drivers worked:Laredo, TX; Gallup, NM; Adirondack, NY; and Eagle Pass, TX. Removing these cities left105 control cities and 104 treated cities.

One final complication in the data is that there may be correlated outcomes for driversacross cities that are near each other. For instance, there are drivers who work in SanFrancisco but commute from Sacramento. This case is particularly problematic because SanFrancisco is in the treatment group and Sacramento is in the control group, meaning theremay be contamination between the treatment arms.

5.2 Balance Between Treatment and Control

Prior to the tipping launch, drivers in treated cities tended to work more than drivers incontrol cities. Drivers in control cities worked about 653 minutes per week on average anddrivers in treated cities worked about 683 minutes per week1. There are also more than90,000 fewer treated individuals than control individuals (362,057 in treatment, 452,525 incontrol). A time series plot of minutes worked by day for the treated and control groups is inFigure 18. A t-test at the individual-level clustered by city fails to reject a test for equalityof means across groups at the driver level because the variance in outcomes across cities isso high. Results at the city level are similar. Estimates are in Table 17.

The differences between the treated and control groups are driven by the variation incity sizes and heterogeneity in mean labor supply across cities. If instead of a t-test weconduct a randomization test on the pre-treatment period means we find similar results.The difference in means between the treated and control individuals is about 29 minutes.Over possible randomizations of the treatment, this is at the 34th percentile of absolutevalue of difference in means. When collapsing to the city-level, the realized absolute value ofdifference in means, 12.5 minutes, is at the 43rd percentile. Even though the randomizationwas valid, the lack of balance at the individual-level in the pre-treatment period means asimple difference in means in the treatment period will be an underpowered test statistic

1If the tipping launch has an effect along the extensive margin then this could be caused by drivers leavingor coming onto the platform in the week after the launch, which affects whether they are included in thedata set for the pre-treatment period. However, the results presented in this section also hold when onlyconsidering drivers who drove at some point in the pre-treatment period, suggesting the effect is not drivensolely by this mechanism.

18

Page 20: Design and Analysis of Cluster-Randomized Field

for evaluating the effect of the tipping launch. Table 51 in the appendix shows that a t-testat the city-level has a minimum detectable effect size of more than 10%. Estimating theeffect of the experiment with sufficient power requires adopting assumptions about the timetrends, such as in the panel data framework used in the preceding simulations.

5.3 Labor Supply Results

We run regressions on the labor supply data using insights from the previous simulations.Results for specifications that include city and time fixed effects are shown in Table 18.Tipping does not have a statistically significant effect in any of the specifications. Thepoint estimates differ between the individual- and unweighted city-level models substantiallybecause of heterogeneous mean outcomes across city sizes. To estimate the counterfactualexpected outcome we take the mean outcome in the treated period, subtracting the estimatedtreatment effect for treated individuals. The unweighted city-level models estimate roughly a1.83% change relative to the counterfactual outcome, while the individual-level and weightedcity-level models estimate a 0.39% change. The standard error is higher in all specificationswhen clustering is done at the city level. We also see that the standard error is lower in theclustered individual-level model than in the clustered city-level models.

To check whether coverage is correct under each specification, we re-randomize treatmentacross cities to simulate placebo experiments and examine coverage under the null for therandomization distribution. While the results should not be interpreted as coverage of thetrue treatment effect of launching tipping, especially if the treatment effect is heterogeneousacross clusters, they can still give some information about Type I error rates. Results areshown in Table 19 for 250 re-randomizations.

The mean of the estimated coefficient is slightly larger than 0 across the specifications,but assuming normality of the estimates we would fail to reject a test that they are in fact0 on average. This provides suggestive evidence that bias may not be a major problem inthe regression models. Unclustered estimators generally underestimate the true standarddeviation of the estimated coefficient, leading to poor coverage of the treatment effect. Theindividual-level regression with city clustering also underestimates the standard error, leadingto less than perfect coverage under the null. The clustered city-level regressions with andwithout weighting are the only ones for which the mean standard errors exceed the actualstandard deviation of the coefficients, suggesting these estimators may be more reliable. Ofthe two, the weighted model has a lower standard deviation of the estimated coefficients andalso lower mean standard errors.

Next we see if more granular fixed effects in the individual-level model assist in estimatingthe underlying correlation structure in cities and across time. We try individual-level fixedeffects and also try to control for where and when drivers typically work. We estimate wheredrivers usually work by taking the level-4 geohash2 in which they took the most trips inthe period prior to June 8, 2017. There are 1,584 unique geohashes in which drivers took

2A geohash is a geocoding system that divides the world into a grid of squares of arbitrary precision. Fora more detailed discussion see Cook et al. (2018).

19

Page 21: Design and Analysis of Cluster-Randomized Field

their most trips in the dataset. For context, 153 of them are within Uber’s boundaries forChicago. To estimate when drivers usually work, we take the 6-hour block of the week, offsetby 4 hours to more accurately reflect driving patterns, in which they took the most tripsprior to June 8, 2017. For example, the first block starts on Sunday at 4:00 am and goesuntil Sunday at 10:00 am, followed by Sunday at 10:00 am to Sunday at 4:00 pm, etc. Tocapture both when and where drivers work, we take the geohash × time block in which theytook the most trips.

Results are in Table 20. Adding more granular fixed effects seems to have little effect onthe point estimate and leads to slightly higher standard errors compared to the model withcity fixed effects. Results in placebo test simulations are in Table 21. Including fixed effectsfor when and where drivers typically work seems to lead to a higher standard deviation inthe coefficients compared to using city fixed effects. In contrast, using individual fixed effectsinstead of city fixed effects results in a lower standard deviation for the estimated treatmenteffect. The mean standard error is higher than in the case when including city fixed effectsinstead, but the individual-level model now gives a properly sized test. These findings mimicwhat we found in our simulations. The clustered city-level models and clustered individual-level model with individual fixed effects seem to be most robust to complex features of thedata set.

In Table 22, we collapse to the geohash x city, time x city, and geohash x time x city levelrather than the individual or city level to explore if estimation at a level more granular thanthe city leads to more precise results. All regressions are weighted by the number of driversand clustered by city. All three specifications have very similar point estimates and lowerstandard errors than the weighted city-level model clustered by city. Results in simulationsover the randomization distribution are given in Table 23. All three models have standarderrors that underestimate the standard deviation of the estimated treatment effect.

Finally, we consider estimates where standard errors are stratified. Results from runningstratified regressions are in Table 24 for select specifications. The individual-level model withcity fixed effects does not see a reduction in standard errors, but the model with individualfixed effects does. The city-level models with and without weighting also see a reductionin standard errors. Results in simulations over the randomization distribution are in Table25. Over the randomization distribution there is no heterogeneity in the treatment effectexplained by observables since the true treatment effect is 0. However, the results indicatethat all of the models underestimate the standard error except the unweighted city-levelspecification. While the unweighted stratified city-level model gives proper standard errors,Table 16 shows that the estimate of the treatment effect may be biased under the alternativehypothesis if the treatment effect varies with city size.

6 Market Response to Tipping

Using the results from previous sections we now present more general findings on the effectof the tipping launch on the Uber marketplace. All analyses use weighted city-level modelswith city and time fixed effects. Estimates are clustered by city. Alternative specifications

20

Page 22: Design and Analysis of Cluster-Randomized Field

are left for the appendix. The metrics we focus on are overall labor supply (discussed inSection 5), rider demand, driver utilization, and hourly earnings.

We measure rider demand using a similar method as was used for labor supply. Weinclude all riders who opened the Uber app and entered a destination at some point in thefour weeks prior to the experiment or during the first week of the experiment. Riders areassociated with the city in which they had the most unique sessions in the period, where aunique session is defined as a block of time when a rider opens the app, enters a destination,and considers taking a trip of any product type (e.g. UberX or UberPOOL). The number ofunique sessions is recorded as 0 in weeks when a rider does not consider taking a trip. Weexclude the same cities, most notably New York, as we did for labor supply. There are over16 million riders in total in the data set. A plot of the number of riders in each city is inFigure 19. The distribution of riders per city is highly skewed.

A driver’s utilization is the amount of time ”working” divided by the total amount of timespent on the platform. For example, if a driver spends 10 minutes waiting for a dispatch, 10minutes driving to pick up a passenger, 40 minutes driving that passenger to their destination,and then goes offline, their utilization rate will be 83.33% (50 minutes producing a trip outof 60 minutes online). A complication with analyzing the effect of tipping on utilization isthat it can only be measured in weeks in which a driver works. This creates selection inwhich drivers are observed in each week and makes the panel unbalanced. Even withoutaccounting for selection, changes in utilization are useful for explaining observed changes inhourly earnings.

We analyze three different metrics for hourly earnings: gross earnings before Uber’s com-mission and including earnings from incentives and tips; earnings net of Uber’s commission(which does not apply to tips); and gross earnings before the commission but excluding in-centives and tips. We refer to these metrics as total hourly earnings, post-commission hourlyearnings, and organic hourly earnings, respectively. Post-commission hourly earnings are thedriver’s revenue from Uber. Organic hourly earnings are what the driver earns exclusivelythrough providing trips, factoring in base fare, time spent on trips, distance traveled, andthe surge multiplier.

Weighted city-level estimates are in Table 26. None of the estimates are significant. Thepoint estimates suggest an increase in minutes worked and a decrease in demand, with thecombination of these factors leading to a decrease in utilization. This results in a dropin gross and organic earnings. The fall in post-commission earnings is more muted, in partbecause Uber does not collect commissions on tip earnings. The stability in post-commissionearnings seems consistent with the market equilibration found in Hall et al. (2018).

While our estimates are noisy, they provide an advance for the tipping literature. Forinstance, even though tipping has played an important role for centuries in modern economiesand remains a hallmark of service economies, until recently economists have not made deepcontributions in the area. Notable recent exceptions include the excellent work of Conlinet al. (2003), Azar (2007), and Lynn (2016). While these authors have made importantinroads into the phenomenon of tipping, many open questions remain. For instance, howdoes tipping influence market outcomes such as demand for the good or service and supply of

21

Page 23: Design and Analysis of Cluster-Randomized Field

work hours? All of these issues remain unsettled, but our estimates provide a first indicationthat the market results are not considerable. More work is necessary.

7 Conclusion

This paper helps to clarify questions surrounding design issues and which methods of es-timation give properly sized tests in cluster-randomized settings with panel data. We findthat when clusters vary in size, unit-level estimation with unit fixed effects and cluster-level estimates weighted by the number of observations per cluster can give more preciseestimates than unweighted cluster-level estimates in short panels. If the interest is in mea-suring individual-level treatment effects, these methods may also be more robust than un-weighted cluster-level estimation to cluster-level treatment heterogeneity. We also verify thatstratifying standard errors based on observable characteristics that are correlated with thetreatment effect can lead to substantial improvements in precision, but show that when thedata-generating process is complex it can lead to underestimates of the true standard error.These findings can be applied to other cluster-randomized experiments with big data, bothin the gig economy and in field experiments more generally.

We use the simulation insights to evaluate the effect of introducing tipping on Uber’smarketplace. While measured results are imprecise, we find a short-run increase in laborsupply, decrease in demand, fall in utilization, and fall in hourly earnings for drivers afterthe introduction of tipping. In equilibrium, introducing tipping did not have a dramaticeffect on the Uber marketplace.

References

Abadie, A., S. Athey, G. W. Imbens, and J. M. Wooldridge (2014): “Finite Pop-ulation Causal Standard Errors,” NBER Working Paper 20325.

——— (2017): “When Should You Adjust Standard Errors for Clustering?” NBER WorkingPaper 24003.

Athey, S. and G. W. Imbens (2016): “The Econometrics of Randomized Experiments,”arXiv, https://arxiv.org/abs/1607.00698.

Azar, O. H. (2007): “The Social Norm of Tipping: A Review,” Journal of Applied SocialPsychology, 37, 380–402.

Bertrand, M., E. Duflo, and S. Mullainathan (2003): “How Much Should We TrustDifferences-in-Differences Estimates?” Quarterly Journal of Economics, 119, 249–275.

Brewer, M., T. F. Crossley, and R. Joyce (2017): “Inference with Differences-in-Differences Revisited,” Journal of Econometric Methods, 7.

22

Page 24: Design and Analysis of Cluster-Randomized Field

Cameron, A. C. and D. L. Miller (2015): “A Practitioner’s Guide to Cluster-RobustInference,” Journal of Human Resources, 50, 317–372.

Chandar, B. K., U. Gneezy, J. A. List, and I. Muir (2019): “The Drivers of SocialPreferences: Evidence from a Nationwide Tipping Field Experiment,” NBER WorkingPaper.

Chetverikov, D., B. Larsen, and C. Palmer (2016): “IV Quantile Regression forGroup-Level Treatments,” Econometrica, 84, 809–833.

Conlin, M., M. Lynn, and T. O’Donoghue (2003): “The Norm of Restaurant Tipping,”Journal of Economic Behavior and Organization, 52, 297–321.

Cook, C., R. Diamond, J. Hall, J. A. List, and P. Oyer (2018): “The GenderEarnings Gap in the Gig Economy: Evidence from over a Million Rideshare Drivers,”NBER Working Paper 24732.

Hall, J. V., J. J. Horton, and D. T. Knoepfle (2018): “Pricing Efficiently inDesigned Markets: Evidence from Ride-Sharing,” Working Paper, https://john-joseph-horton.com/papers/uber price.pdf.

Hansen, C. B. (2007): “Generalized Least Squares Inference in Panel and Multilevel Modelswith Serial Correlation and Fixed Effects,” Journal of Econometrics, 140, 670–694.

Liang, K.-Y. and S. L. Zeger (1986): “Longitudinal Data Analysis Using GeneralizedLinear Models,” Biometrika, 73, 13–22.

Lynn, M. (2016): “Why Are We More Likely to Tip Some Service Occupations than Others?:Theory, Evidence, and Implications,” Journal of Economic Psychology, 54, 134–150.

Moulton, B. R. (1990): “An Illustration of a Pitfall in Estimating the Effects of AggregateVariables on Micro Units,” The Review of Economics and Statistics, 72, 334–338.

White, H. (2014): Asymptotic Theory for Econometricians, Academic Press.

23

Page 25: Design and Analysis of Cluster-Randomized Field

Figure 1

1

2

3

4

0.0 0.1 0.2 0.3 0.4 0.5

Treatment Heterogeneity − Observables

Mea

n S

E /

SD

Est

imat

e

Mean SE of Estimate / SD of Estimate − Stratified

1

2

3

4

0.0 0.1 0.2 0.3 0.4 0.5

Treatment Heterogeneity − Observables

Mea

n S

E /

SD

Est

imat

e

Mean SE of Estimate / SD of Estimate − Unstrat.

0.6

0.7

0.8

0.9

1.0

1.1

0.0 0.1 0.2 0.3 0.4 0.5

Treatment Heterogeneity − Observables

Str

at. S

E /

Uns

trat

. SE

Mean Stratified SE / Mean Unstratified SE

Model

City−Level Unweighted

City−Level Weighted

Ind. City FE

Ind. Ind. FE

Stratified SE − Equal Clusters

Note: These figures show the effect of varying the observable treatment heterogeneity parameter on estimated standarderrors when clusters are all the same size. The top-left plot shows the mean standard error divided by the standarddeviation of the estimated treatment effects when standard errors are stratified by the observable characteristics.Averages are taken over 500 different randomizations of treatment status across clusters. If the test is correctly sized,then the ratio of these quantities should equal 1 because the standard error would correctly measure the true variabilityin the estimated treatment effect (the converse may not be true if the standard error is sometimes an underestimate andsometimes an overestimate of the true variability). The top-right plot shows the same plot but when standard errors arenot stratified. The bottom-left figure shows the ratio of the mean stratified standard error to the mean unstratifiedstandard error. If this ratio is less than 1, then stratification may give higher rejection frequencies. Full DGP parametervalues are in Table 36.

24

Page 26: Design and Analysis of Cluster-Randomized Field

Figure 2

1.5

2.0

0.0 0.1 0.2 0.3 0.4 0.5

Treatment Heterogeneity − Observables

Mea

n S

E /

SD

Est

imat

e

Mean SE of Estimate / SD of Estimate − Stratified

1.5

2.0

0.0 0.1 0.2 0.3 0.4 0.5

Treatment Heterogeneity − Observables

Mea

n S

E /

SD

Est

imat

e

Mean SE of Estimate / SD of Estimate − Unstrat.

0.7

0.8

0.9

1.0

0.0 0.1 0.2 0.3 0.4 0.5

Treatment Heterogeneity − Observables

Str

at. S

E /

Uns

trat

. SE

Mean Stratified SE / Mean Unstratified SE

Model

City−Level Unweighted

City−Level Weighted

Ind. City FE

Ind. Ind. FE

Stratified SE − Unequal Clusters

Note: These plots are similar to the ones in Figure 1, but in this case clusters vary in size. Full DGP parameter values arein Table 37.

25

Page 27: Design and Analysis of Cluster-Randomized Field

Figure 3

0.6

0.7

0.8

0.9

1.0

5000 10000 15000 20000Number of Observations

Cov

erag

e

Coverage Under the Null

0.6

0.7

0.8

0.9

1.0

5000 10000 15000 20000Number of Observations

Cov

erag

e

Coverage Under the Alternative

0.00

0.25

0.50

0.75

1.00

5000 10000 15000 20000Number of Observations

Rej

ectio

n F

requ

ency

Rejection Frequency

Model Type

City−Level Unweighted

City−Level Weighted

Ind. City FE

Ind. Ind. FE

Number of Observations

Note: These plots show the effect of varying the number of observations when clusters are all the same size. The top-leftplot shows coverage results under the null hypothesis of no treatment effect as the number of observations varies. Thetop-right plot shows coverage results under the alternative hypothesis. The bottom-left plot shows rejection frequenciesunder the alternative hypothesis. Full DGP parameter values are in Table 38.

26

Page 28: Design and Analysis of Cluster-Randomized Field

Figure 4

0.6

0.7

0.8

0.9

1.0

0 50 100 150 200Number of Clusters

Cov

erag

e

Coverage Under the Null

0.6

0.7

0.8

0.9

1.0

0 50 100 150 200Number of Clusters

Cov

erag

e

Coverage Under the Alternative

0.00

0.25

0.50

0.75

1.00

0 50 100 150 200Number of Clusters

Rej

ectio

n F

requ

ency

Rejection Frequency

Model Type

City−Level Unweighted

City−Level Weighted

Ind. City FE

Ind. Ind. FE

Number of Clusters

Note: These plots show the effect of varying the number of clusters when clusters are all the same size. Full DGPparameter values are in Table 39.

27

Page 29: Design and Analysis of Cluster-Randomized Field

Figure 5

0.6

0.7

0.8

0.9

1.0

0.00 0.25 0.50 0.75 1.00Cluster Size Dispersion Parameter

Cov

erag

e

Coverage Under the Null

0.6

0.7

0.8

0.9

1.0

0.00 0.25 0.50 0.75 1.00Cluster Size Dispersion Parameter

Cov

erag

e

Coverage Under the Alternative

0.00

0.25

0.50

0.75

1.00

0.00 0.25 0.50 0.75 1.00Cluster Size Dispersion Parameter

Rej

ectio

n F

requ

ency

Rejection Frequency

Model Type

City−Level Unweighted

City−Level Weighted

Ind. City FE

Ind. Ind. FE

Cluster Size Dispersion

Note: These plots show the effect of varying the dispersion in cluster sizes. Cluster dispersion varies linearly from equalityto the Uber data calibration. Full DGP parameter values are in Table 40.

28

Page 30: Design and Analysis of Cluster-Randomized Field

Figure 6

0.6

0.7

0.8

0.9

1.0

0.00 0.25 0.50 0.75 1.00Intracluster Dependence Parameter

Cov

erag

e

Coverage Under the Null

0.6

0.7

0.8

0.9

1.0

0.00 0.25 0.50 0.75 1.00Intracluster Dependence Parameter

Cov

erag

e

Coverage Under the Alternative

0.00

0.25

0.50

0.75

1.00

0.00 0.25 0.50 0.75 1.00Intracluster Dependence Parameter

Rej

ectio

n F

requ

ency

Rejection Frequency

Model Type

City−Level Unweighted

City−Level Weighted

Ind. City FE

Ind. Ind. FE

Intracluster Dependence − Equal Clusters

Note: These plots show the effect of varying the amount of spatial correlation in errors for each city-week. Clusters are allthe same size. Full DGP parameter values are in Table 41.

29

Page 31: Design and Analysis of Cluster-Randomized Field

Figure 7

0.6

0.7

0.8

0.9

1.0

0.00 0.25 0.50 0.75 1.00Intracluster Dependence Parameter

Cov

erag

e

Coverage Under the Null

0.6

0.7

0.8

0.9

1.0

0.00 0.25 0.50 0.75 1.00Intracluster Dependence Parameter

Cov

erag

e

Coverage Under the Alternative

0.00

0.25

0.50

0.75

1.00

0.00 0.25 0.50 0.75 1.00Intracluster Dependence Parameter

Rej

ectio

n F

requ

ency

Rejection Frequency

Model Type

City−Level Unweighted

City−Level Weighted

Ind. City FE

Ind. Ind. FE

Intracluster Dependence − Unequal Clusters

Note: These plots show the effect of varying the amount of spatial correlation in errors for each city-week. Cluster sizes arecalibrated to the Uber data. Full DGP parameter values are in Table 42.

30

Page 32: Design and Analysis of Cluster-Randomized Field

Figure 8

0.6

0.7

0.8

0.9

1.0

0 50 100 150 200Number of Treated Clusters

Cov

erag

e

Coverage Under the Null

0.5

0.6

0.7

0.8

0.9

1.0

0 50 100 150 200Number of Treated Clusters

Cov

erag

e

Coverage Under the Alternative

0.00

0.25

0.50

0.75

1.00

0 50 100 150 200Number of Treated Clusters

Rej

ectio

n F

requ

ency

Rejection Frequency

Model Type

City−Level Unweighted

City−Level Weighted

Ind. City FE

Ind. Ind. FE

Number of Treated Clusters

Note: These plots show the effect of varying the number of treated clusters. Cluster sizes are calibrated to the Uber data.Full DGP parameter values are in Table 43.

31

Page 33: Design and Analysis of Cluster-Randomized Field

Figure 9

0.6

0.7

0.8

0.9

1.0

5 10 15 20Time Periods Before Treatment

Cov

erag

e

Coverage Under the Null

0.6

0.7

0.8

0.9

1.0

5 10 15 20Time Periods Before Treatment

Cov

erag

e

Coverage Under the Alternative

0.00

0.25

0.50

0.75

1.00

5 10 15 20Time Periods Before Treatment

Rej

ectio

n F

requ

ency

Rejection Frequency

Model Type

City−Level Unweighted

City−Level Weighted

FGLS BC

FGLS BC Weighted

Ind. City FE

Ind. Ind. FE

Time Periods Before Treatment

Note: These plots show the effect of varying the number of time periods before the treatment event. Treatment occursduring the last period for all results displayed. Cluster sizes are calibrated to the Uber data. Full DGP parameter valuesare in Table 44.

32

Page 34: Design and Analysis of Cluster-Randomized Field

Figure 10

0.6

0.7

0.8

0.9

1.0

5 10 15 20Time Periods After Treatment

Cov

erag

e

Coverage Under the Null

0.6

0.7

0.8

0.9

1.0

5 10 15 20Time Periods After Treatment

Cov

erag

e

Coverage Under the Alternative

0.00

0.25

0.50

0.75

1.00

5 10 15 20Time Periods After Treatment

Rej

ectio

n F

requ

ency

Rejection Frequency

Model Type

City−Level Unweighted

City−Level Weighted

FGLS BC

FGLS BC Weighted

Ind. City FE

Ind. Ind. FE

Time Periods After Treatment

Note: These plots show the effect of varying the number of time periods after the treatment event. Treatment occursduring period 2 for all results displayed. Cluster sizes are calibrated to the Uber data. Full DGP parameter values are inTable 45.

33

Page 35: Design and Analysis of Cluster-Randomized Field

Figure 11

0.6

0.7

0.8

0.9

1.0

0.0 0.2 0.4 0.6Serial Correlation Parameter

Cov

erag

e

Coverage Under the Null

0.6

0.7

0.8

0.9

1.0

0.0 0.2 0.4 0.6Serial Correlation Parameter

Cov

erag

e

Coverage Under the Alternative

0.00

0.25

0.50

0.75

1.00

0.0 0.2 0.4 0.6Serial Correlation Parameter

Rej

ectio

n F

requ

ency

Rejection Frequency

Model Type

City−Level Unweighted

City−Level Weighted

FGLS BC

FGLS BC Weighted

Ind. City FE

Ind. Ind. FE

Serial Correlation

Note: These plots show the effect of varying the amount of serial correlation in the errors. In each case there are five timeperiods, with treatment occurring during the fifth period. Cluster sizes are calibrated to the Uber data. Full DGPparameter values are in Table 46.

34

Page 36: Design and Analysis of Cluster-Randomized Field

Figure 12

Note: This plot shows the number of drivers across cities in the United States and Canada. Cities are ordered from fewestdrivers to most drivers.

Figure 13

Note: This plot shows the percentage of driver-weeks with 0 minutes worked by city. Cities are ordered from fewest driversto most drivers.

35

Page 37: Design and Analysis of Cluster-Randomized Field

Figure 14

Note: This plot shows the distribution over driver-weeks of minutes worked conditional on working for four different cities:San Francisco, Portland, Montreal, and Connecticut.

Figure 15

Note: This plot shows the density across city-weeks of the mean number of minutes worked by drivers.

36

Page 38: Design and Analysis of Cluster-Randomized Field

Figure 16

Note: This plot shows the mean number of minutes worked for each week in various cities. Each line in the plot is adifferent city.

Figure 17

Note: This plot shows the mean number of minutes worked for each week in Anchorage and Corpus Christi, two cities thatare removed from the sample used in the analysis. Uber began operating in Anchorage on June 15, 2017 following ahiatus due to regulatory reasons. This date corresponds to the start of week 2. Uber started operating in Corpus Christion June 29, 2017, or at the start of week 4.

37

Page 39: Design and Analysis of Cluster-Randomized Field

Figure 18

Note: This plot shows the time series for minutes worked for the treated and control groups. Tipping launched in treatedcities on July 06.

38

Page 40: Design and Analysis of Cluster-Randomized Field

Figure 19

Note: This plot shows the number of riders across cities in the United States and Canada. Cities are ordered from fewestriders to most riders.

Table 1

Mean Coef. SD Coef. Mean SE Coverage

Individual, No Clustering -0.001 0.012 0.014 0.980Individual, City Clustering -0.001 0.012 0.013 0.966

City, No Clustering -0.001 0.012 0.013 0.968

Note: This table shows simulation results under the null hypothesis of no effect for a one-periodmodel with no cluster heterogeneity. All cities are the same size. Full DGP parameter valuesare in Table 27.

Table 2

Mean Coef. SD Coef. Mean SE Cov. Rej. Freq.

Individual, No Clustering 0.035 0.013 0.014 0.968 0.720Individual, City Clustering 0.035 0.013 0.013 0.958 0.782

City, No Clustering 0.035 0.013 0.013 0.958 0.772

Note: This table shows simulation results under the alternative hypothesis for a one-period model with nocluster heterogeneity. All cities are the same size. τp = 0.035. Full DGP parameter values are in Table27.

39

Page 41: Design and Analysis of Cluster-Randomized Field

Table 3

Mean Coef. SD Coef. Mean SE Coverage

Individual, No Clustering 0.000 0.015 0.015 0.932Individual, City Clustering 0.000 0.015 0.015 0.922

City, No Clustering 0.001 0.054 0.053 0.958City, EHW SE 0.001 0.054 0.053 0.960

City, Size Weighted 0.000 0.016 0.014 0.922City, Size Weighted, EHW SE 0.000 0.016 0.015 0.924

Note: This table shows simulation results under the null hypothesis of no effect for a one-periodmodel with no cluster heterogeneity. Cluster sizes are calibrated to the Uber data. Full DGPparameter values are in Table 28.

Table 4

Mean Coef. SD Coef. Mean SE Cov. Rej. Freq.

Individual, No Clustering 0.035 0.015 0.015 0.932 0.646Individual, City Clustering 0.035 0.015 0.015 0.922 0.622

City, No Clustering 0.036 0.054 0.053 0.958 0.092City, EHW SE 0.036 0.054 0.053 0.96 0.096

City, Size Weighted 0.035 0.016 0.014 0.922 0.664City, Size Weighted, EHW SE 0.035 0.016 0.015 0.924 0.616

Note: This table shows simulation results under the alternative hypothesis for a one-period model with nocluster heterogeneity. Cluster sizes are calibrated to the Uber data. τp = 0.035. Full DGP parameter valuesare in Table 28.

Table 5

Mean Coef. SD Coef. Mean SE Coverage

Individual, No Clustering 0.002 0.018 0.014 0.896Individual, City Clustering 0.002 0.018 0.019 0.958

City, No Clustering 0.002 0.018 0.019 0.954

Note: This table shows simulation results under the null hypothesis of no effect for a one-period modelwith city-specific intercepts in the DGP. All cities are the same size. Full DGP parameter values arein Table 29.

Table 6

Mean Coef. SD Coef. Mean SE Cov. Rej. Freq.

Individual, No Clustering 0.037 0.018 0.014 0.896 0.682Individual, City Clustering 0.037 0.018 0.019 0.958 0.522

City, No Clustering 0.037 0.018 0.019 0.954 0.478

Note: This table shows simulation results under the alternative hypothesis for a one-period model withcity-specific intercepts in the DGP. All cities are the same size. τp = 0.035. Full DGP parameter valuesare in Table 29.

40

Page 42: Design and Analysis of Cluster-Randomized Field

Table 7

Mean Coef. SD Coef. Mean SE Coverage

Individual, No Clustering -0.002 0.051 0.015 0.380Individual, City Clustering -0.002 0.051 0.045 0.862

City, No Clustering -0.002 0.065 0.066 0.950City, EHW SE -0.002 0.065 0.066 0.950

City, Size Weighted -0.002 0.051 0.021 0.532City, Size Weighted, EHW SE -0.002 0.051 0.045 0.876

Note: This table shows simulation results under the null hypothesis of no effect for a one-periodmodel with city-specific intercepts in the DGP. Cluster sizes are calibrated to the Uber data. FullDGP parameter values are in Table 30.

Table 8

Mean Coef. SD Coef. Mean SE Cov. Rej. Freq.

Individual, No Clustering 0.033 0.051 0.015 0.380 0.532Individual, City Clustering 0.033 0.051 0.045 0.862 0.182

City, No Clustering 0.033 0.065 0.066 0.950 0.074City, EHW SE 0.033 0.065 0.066 0.950 0.072

City, Size Weighted 0.033 0.051 0.021 0.532 0.438City, Size Weighted, EHW SE 0.033 0.051 0.045 0.876 0.174

Note: This table shows simulation results under the alternative hypothesis for a one-period model with city-specific intercepts in the DGP. Cluster sizes are calibrated to the Uber data. τp = 0.035. Full DGP parametervalues are in Table 30.

Table 9

Mean Coef. SD Coef. Mean SE Coverage

Individual, No Clustering -0.001 0.029 0.021 0.816Individual, No Clus. FE, City Clustering 0.000 0.041 0.037 0.898

Individual, City Clustering -0.001 0.029 0.027 0.896City, No Clustering 0.001 0.078 0.080 0.952

City, City Clustering 0.001 0.078 0.113 0.994City, Clustered, No Clus. FE 0.001 0.062 0.062 0.952

City, Size Weighted, No Clustering -0.001 0.029 0.021 0.826City, Clustered, Size Weighted -0.001 0.029 0.038 0.988

City, Size Weighted + Clustered, No Clus. FE 0.000 0.041 0.037 0.898

Note: This table shows simulation results under the null hypothesis of no effect for a two-period model with city ef-fects and time effects in the DGP. Cluster sizes are calibrated to the Uber data. All model specifications includetime and city fixed effects unless otherwise noted. Full DGP parameter values are in Table 31.

41

Page 43: Design and Analysis of Cluster-Randomized Field

Table 10

Mean Coef. SD Coef. Mean SE Cov. Rej. Freq.

Individual, No Clustering 0.033 0.021 0.021 0.95 0.338Individual, No Clus. FE, City Clustering 0.032 0.034 0.031 0.906 0.19

Individual, City Clustering 0.033 0.021 0.02 0.93 0.35City, No Clustering 0.026 0.075 0.076 0.952 0.06

City, City Clustering 0.026 0.075 0.107 1 0.004City, Clustered, No Clus. FE 0.027 0.057 0.058 0.95 0.056

City, Size Weighted, No Clustering 0.033 0.021 0.021 0.948 0.354City, Clustered, Size Weighted 0.033 0.021 0.029 0.992 0.136

City, Size Weighted + Clustered, No Clus. FE 0.032 0.034 0.032 0.906 0.188

Note: This table shows simulation results under the alternative hypothesis for a two-period model with city effects and time effectsin the DGP. The treatment effect varies by individual. Cluster sizes are calibrated to the Uber data. All model specificationsinclude time and city fixed effects unless otherwise noted. τp = 0.0326. Full DGP parameter values are in Table 31.

Table 11

Mean Coef. SD Coef. Mean SE Coverage

Individual, No Clustering 0.000 0.019 0.021 0.966Individual, City Clustering 0.000 0.019 0.018 0.936

Individual, Ind. FE, City Clustering 0.000 0.019 0.026 0.988City, No Clustering 0.004 0.072 0.073 0.956

City, City Clustering 0.004 0.072 0.103 0.998City, Size Weighted, No Clustering 0.000 0.019 0.020 0.962

City, Size Weighted, City Clustering 0.000 0.019 0.026 0.990

Note: This table shows simulation results under the null hypothesis of no effect for a two-period modelwith city effects, time effects, and individual effects in the DGP. Cluster sizes are calibrated to the Uberdata. All model specifications include time and city fixed effects unless otherwise noted. Full DGP pa-rameter values are in Table 32.

Table 12

Mean Coef. SD Coef. Mean SE Cov. Rej. Freq.

Individual, No Clustering 0.064 0.026 0.021 0.89 0.808Individual, City Clustering 0.064 0.026 0.028 0.932 0.548

Individual, Ind. FE, City Clustering 0.064 0.026 0.039 0.982 0.408City, No Clustering 0.133 0.069 0.071 0.854 0.456

City, City Clustering 0.133 0.069 0.101 0.958 0.166City, Size Weighted, No Clustering 0.064 0.026 0.021 0.874 0.810

City, Size Weighted, City Clustering 0.064 0.026 0.040 0.982 0.404

Note: This table shows simulation results under the alternative hypothesis for a two-period model with city effects,time effects, and individual effects in the DGP. The treatment effect varies by individual and by cluster. Clustersizes are calibrated to the Uber data. All model specifications include time and city fixed effects unless otherwisenoted. τp = 0.116. Full DGP parameter values are in Table 32.

42

Page 44: Design and Analysis of Cluster-Randomized Field

Table 13

Mean Coef. SD Coef. Mean SE Coverage

Individual, No Clustering -0.001 0.025 0.018 0.862Individual, City Clustering -0.001 0.025 0.024 0.930

Individual, Ind. FE, City Clustering -0.001 0.025 0.035 0.994City, No Clustering 0.002 0.088 0.087 0.948

City, City Clustering 0.002 0.088 0.123 0.996City, Size Weighted -0.001 0.025 0.023 0.932

City, Size Weighted, City Clustering -0.001 0.025 0.035 0.994

Note: This table shows simulation results under the null hypothesis of no effect for a two-period modelwith city effects, time effects, and individual effects in the DGP. Individual error terms within a city-week are drawn from Algorithm 1. Cluster sizes are calibrated to the Uber data. All model specificationsinclude time and city fixed effects unless otherwise noted. Full DGP parameter values are in Table 33.

Table 14

Mean Coef. SD Coef. Mean SE Cov. Rej. Freq.

Individual, No Clustering 0.064 0.027 0.019 0.83 0.822Individual, City Clustering 0.064 0.027 0.029 0.958 0.582

Individual, Ind. FE, City Clustering 0.064 0.027 0.041 0.994 0.36City, No Clustering 0.132 0.077 0.077 0.854 0.402

City, City Clustering 0.132 0.077 0.109 0.976 0.13City, Size Weighted 0.064 0.027 0.024 0.92 0.72

City, Size Weighted, City Clustering 0.064 0.027 0.041 0.994 0.356

Note: This table shows simulation results under the alternative hypothesis for a two-period model with city effects,time effects, and individual effects in the DGP. The treatment effect varies by individual and by cluster. Individualerror terms within a city-week are drawn from Algorithm 1. Cluster sizes are calibrated to the Uber data. All modelspecifications include time and city fixed effects unless otherwise noted. τp = 0.116. Full DGP parameter values arein Table 33.

Table 15

Mean Coef. SD Coef. Mean SE Cov. Rej. Freq.

Individual, City Clustering 0.35 0.026 0.028 0.962 1Individual, City Clustering, Stratified 0.35 0.025 0.026 0.948 1

Individual, Ind. FE, City Clus. 0.35 0.026 0.039 0.998 1Individual, Ind. FE, City Clus., Stratified 0.35 0.026 0.026 0.956 1

City, City Clustering 0.35 0.026 0.039 0.998 1City, City Clustering, Stratified 0.35 0.026 0.026 0.956 1

City, Size Weighted, City Clustering 0.35 0.026 0.039 0.998 1City, Weighted + Clustered, Stratified 0.35 0.026 0.026 0.956 1

Note: This table shows simulation results under the alternative hypothesis for a two-period model with city effects, timeeffects, and individual effects in the DGP. The treatment effect varies by individual and by cluster. Individual error termswithin a city-week are drawn from Algorithm 1. Further, treatment effects vary by an observable cluster-level parameter.Clusters are equal sizes. All model specifications include time and city fixed effects unless otherwise noted. τp = 0.351.Full DGP parameter values are in Table 34.

43

Page 45: Design and Analysis of Cluster-Randomized Field

Table 16

Mean Coef. SD Coef. Mean SE Cov. Rej. Freq.

Individual, City Clustering 0.524 0.025 0.025 0.942 1Individual, City Clustering, Stratified 0.522 0.023 0.025 0.96 1

Individual, Ind. FE, City Clus. 0.524 0.025 0.035 0.988 1Individual, Ind. FE, City Clus., Stratified 0.524 0.025 0.025 0.942 1

City, City Clustering 0.386 0.077 0.108 0.834 0.986City, City Clustering, Stratified 0.386 0.077 0.076 0.574 1

City, Size Weighted, City Clustering 0.524 0.025 0.035 0.988 1City, Weighted + Clustered, Stratified 0.524 0.025 0.025 0.942 1

Note: This table shows simulation results under the alternative hypothesis for a two-period model with city effects, timeeffects, and individual effects in the DGP. The treatment effect varies by individual and by cluster. Individual error termswithin a city-week are drawn from Algorithm 1. Further, treatment effects vary by cluster size. Cluster sizes are calibratedto the Uber data. All model specifications include time and city fixed effects unless otherwise noted. τp = 0.534. Full DGPparameter values are in Table 35.

Table 17

Level Outcome T P Control Mean Treat. Mean

Individual Minutes Worked 0.440 0.660 653.498 682.553City Minutes Worked 0.553 0.580 500.417 512.871

Individual Did Work 0.501 0.617 0.636 0.646City Did Work -0.114 0.909 0.569 0.568

Individual Minutes Worked |Worked 0.393 0.694 1, 027.669 1, 056.076City Minutes Worked |Worked 1.019 0.308 863.138 889.840

Note: This table shows tests for equality of means between treatment and control in the period before the tipping launch.All estimates are clustered by city. The first row compares minutes worked between treatment and control at the driverlevel. The second row compares minutes worked after first collapsing to means across drivers for each city-week. Rows3 and 4 show t-test estimates for whether the driver worked at least one minute. Rows 5 and 6 show estimates for howmuch drivers work conditional on working at least one minute.

44

Page 46: Design and Analysis of Cluster-Randomized Field

Table 18

Dependent variable:

Minutes Worked

(1) (2) (3) (4) (5) (6)

Treated 2.725 2.725 9.637 9.637 2.725 2.725(2.408) (8.691) (7.188) (10.498) (4.579) (9.742)

Unit Driver Driver City City City Weighted City WeightedCity Clustering No Yes No Yes No YesCounterfactual 697 697 527 527 697 697% Change 0.39% 0.39% 1.83% 1.83% 0.39% 0.39%

Observations 4,072,910 4,072,910 1,045 1,045 1,045 1,045R2 0.032 0.032 0.941 0.941 0.978 0.978Adjusted R2 0.031 0.031 0.925 0.925 0.972 0.972

Note: ∗p<0.1; ∗∗p<0.05; ∗∗∗p<0.01

Note: This table shows regression estimates for the effect of the tipping launch on labor supply. Allmodel specifications include time and city effects. In columns 1 and 2, observations are at the driver-week level. In columns 3 and 4, observations are at the city-week level, and regressions are unweighted.In columns 5 and 6, observations are at the city-week level, and regressions are weighted by the numberof drivers for each city-week.

Table 19

Mean Coef. SD Coef. Mean SE Coverage

Individual, No Clustering 0.064 8.343 2.435 0.420Individual, City Clustering 0.064 8.343 7.707 0.900

City, No Clustering 0.239 9.221 7.188 0.883City, Size Weighted 0.239 9.221 10.481 0.983

City, Clustered 0.064 8.343 4.621 0.727City, Clustered + Weighted 0.064 8.343 8.638 0.940

Note: This table shows results over placebo randomizations in the real Uber labor supply data. All modelspecifications include time and city fixed effects.

45

Page 47: Design and Analysis of Cluster-Randomized Field

Table 20

Dependent variable:

Minutes Worked

(1) (2) (3) (4)

Treated 2.725 2.725 2.725 2.725(9.717) (8.693) (8.692) (8.712)

Unit Driver Driver Driver DriverFE Driver Geohash Time of Week Geohash x Time of WeekCity Clustering Yes Yes Yes YesCounterfactual 697 697 697 697% Change 0.39% 0.39% 0.39% 0.39%

Observations 4,072,910 4,072,910 4,072,910 4,072,910R2 0.808 0.085 0.075 0.109Adjusted R2 0.760 0.085 0.075 0.105

Note: ∗p<0.1; ∗∗p<0.05; ∗∗∗p<0.01

Note: This table shows regression results for the effect of the tipping launch on labor supply when usingmore granular fixed effects in addition to city and week effects. Observations are at the driver-weeklevel. All estimates are clustered by city. The specification in column 1 uses driver effects. Column 2uses more granular location effects based on where drivers work, while column 3 uses more granulartime effects based on when drivers typically work. Column 4 includes effects for the interaction of thelocation and time effects.

Table 21

Mean Coef. SD Coef. Mean SE Coverage

Individual FE 0.222 8.035 8.669 0.948Top Geohash FE 0.858 8.547 7.663 0.892

Top Week Time FE 0.858 8.547 7.661 0.892Geohash x Time 0.858 8.547 7.679 0.892

Note: This table shows results over placebo randomizations in the real Uber labor supplydata with more granular fixed effects. All model specifications include time and city fixedeffects.

46

Page 48: Design and Analysis of Cluster-Randomized Field

Table 22

Dependent variable:

Minutes Worked

(1) (2) (3)

Treated 2.725 2.725 2.725(8.715) (8.725) (8.697)

Unit City x Geohash City x Time City x Geohash x TimeCity Clustering Yes Yes YesCounterfactual 537 580 577% Change 0.51% 0.47% 0.47%

Observations 38,400 27,740 160,580R2 0.329 0.355 0.242Adjusted R2 0.325 0.350 0.241

Note: ∗p<0.1; ∗∗p<0.05; ∗∗∗p<0.01

Note: This table shows regression results for the effect of the tipping launch on labor sup-ply when using observations more granular than city-weeks. All estimates are clusteredby city and weighted by the number of drivers. All model specifications include time andcity fixed effects. The specification in column 1 collapses to the city x geohash level. Col-umn 2 collapses to the city x typical time block level, while column 3 collapses to the cityx geohash x time block level.

Table 23

Mean Coef. SD Coef. Mean SE Coverage

Geohash x City 0.064 8.343 7.728 0.900Week Time x City 0.064 8.343 7.736 0.900

Geohash x Week Time x City 0.064 8.343 7.712 0.900

Note: This table shows results over placebo randomizations in the Uber labor supply data collapsing tolevel more granular than cities. All model specifications include time and city fixed effects.

47

Page 49: Design and Analysis of Cluster-Randomized Field

Table 24

Dependent variable:

Minutes Worked

(1) (2) (3) (4)

Treated 2.725 2.725 9.637 2.725(8.734) (8.734) (9.394) (8.734)

Unit Individual Individual City City WeightedFE City Individual City CityCity Clustering Yes Yes Yes YesStratification Yes Yes Yes YesCounterfactual 697 697 527 697% Change 0.39% 0.39% 1.83% 0.39%Observations 4,072,910 4,072,910 1,045 1,045

Note: ∗p<0.1; ∗∗p<0.05; ∗∗∗p<0.01

Note: This table shows regression results for the effect of the tipping launch on la-bor supply when stratifying standard errors by cluster size tercile. All estimatesare clustered by city and include time and city fixed effects.

Table 25

Mean Coef. SD Coef. Mean SE Coverage

Individual, Driver FE 0.124 8.278 7.725 0.920Individual, City FE 0.187 8.353 7.787 0.900

City, Clustered -0.526 9.220 9.351 0.968City, Clustered + Weighted 0.124 8.278 7.725 0.920

Note: This table shows results over placebo randomizations in the real Uber labor supply data withstandard errors stratified by city size tercile.

48

Page 50: Design and Analysis of Cluster-Randomized Field

Table 26

Dependent variable:

Minutes Worked Uniq. Rider Sess. Utilization Total Post-Comm. Organic

(1) (2) (3) (4) (5) (6)

Treated 2.725 −0.013 −0.004 −0.227 −0.086 −0.173(9.742) (0.028) (0.010) (0.554) (0.436) (0.498)

Counterfactual 697 0.997 0.489 16.22 12.42 15.03% Change 0.39% -1.320% -0.750% -1.40% -0.69% -1.15%

Observations 1,045 1,045 1,045 1,045 1,045 1,045R2 0.978 0.964 0.964 0.966 0.969 0.953Adjusted R2 0.972 0.955 0.954 0.958 0.961 0.941

Note: ∗p<0.1; ∗∗p<0.05; ∗∗∗p<0.01

Note: This table shows regression estimates of the effect of tipping on labor supply, demand, utilization, andearnings. Observations are at the city-week level and weighted by the number of drivers (or weighted by thenumber of riders in the case of sessions). Standard errors are clustered by city. Minutes worked estimates arethe same as in Column 6 in Table 18. Unique rider sessions are blocks of time when a rider opens the app, en-ters a destination, and considers taking a trip of any product type (e.g. UberX or UberPOOL). Gross hourlyearnings include tips and non-trip driving incentives but are calculated before the commission is deducted. Post-commission hourly earnings are the same as gross hourly earnings minus the commission paid to Uber. Organichourly earnings are calculated before the commission is deducted but exclude incentives and tips.

49

Page 51: Design and Analysis of Cluster-Randomized Field

Appendix A Simulation Calibrations

Table 27

Parameter Meaning Valueτ Treatment Effect Parameter 0.035C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 1Ttreated Time of Treatment 1R Number of Iterations in Simulation 500Note: DGP parameter settings for Tables 1 and 2. Clusters are equal sizes.

Table 28

Parameter Meaning Valueτ Treatment Effect Parameter 0.035C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 1Ttreated Time of Treatment 1R Number of Iterations in Simulation 500Note: DGP parameter settings for Tables 3 and 4. Cluster sizes are calibrated

to the Uber data.

Table 29

Parameter Meaning Valueτ Average Treatment Effect 0.035σc SD of City FE 0.1C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 1Ttreated Time of Treatment 1R Number of Iterations in Simulation 500Note: DGP parameter settings for Tables 5 and 6. Clusters are equal sizes.

50

Page 52: Design and Analysis of Cluster-Randomized Field

Table 30

Parameter Meaning Valueτ Average Treatment Effect 0.035σc SD of City FE 0.1C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 1Ttreated Time of Treatment 1R Number of Iterations in Simulation 500Note: DGP parameter settings for Tables 7 and 8. Cluster sizes are calibrated

to the Uber data.

Table 31

Parameter Meaning Valueτ Average Treatment Effect 0.035στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Iterations in Simulation 500Note: DGP parameter settings for Tables 9 and 10. Cluster sizes are

calibrated to the Uber data.

51

Page 53: Design and Analysis of Cluster-Randomized Field

Table 32

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Iterations in Simulation 500Note: DGP parameter settings for Tables 11 and 12. Cluster sizes are

calibrated to the Uber data.

Table 33

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1ρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Iterations in Simulation 500Note: DGP parameter settings for Tables 13 and 14. Cluster sizes are

calibrated to the Uber data.

52

Page 54: Design and Analysis of Cluster-Randomized Field

Table 34

Parameter Meaning Valueτ Average Treatment Effect 0.35στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.02τz Observable Het. Param. 0.1ρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Iterations in Simulation 500Note: DGP parameter settings for Table 15. Clusters are equal sizes.

Table 35

Parameter Meaning Valueτ Average Treatment Effect 0.35στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.02τz Observable Het. Param. 0.1ρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Iterations in Simulation 500Note: DGP parameter settings for Table 16. Cluster sizes are calibrated to

Uber data.

53

Page 55: Design and Analysis of Cluster-Randomized Field

Table 36

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. Variesρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Iterations in Simulation 500Note: DGP parameter settings for Figure 1. Clusters are equal sizes.

Table 37

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. Variesρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Iterations in Simulation 500Note: DGP parameter settings for Figure 2. Cluster sizes are calibrated to

the Uber data.

54

Page 56: Design and Analysis of Cluster-Randomized Field

Table 38

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. 0ρ Spatial Correlation Parameter 0.5C Number of Clusters 200N Number of Individuals Variesσ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Re-randomizations 200P Number of Populations Drawn 25Note: DGP parameter settings for Figure 3. Clusters are equal sizes.

Table 39

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. 0ρ Spatial Correlation Parameter 0.5C Number of Clusters VariesN Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Re-randomizations 200P Number of Populations Drawn 25Note: DGP parameter settings for Figure 4. Clusters are equal sizes.

55

Page 57: Design and Analysis of Cluster-Randomized Field

Table 40

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. 0ρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000δ Cluster Size Variation Variesσ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Re-randomizations 200P Number of Populations Drawn 25Note: DGP parameter settings for Figure 5.

Table 41

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. 0ρ Spatial Correlation Parameter VariesC Number of Clusters 200N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Re-randomizations 200P Number of Populations Drawn 25Note: DGP parameter settings for Figure 6. Clusters are equal sizes.

56

Page 58: Design and Analysis of Cluster-Randomized Field

Table 42

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. 0ρ Spatial Correlation Parameter VariesC Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Re-randomizations 200P Number of Populations Drawn 25Note: DGP parameter settings for Figure 7. Cluster sizes calibrated to the

Uber data.

Table 43

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. 0ρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000|Ctreated| Number of Treated Clusters Variesσ SD of Error Term 1T Number of Time Periods 2Ttreated Time of Treatment 2R Number of Re-randomizations 200P Number of Populations Drawn 25Note: DGP parameter settings for Figure 8. Cluster sizes calibrated to the

Uber data.

57

Page 59: Design and Analysis of Cluster-Randomized Field

Table 44

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. 0ρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods VariesTtreated Time of Treatment Same as TR Number of Re-randomizations 200P Number of Populations Drawn 25Note: DGP parameter settings for Figure 9. Cluster sizes calibrated to the Uber

data.

Table 45

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1τz Observable Het. Param. 0ρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods VariesTtreated Time of Treatment 2R Number of Re-randomizations 200P Number of Populations Drawn 25Note: DGP parameter settings for Figure 10. Cluster sizes calibrated to the

Uber data.

58

Page 60: Design and Analysis of Cluster-Randomized Field

Table 46

Parameter Meaning Valueτ Average Treatment Effect 0.1στ SD of Ind. Treatment Effect 0.5σc SD of City FE 0.1σt SD of Time FE 0.1σγ SD of Ind. FE 0.1σh SD of City Treatment Effect Het. 0.1φ Serial Correlation Parameter Variesτz Observable Het. Param. 0ρ Spatial Correlation Parameter 0.5C Number of Clusters 212N Number of Individuals 20,000σ SD of Error Term 1T Number of Time Periods 5Ttreated Time of Treatment 5R Number of Re-randomizations 200P Number of Populations Drawn 25Note: DGP parameter settings for Figure 11. Cluster sizes calibrated to the

Uber data.

Appendix B Alternative Regression Specifications and

Outcomes

Table 47

Var Coef. SE tstat pval dof mean % change

1 Unique Sessions -0.013 0.028 -0.461 0.645 831 0.997 -1.3192 Riders w/ Sessions -0.002 0.006 -0.331 0.741 831 0.342 -0.6053 Unique Sessions Cond. >0 0.034 0.034 0.985 0.325 831 2.834 1.199

Note: This table shows regression results for rider-level demand outcomes. Regressions are estimated at the citylevel and weighted by the number of riders in each city-week. Standard errors are clustered by city. Row 1 showsestimates for the number of unique rider sessions. Row 2 shows estimates for an indicator for whether the riderhad at least on session. Row 3 shows estimates for the number of unique sessions conditional on the rider havingat least one unique session for the week.

59

Page 61: Design and Analysis of Cluster-Randomized Field

Table 48

Var Coef. SE tstat pval dof mean % change

1 Total Earnings -1.940 6.819 -0.285 0.776 831 209.688 -0.9252 Organic Earnings -3.953 6.835 -0.578 0.563 831 192.500 -2.0543 Minutes Worked 2.725 9.742 0.280 0.780 831 697.208 0.3914 Incentive Earnings -0.161 0.910 -0.177 0.859 831 6.006 -2.6855 Surge Earnings -2.283 3.185 -0.717 0.474 831 5.228 -43.6656 Post-Commission Earn. -0.486 5.261 -0.092 0.926 831 161.280 -0.3017 Hourly Earnings -0.227 0.554 -0.410 0.682 831 16.219 -1.3998 Organic Hourly -0.173 0.498 -0.347 0.729 831 15.030 -1.1519 Utilization -0.004 0.010 -0.378 0.706 831 0.489 -0.74610 Incentive Hourly -0.039 0.051 -0.752 0.452 831 0.462 -8.34811 Surge Hourly -0.173 0.212 -0.816 0.415 831 0.459 -37.73112 Post-Commission Hourly -0.086 0.436 -0.196 0.844 831 12.423 -0.69013 Minutes Worked 2.725 9.742 0.280 0.780 831 697.208 0.39114 Min. Worked - Intensive -3.339 13.065 -0.256 0.798 831 1, 054.207 -0.31715 Did Work 0.004 0.003 1.225 0.221 831 0.662 0.62316 Completed Trips -0.128 0.252 -0.508 0.612 831 28.323 -0.45217 Mean Fare -0.317 0.277 -1.144 0.253 831 12.905 -2.45718 Trip Distance 0 0.034 0.003 0.998 831 6.625 0.00219 Trip Duration 5.510 8.625 0.639 0.523 831 972.191 0.56720 Trip Surge -0.007 0.011 -0.650 0.516 831 1.027 -0.721

Note: This table shows regression results for driver-level and trip-level outcomes. Regressions are estimated at the citylevel and weighted by the number of drivers in each city-week. Standard errors are clustered by city. Rows 1-6 refer toaggregate earnings and labor supply measures for each driver-week. Rows 7-12 report hourly measures, excludingdriver-weeks in which the driver did not work. Rows 13-15 report estimates for minutes worked, minutes workedconditional on working at least one minute, and an indicator for whether or not the driver worked at least one minute,respectively. Rows 16-20 report results for trip-level outcomes.

Table 49

Var Coefficient SE tstat pval dof mean % change

1 Unique Sessions -0.013 0.026 -0.515 0.607 201 0.997 -1.3192 Riders w/ Sessions -0.002 0.006 -0.369 0.712 201 0.342 -0.6053 Unique Sessions Cond. >0 0.034 0.031 1.099 0.273 201 2.834 1.199

Note: This table shows regression results for rider-level demand outcomes. Regressions are estimated at the city leveland weighted by the number of riders in each city-week. Standard errors are clustered by city and stratified by thetercile of the number of riders in each city.

60

Page 62: Design and Analysis of Cluster-Randomized Field

Table 50

Var Coef. SE tstat pval dof mean % change

1 Total Earnings -1.940 6.113 -0.317 0.751 201 209.688 -0.9252 Organic Earnings -3.953 6.127 -0.645 0.520 201 192.500 -2.0543 Minutes Worked 2.725 8.734 0.312 0.755 201 697.208 0.3914 Incentive Earnings -0.161 0.816 -0.198 0.844 201 6.006 -2.6855 Surge Earnings -2.283 2.855 -0.800 0.425 201 5.228 -43.6656 Post-Commission Earn. -0.486 4.717 -0.103 0.918 201 161.280 -0.3017 Hourly Earnings -0.228 0.496 -0.459 0.647 201 16.219 -1.4048 Organic Earnings -0.174 0.447 -0.389 0.697 201 15.031 -1.1579 Utilization -0.004 0.009 -0.423 0.672 201 0.489 -0.75010 Incentive Hourly -0.039 0.046 -0.839 0.403 201 0.462 -8.34311 Surge Hourly -0.173 0.190 -0.910 0.364 201 0.459 -37.71512 Post-Commission Hourly -0.086 0.391 -0.221 0.825 201 12.423 -0.69513 Minutes Worked 2.725 8.734 0.312 0.755 201 697.208 0.39114 Min. Worked - Intensive -3.291 11.719 -0.281 0.779 201 1, 054.186 -0.31215 Did Work 0.004 0.003 1.368 0.173 201 0.662 0.62316 Completed Trips -0.129 0.227 -0.571 0.569 201 28.324 -0.45717 Mean Fare -0.316 0.248 -1.273 0.204 201 12.904 -2.45218 Trip Distance 0 0.031 0.016 0.987 201 6.624 0.00719 Trip Duration 5.533 7.730 0.716 0.475 201 972.181 0.56920 Trip Surge -0.007 0.010 -0.725 0.469 201 1.027 -0.721

Note: This table shows regression results for driver-level and trip-level outcomes. Regressions are estimated at the citylevel and weighted by the number of drivers in each city-week. Standard errors are clustered by city and stratified bythe tercile of the number of drivers in each city.

Appendix C City Level t-test

Since labor supply seems to be balanced in the pre-treatment period at the city level withoutany weighting, we can try a t-test to check for differences in the experiment period. Resultsare in Table 51. The minimum detectable effect size is about 9%.

Table 51

Mean Treated Mean Control T Stat P Value CI

542.70 520.61 0.88 0.38 (-27.33, 71.51)

Note: This table mean labor supply outcomes across cities in the control vs treatment group inthe post-treatment period. Cities are not weighted by the number of drivers. The minimumdetectable effect size for the t-test is about 9%.

Appendix D Quantile Regressions

As a robustness check we report quantile regression estimates using the methods in Chetverikovet al. (2016). These methods can help estimate differential effects across drivers and may be

61

Page 63: Design and Analysis of Cluster-Randomized Field

more robust to outliers. In this application, estimating the quantile effects can be performedby simply computing the quantile of the outcome variable within each city-by-week cell andtreating this computed quantile as the dependent variable in an OLS regression, using thesame right-hand side specification from our city-level models. Estimates along with confi-dence intervals are shown in Figures 20 through 25. We do not plot quantiles that havelittle to no variation across city-weeks. For example, the tenth percentile of the labor supplyis 0 across all city-weeks in the data and is not shown in Figure 20. As before, results areunderpowered and should be interpreted with caution. The point estimates would suggestfull-time drivers increased labor supply the most. The drivers with highest hourly gross earn-ings and utilization may have seen a larger drop in these metrics, though post-commissionearnings are roughly flat, perhaps offset by tip income. Changes in demand do not seem tovary much by quantile.

Figure 20

Note: This plot shows quantile regression estimates for labor supply. We exclude percentiles under 25% because below thisthreshold labor supply is zero for every city-week. Quantile regression estimates are computed as per Chetverikov et al.(2016).

62

Page 64: Design and Analysis of Cluster-Randomized Field

Figure 21

Note: This plot shows quantile regression estimates for hourly earnings. Quantiles for each city-week are estimated overonly drivers who worked in that week.

Figure 22

Note: This plot shows quantile regression estimates for organic hourly earnings. Quantiles for each city-week are estimatedover only drivers who worked in that week.

63

Page 65: Design and Analysis of Cluster-Randomized Field

Figure 23

Note: This plot shows quantile regression estimates for utilization. Quantiles for each city-week are estimated over onlydrivers who worked in that week.

Figure 24

Note: This plot shows quantile regression estimates for post-commission hourly earnings. Quantiles for each city-week areestimated over only drivers who worked in that week.

64

Page 66: Design and Analysis of Cluster-Randomized Field

Figure 25

Note: This plot shows quantile regression estimates for the number of unique sessions for each rider. A rider records aunique session when they open the app and search for a potential destination, whether or not they book and complete atrip. We exclude percentiles under 60% because below this threshold the number of unique sessions is zero for everycity-week.

65