Experimental Design Details
Heterogeneity
We expect platforms to optimally set ad loads depending on 1) the elasticity of a user with respect to their ad load, 2) the network externalities that the user imposes on others, and 3) the value of targeting the user with ads. Based on this intuition, we will report the average ad load by:
Users with above vs. below median elasticity (predicted elasticity). We will predict the elasticity using our ads reduction intervention and demographics/baseline variables.
Users with above vs. below median network effects. Our measure of network effects will be the self-reported number of friends on Facebook, but we will also use other measures such as the number of posts produced or the composition of the feed.
Users with above vs. below value of ads. The value of ads is driven both by the probability that a user engages with the ad (e.g. clicks) and the value of the click for the advertiser. We will measure engagement through users’ actions (e.g., clicks) and the time they spend viewing ads. We also obtain a measure of the cost-per-click of targeting users with ads based on the predicted number of clicks from the Facebook Ads manager, targeting on gender and age.
Besides these variables, we will conduct more systematic approaches to understand which variables predict variability in ad loads.
First, we will conduct a variance decomposition exercise to test whether variation between individuals is a bigger source of variation than that of within individuals.
We will also analyze which variables predict cross-sectional differences in ad-loads. We will use OLS and machine-learning methods, including as predictors: baseline activity on and off the platforms, demographics, baseline content produced, baseline toxicity produced and consumed, predicted ad elasticity, measures of network effects interacted with ad engagement by users (the share of time allocated to ads versus other content by user), the percent of content that individuals see from friends vs from non-friends.
Lastly, we will analyze heterogeneous treatment effects using the same predictors and the Generalized Random Forest approach (Athey, Tibshirani, & Wager, 2019).
Empirical Analysis
To increase power, we will exploit both between and within variation, taking advantage that for most individuals we will have at least six weeks of baseline period without intervention and six weeks of intervention. In our main analyses we will run daily difference-in-difference two-way fixed effect regressions. We will also report specifications that are robust to the staggered nature of our treatment.
We will exploit that we have a large number of periods and hence will use Driscoll and Kraay (1998) standard errors to increase power. For those outcomes for which we do not have a long enough time dimension, we will also report standard errors clustered at the individual level. We will also report event-study estimates.
We will report two main specifications: 1) comparing the quantity reduction treatment arm with the passive control group, and 2) comparing the targeting reduction group with the replacement control group. We will also report a third version, which compares both control groups with each other. For some outcomes for which we do not have daily observations (e.g., valuation outcomes), we will report OLS regressions controlling for baseline outcomes, demographics, and enrollment and intervention start dates.
In terms of time frame, our main specification will include up to six weeks of intervention period. However, we will also report longer-run estimates.
In terms of attrition, we expect based on our previous work to have a low attrition rate (less than 15% over the main period of interest). We also expect this attrition rate to not be differential across treatment groups.
Based on our estimates, we will report “price” elasticities and diversion ratios.