Pre-Experiment: Discrimination of Toddlers – a nationwide field experiment

Last registered on January 22, 2026

Pre-Trial

Trial Information

General Information

Title
Pre-Experiment: Discrimination of Toddlers – a nationwide field experiment
RCT ID
AEARCTR-0017647
Initial registration date
January 13, 2026

Initial registration date is when the trial was registered.

It corresponds to when the registration was submitted to the Registry to be reviewed for publication.

First published
January 22, 2026, 6:20 AM EST

First published corresponds to when the trial was first made public on the Registry after being reviewed.

Locations

Region

Primary Investigator

Affiliation
Max Planck Institute for Behavioral Economics

Other Primary Investigator(s)

PI Affiliation
University of Mannheim

Additional Trial Information

Status
In development
Start date
2026-01-19
End date
2026-10-01
Secondary IDs
Prior work
This trial does not extend or rely on any prior RCTs.
Abstract
This is the pre-experiment for our main study that will study racial discrimination against toddlers (independent of parental race). In our study (both the main experiment and this pre-experiment), we send emails to German child care facilities, mimicking parents asking for information on the application process and a free slot for their child. We use our own AI-generated images to signal both parental and children's race. Other than that, the emails are identical across treatment conditions, including names of writers. The main aim of this pre-experiment is to validate that the treatment works, that is, that recipients, in fact, see the image and take it into account when making a decision of whether, what, and how fast to answer. To do so, this study uses Black, and white parents with a toddler in-between them as the main treatment conditions. Our primary aim in this pre-experiment is to test whether responses towards Black and white parents differ, which would show that pictures are, in fact, viewed before making a decision on whether or not (and how) to respond.
External Link(s)

Registration Citation

Citation
Mill, Wladislaw and Felix Rusche. 2026. "Pre-Experiment: Discrimination of Toddlers – a nationwide field experiment." AEA RCT Registry. January 22. https://doi.org/10.1257/rct.17647-1.0
Sponsors & Partners

There is information in this trial unavailable to the public. Use the button below to request access.

Request Information
Experimental Details

Interventions

Intervention(s)
Intervention Start Date
2026-01-19
Intervention End Date
2026-02-20

Primary Outcomes

Primary Outcomes (end points)
We define three families of outcomes as our primary outcomes of interest:
- Response rate
- Helpfulness of responses. We follow Hermes et al. (2023) and code six variables:
1) slot offer
2) placement on waiting list
3) response length (above median length after removal of names, signatures, email histories)
4) helpful content (anything that helps the applicant in searching for a spot, e.g. a contact telephone number, a link to a registration portal or the center’s website, mentions of alternative institutions, or an application form.
5) encouraging: RAs had to answer whether the email was encouraging as a measure of tone
6) recommendation: RAs had to answer whether they would recommend the kita to a friend as a measure of tone.
- Response time
Primary Outcomes (explanation)

Secondary Outcomes

Secondary Outcomes (end points)
We may run heterogeneity regressions on the AfD vote share, urbanity (measured in 3 levels), provider type (public, church, other), and share of migrants in region.
Secondary Outcomes (explanation)

Experimental Design

Experimental Design
We send emails to 7,500 child care facilities in Germany. Each facility receives a single email with the following wording (the original email is in German):
_________
ENGLISH TRANSLATION
Dear Sir or Madam,
We are looking for a childcare place for [our baby's name], who is currently [2/5/8/12/18/24 months] old (see attached picture). As [my wife/husband] and I are planning to move to the area in [April/July/October] and to return to work, we are interested in a childcare place starting in [corresponding month] 2026.
Do you have any slots available? And how can we apply for a slot?
Thank you very much!
Sincerely,
[Mother's name / Father's name]
_________
ORIGINAL GERMAN VERSION
Sehr geehrte Damen und Herren,
wir sind auf der Suche nach einem Betreuungsplatz für [unsere/unseren Babyname], [der/die], derzeit [2/5/8/12/18/24 Monate] alt ist (siehe beigefügtes Foto). Da [meine/mein Frau/Mann ] und ich planen, im [April/Juli/Oktober] neu in die Gegend zu ziehen und wieder arbeiten zu gehen, sind wir an einem Betreuungsplatz ab [entsprechender Monat] 2026 interessiert.
Haben Sie noch einen freien Platz? Und wie können wir uns für einen Platz bewerben?
Vielen Dank!
Mit freundlichen Grüßen,
[Name Mother / Name Father]
_________

While we randomly vary various characteristics in the email above (as indicated), the email’s main variation lies in the race of the parents as signaled via images. Specifically, each email contains an AI-generated image that shows a toddler sitting in the middle and its parents on the left and right side respectively (the toddler’s face is visible, the parents’ is not). We use typical German first and last names in order to not signal race or other characteristics (e.g. SES) via names. The names have been validated via a Prolific survey with German-speaking participants. Using an algorithm to gradually vary toddlers’ race, we vary its race from 0 to 1 in steps of 0.25. Our main dimension of interest lies in varying parents’ race in a binary fashion with the following two conditions:

1. Black father, Black mother
2. white father, white mother

Within each strata, we randomly assign observations to one of the two conditions. In addition, within each strata and condition we assign one fifths of observations to one race category of the toddler (0, 0.25, 0.5, 0.75, 1). Similarly, within each strata and condition, we assign observations to the different email contents (like age of child, start date, etc.).

Our main hypothesis is that Black parents are treated different to white parents, specifically:
H1: Black parents have lower response rates than white parents.
H2: Black parents receive fewer helpful emails than white parents (both conditional and unconditional on receiving a response).
H3: Response times of Black parents are longer (both conditional and unconditional on receiving a response).

Our pre-study’s main aim is to validate that participants view images before making a decision of whether, when, and what to respond. We expect the above hypotheses to hold based on a previous study by Hermes et al. (2023), which varies immigrant/native status via names in an email to German child care facilities. As such, finding differences in response will be interpreted as validation that participants did, in fact, view images before making a decision.
Our study has the following exclusion restrictions: We start with a list of the universe of child care facilities in Germany. Prior to randomization we then, first, exclude all childcare facilities for which we could not identify an email address and, second, all those for which the identified email address was found to be unreachable or false according to Zerobounce, an email testing service. In the experiment, we will further exclude all email addresses that bounce back. Finally, prior to assigning facilities to treatment conditions, we randomly assign 7,500 out of 43,291 (or 17.32%) of child care facilities within each strata to participate in this experiment. The remainder is left for other parts of the experiments.

Methods: we will conduct regressions using a linear probability model. Standard errors will be clustered at the individual level given that each observation is independent and we have no repeated treatment. Our main regression will not include control variables.

Coding of Responses: We will code the helpfulness of responses using human RAs and LLMs on 500 responses. If interrater-reliability is comparable in-between humans to that in-between humans and the LLM, we will rate the remaining responses using the LLM only. If it is not comparable, humans will code the remaining responses.

_____________
References:
Hermes, H., Lergetporer, P., Mierisch, F., Peter, F., & Wiederhold, S. (2023). Discrimination in Universal Social Programs? A Nationwide Field Experiment on Access to Child Care, CESifo Working Paper No. 10368 CESifo Munich.
Experimental Design Details
Not available
Randomization Method
Stratified randomization (block randomization) is done using R.
We stratify treatment by state (16 states), urbanity (large city, middle city, other), AfD vote share (median split within state), and provider type (church/public/other). We have a total of potential 288 strata. Because some strata do not exist within a given state (e.g. in city states there is only one level of urbanity), we end up with a total of 227 strata.
Randomization Unit
Child care center level.
Details: We randomly assign units (child care facilities) within a single strata to one of the two treatment conditions. Within strata, we ensure that the same number of units is assigned to each of the conditions (with minor differences due to indivisibility of the number of units by two). In addition, we randomly assign other email characteristics within strata, including age and gender of child, requested start date, date of move, and gender of the emailing person (father or mother). In addition, within each strata and condition we randomly assign one fifths of observations to one race category of the toddler (0, 0.25, 0.5, 0.75, 1).
Was the treatment clustered?
No

Experiment Characteristics

Sample size: planned number of clusters
7,500 participating child care centers (we expect a reduction due to emails bouncing back)
Sample size: planned number of observations
We include 7,500 participating institutions, but expect this number to drop due to false or unreachable email addresses, that is, emails bouncing back. These will be removed from the data after the intervention is over. To avoid loss of too many, we checked the availability of email-addresses via Zerobounce prior to the experiment and excluded those flagged as invalid.
Sample size (or number of clusters) by treatment arms
The trial includes two main treatment arms within each of the strata.
Minimum detectable effect size for main outcomes (accounting for sample design and clustering)
Regarding our primary outcome (response rate), the study by Hermes et al. (2023) found a native response rate of 71% and a native-migrant gap of 4.4pp. With 80% power and a confidence interval of 95%, our study would require a sample size of 2,584. To ensure that we detect even substantially smaller effect sizes, we roughly tripled the required sample size. Our 7,500 observations allow us to identify gaps of around 2.56pp with the same power and confidence interval.
IRB

Institutional Review Boards (IRBs)

IRB Name
Ethics Committee of the University of Mannheim
IRB Approval Date
2025-12-18
IRB Approval Number
EC Mannheim 95/2025