Self-Selection in Financial Literacy Education

Last registered on September 04, 2025

Pre-Trial

Trial Information

General Information

Title
Self-Selection in Financial Literacy Education
RCT ID
AEARCTR-0016395
Initial registration date
July 30, 2025

Initial registration date is when the trial was registered.

It corresponds to when the registration was submitted to the Registry to be reviewed for publication.

First published
August 01, 2025, 2:29 PM EDT

First published corresponds to when the trial was first made public on the Registry after being reviewed.

Last updated
September 04, 2025, 4:40 PM EDT

Last updated is the most recent time when changes to the trial's registration were published.

Locations

There is information in this trial unavailable to the public. Use the button below to request access.

Request Information

Primary Investigator

Affiliation
UC Berkeley

Other Primary Investigator(s)

PI Affiliation
UC Berkeley
PI Affiliation
UC Berkeley

Additional Trial Information

Status
In development
Start date
2025-09-04
End date
2026-10-01
Secondary IDs
Prior work
This trial does not extend or rely on any prior RCTs.
Abstract
Persistent disparities in financial literacy contribute to income and wealth inequality. This project investigates how individuals self-select into financial education and explores the welfare impacts of such selection. We pair rich administrative data with a lab experiment that incentivizes participation in financial education among individuals with different levels of demand. To examine how different types of education material influence take-up, we experimentally vary the format and perceived difficulty of the education material.
External Link(s)

Registration Citation

Citation
Shen, Elaine, Junru Lyu and Laila Voss. 2025. "Self-Selection in Financial Literacy Education." AEA RCT Registry. September 04. https://doi.org/10.1257/rct.16395-3.0
Experimental Details

Interventions

Intervention(s)
We will conduct a lab experiment to study selection patterns in financial education. Participants are offered access to different types of education material which are designed to improve their financial decision-making ability. We study how demand (willingness to pay) for education varies with baseline ability and with the format and perceived difficulty of the education material. To better understand the role of confidence in education investment decisions, we also elicit each participant’s perceived baseline ability and the perceived benefits of education.
Intervention Start Date
2025-09-04
Intervention End Date
2026-10-01

Primary Outcomes

Primary Outcomes (end points)
We measure how demand for different types of educational materials vary with an individual's starting level of financial knowledge. We measure an individual’s financial decision-making ability, demand for financial education (willingness to pay), and the treatment effect of education (or no education).
Primary Outcomes (explanation)
To measure financial knowledge, we design scenarios where there is a financial decision that maximizes a participant's monetary payout in the experiment. We then measure the mistake, or the difference between the participant's choices in the scenario and the choice that would have maximized their payout. We construct our measure of financial knowledge as a rescaled function of the mistake. We base our methodology on Ambuehl, Bernheim, and Lusardi (2022), which uses the difference between two equivalently valued complex and simply-framed financial instruments as the outcome. We collect both a baseline and an endline financial decision-making score. We also measure the change in the score, or the treatment effect under the different versions of the education material and under the control condition of no education material. We conduct pilots on the relevant populations to check the difficulty of the questions and to make sure there is room for a treatment effect from the education material.

We use a multiple price list to measure demand (willingness to pay/willingness to accept) for different types of education. We exclude observations where individuals switch at multiple points in the multiple price list. As a robustness check, we will also conduct a version of the analysis using only the willingness to pay/accept elicitation for the first version of the education material presented to account for potential anchoring.

Secondary Outcomes

Secondary Outcomes (end points)
We also measure the confidence, perceived ability, and perceived benefit of the different versions of the education material by asking participants to predict their performance on the baseline and endline financial decision-making questions. We count an estimated score as being correct if it is within a reasonable range of the actual score. To understand the drivers of demand for the different types of education materials, we ask individuals to explain their preferences and classify the responses.
Secondary Outcomes (explanation)
We measure the confidence, perceived ability, and perceived benefit of the different versions of the education material by asking participants to predict their performance on the baseline financial decision-making questions. We also ask participants to predict their performance on the endline questions for each version of the education materials and for no education material. We count an estimate as being correct if it is within a reasonable range of the actual score. We also measure overconfidence using the difference between the participant's actual ability and their perceived ability. To understand the drivers of demand for the different types of education materials, we ask individuals to explain their preferences and classify the responses.

Experimental Design

Experimental Design
We will conduct an experiment where we offer participants different types of education material to identify selection patterns and measure treatment effects.
Experimental Design Details
Not available
Randomization Method
The treatment and control conditions will be randomly assigned by a computer. Treatment conditions will be assigned at the individual level. Within treatment status, the different versions of the education material will be randomly assigned at the individual level as well.
Randomization Unit
Treatment conditions will be randomly assigned at the individual level. A computer will choose a random subset of individuals to have one of their choices in the willingness to pay elicitation implemented. For the majority of participants, the treatment and control conditions will be randomly assigned. Within treatment status, the different versions of the education material will be randomly assigned at the individual level as well.
Was the treatment clustered?
No

Experiment Characteristics

Sample size: planned number of clusters
Treatment will not be clustered.
Sample size: planned number of observations
With our current funding we will target 1,000-1,200 responses, however the final sample size will ultimately depend on the attrition rate. Complete responses are defined as surveys that have been fully answered with only coherent willingness to pay responses and who have passed all relevant comprehension and attention checks. Incoherent willingness to pay responses include participants who switch multiple times on the multiple price list, participants who switch from right to left, or responses with willingness to pay elicitations that otherwise do not make sense. Incoherent willingness to pay responses will be coded as missing and excluded from the analysis. Responses that are flagged as potentially generated by artificial intelligence software (such as ChatGPT) or copied from external sources will be excluded from the analysis. We may use automated plagiarism detection tools and web page navigation tracking, among other methods, to identify such behavior. Our ideal sample size is 1,200-1,600 responses, and we will target larger sample sizes if we are able to obtain additional funding. Treatment will not be clustered. If we conduct a version of the experiment with UC Berkeley students, we will also ask about interest in a real personal finance class at UC Berkeley. For this population, we will exclude responses from participants who have already taken the class if their responses look different from other participants. We will also exclude participants who provide incoherent willingness to pay responses, which will contribute to attrition rates. Based on pilots of the survey on Prolific, we expect the attrition rate from incoherent willingness to pay responses to be approximately 5-15%, however the exact attrition rate may be higher or lower for different participant groups. We will also include a version of the analysis where we drop participants who have their education material choices implemented based on their willingness-to-pay.
Sample size (or number of clusters) by treatment arms
We will target 1,000-1,200 responses. The treatment-to-control ratio is 3:1, reflecting our design with three treatment arms and one control group. Our ideal sample size is 1,200-1,600 responses, but the final sample size will ultimately depend on the attrition rate and our ability to recruit enough participants to complete the survey. We may add additional versions of the education material with different wordings in future versions of the study to disentangle mechanisms. We will target larger sample sizes if we are able to obtain additional funding. Treatment will not be clustered.
Minimum detectable effect size for main outcomes (accounting for sample design and clustering)
We use results from a pilot study on Prolific to calculate the minimum detectable effect size, assuming a Type I error rate of 5%, 80% power, a mean score of 50 (out of 100), and a standard deviation of 30. To be conservative, we use a slightly higher standard deviation than what was observed in the pilot study. The treatment-to-control ratio is 3:1, reflecting our design with three treatment arms and one control group. We will target 1,000-1,200 responses. According to our power calculations, with a sample size of 800 the minimum detectable effect size is 10 percentage points (p.p.), meaning the treated group must improve by at least 10 p.p. more than the control group for the effect to be detectable. As the sample size increases, the minimum detectable effect decreases: with 1,000-1,200 observations, it drops to 9 p.p.; with 1,200-1,400 observations, it decreases further to 8 p.p.; and with approximately 1,600 observations, it reaches 7 p.p. We plan to examine the correlation between participants' willingness to pay for each treatment and their baseline score in the financial decision-making task, then test whether these correlations differ across treatments. Due to funding constraints that limit our sample size, we do not expect to be adequately powered to detect differences in the slope of the demand curve across the different versions of the education material at conventional significance levels. Based on our power calculations, with a sample of 1,000 participants, we can only detect differences in correlations of approximately 15–20 percentage points or more–that is, the correlation between the baseline financial literacy score and the willingness to pay would need to differ by at least 15–20 percentage points across the treatment arms for the differences to be statistically significant. We believe such large differences are unlikely since we highlight the similar nature of the education materials to the participants. We may be able to detect differences in demand for the different types of education materials when we bin the participant groups (for example by high and low baseline ability) or when pooling the AI and non-AI versions of the education material.
IRB

Institutional Review Boards (IRBs)

IRB Name
UC Berkeley Committee for the Protection of Human Subjects
IRB Approval Date
2024-09-27
IRB Approval Number
2024-03-17249