Experimental Evidence on the Learning Impact of Generative AI

Last registered on April 17, 2025

Pre-Trial

Trial Information

General Information

Title
Experimental Evidence on the Learning Impact of Generative AI
RCT ID
AEARCTR-0015806
Initial registration date
April 14, 2025

Initial registration date is when the trial was registered.

It corresponds to when the registration was submitted to the Registry to be reviewed for publication.

First published
April 17, 2025, 7:13 AM EDT

First published corresponds to when the trial was first made public on the Registry after being reviewed.

Locations

Primary Investigator

Affiliation
Middlebury College

Other Primary Investigator(s)

PI Affiliation
Middlebury College

Additional Trial Information

Status
In development
Start date
2025-04-21
End date
2025-05-20
Secondary IDs
Prior work
This trial does not extend or rely on any prior RCTs.
Abstract
This study examines how access to generative artificial intelligence (AI) tools affects student learning outcomes in higher education. Using a randomized controlled experiment, we assign undergraduate students to complete learning tasks on unfamiliar academic topics either with or without access to generative AI tools. The experiment consists of two sessions approximately one week apart, allowing us to measure both immediate performance and longer-term knowledge retention. In the first session, participants complete a baseline assessment, engage in a structured learning exercise on an assigned topic, and produce a written analysis. Only participants in the treatment group are permitted to use generative AI during this process. In the second session, all participants complete knowledge retention assessments without access to AI. By comparing performance across conditions, we aim to identify causal effects of AI access on learning outcomes and explore potential mechanisms behind these effects. The findings will provide empirical evidence on an important question facing higher education as these technologies become increasingly prevalent.
External Link(s)

Registration Citation

Citation
Contractor, Zara and Germán Reyes. 2025. "Experimental Evidence on the Learning Impact of Generative AI." AEA RCT Registry. April 17. https://doi.org/10.1257/rct.15806-1.0
Experimental Details

Interventions

Intervention(s)
Intervention (Hidden)
This experiment investigates the causal impact of generative AI access on learning outcomes through a two-session laboratory study with undergraduate students. In Session One, participants are randomly assigned to either an AI-allowed or AI-forbidden condition and tasked with learning about one of three unfamiliar academic topics (Blockchain, Carbon Capture, or CRISPR). The session begins with participants completing a 5-question baseline knowledge assessment, followed by a 35-minute learning phase during which they research their assigned topic and complete a written analytical assessment of approximately 500 words. Participants in the treatment group have explicit permission to use generative AI tools during this learning phase, while those in the control group are instructed not to use these tools. After completing the written assessment, all participants take a post-learning assessment without access to external resources to measure immediate knowledge acquisition.

Session Two occurs approximately one week later and measures knowledge retention without any external resources. This session includes a follow-up written analytical task using a complementary prompt on the same topic and a 10-item multiple-choice test of factual knowledge. To incentivize effort, we implement a lottery-based compensation structure where participants earn tickets based on their performance on assessments, with each ticket giving them a chance to win monetary prizes.

The experimental design allows us to measure both the immediate effects of AI access on learning performance and its impact on longer-term knowledge retention. We collect detailed data on treatment compliance, assess written responses through blind evaluations by independent graders, and gather self-reported measures of AI experience and usage patterns. This allows us to examine not only whether AI access affects learning outcomes but also the potential mechanisms through which these effects might operate, such as changes in study behaviors, time allocation, and engagement with learning materials.
Intervention Start Date
2025-04-21
Intervention End Date
2025-05-20

Primary Outcomes

Primary Outcomes (end points)
1. Written assessment quality in Session One (as measured by blind grader evaluations)

2. Post-learning assessment performance in Session One (multiple-choice test)

3. Follow-up written assessment quality in Session Two (as measured by blind grader evaluations)

4. Knowledge retention test performance in Session Two (multiple-choice test)

5. Learning gain (difference between post-learning and baseline assessments)
Primary Outcomes (explanation)
Knowledge retention is measured as the score on a 10-item multiple-choice test administered in Session Two, while learning gain represents the difference between the post-learning assessment and baseline knowledge assessment scores.

Secondary Outcomes

Secondary Outcomes (end points)
1. Time allocation during the learning phase

2. AI usage patterns (for treatment group)

3. Computational text metrics (lexical diversity, word count, syntactic complexity)

4. Retention rate (ratio of Session Two to Session One performance)
Secondary Outcomes (explanation)
Time allocation measures will track how participants distribute their time across different activities (researching, writing, editing) during the learning phase. For the treatment group, we will analyze AI usage patterns including frequency and types of AI interactions. Computational text metrics will be derived from participants' written submissions using natural language processing methods.

Experimental Design

Experimental Design
This experiment uses a between-subjects design with random assignment. Undergraduate participants complete two sessions approximately one week apart. In Session One, participants learn about an unfamiliar academic topic and complete assessments. In Session Two, all participants complete retention assessments.
Experimental Design Details
This experiment uses a between-subjects design with random assignment to either an AI-allowed or AI-forbidden condition. Undergraduate participants complete two sessions approximately one week apart. In Session One, participants learn about an unfamiliar academic topic and complete assessments, with only the treatment group permitted to use generative AI. In Session Two, all participants complete retention assessments without AI access.
Randomization Method
Randomization done in office by a computer.
Randomization Unit
Individual participant
Was the treatment clustered?
No

Experiment Characteristics

Sample size: planned number of clusters
Not applicable (individual-level randomization)
Sample size: planned number of observations
~200-300 students
Sample size (or number of clusters) by treatment arms
100-150 students in AI-allowed condition, 100-150 students in AI-forbidden condition
Minimum detectable effect size for main outcomes (accounting for sample design and clustering)
With our minimum target sample size of 200 participants (100 per treatment arm), significance level = 0.05, and 80% power, the minimum detectable effect size is 0.40 standard deviations for main between-group comparisons of learning outcomes. For knowledge retention measured after one week, the literature on generation effects in learning suggests potentially larger effects (0.64 SD), allowing us to detect meaningful differences even with some attrition. Power calculations are based on two-sided t-tests.
IRB

Institutional Review Boards (IRBs)

IRB Name
Middlebury College IRB
IRB Approval Date
2025-04-10
IRB Approval Number
447
Analysis Plan

There is information in this trial unavailable to the public. Use the button below to request access.

Request Information

Post-Trial

Post Trial Information

Study Withdrawal

There is information in this trial unavailable to the public. Use the button below to request access.

Request Information

Intervention

Is the intervention completed?
No
Data Collection Complete
Data Publication

Data Publication

Is public data available?
No

Program Files

Program Files
Reports, Papers & Other Materials

Relevant Paper(s)

Reports & Other Materials