Physician Responses to Patient Use of Generative AI for Self-Diagnosis: A Field Experiment

Last registered on January 06, 2026

Pre-Trial

Trial Information

General Information

Title
Physician Responses to Patient Use of Generative AI for Self-Diagnosis: A Field Experiment
RCT ID
AEARCTR-0017590
Initial registration date
January 05, 2026

Initial registration date is when the trial was registered.

It corresponds to when the registration was submitted to the Registry to be reviewed for publication.

First published
January 06, 2026, 7:24 AM EST

First published corresponds to when the trial was first made public on the Registry after being reviewed.

Locations

There is information in this trial unavailable to the public. Use the button below to request access.

Request Information

Primary Investigator

Affiliation

Other Primary Investigator(s)

PI Affiliation
Boston University
PI Affiliation
Boston University
PI Affiliation
China Europe International Business School

Additional Trial Information

Status
On going
Start date
2025-11-20
End date
2026-02-22
Secondary IDs
Prior work
This trial does not extend or rely on any prior RCTs.
Abstract
This study examines how physicians respond when patients share diagnostic suggestions during online medical consultations.
External Link(s)

Registration Citation

Citation
Burtch, Gordon et al. 2026. "Physician Responses to Patient Use of Generative AI for Self-Diagnosis: A Field Experiment." AEA RCT Registry. January 06. https://doi.org/10.1257/rct.17590-1.0
Experimental Details

Interventions

Intervention(s)
Intervention Start Date
2025-11-20
Intervention End Date
2026-01-20

Primary Outcomes

Primary Outcomes (end points)
Physician decision quality and decision-making processes during online consultations
Primary Outcomes (explanation)
Decision quality is measured by the extent to which physicians’ diagnostic assessments and recommended actions align with the ground-truth diagnosis and the clinically desirable care plan (e.g., diagnosis, recommended treatment, or the need for escalation to in-person care). Decision-making processes are measured using behavioral and textual indicators, including response time, hesitation or delay in acceptance, and the informational content and structure of physicians’ written responses.

Secondary Outcomes

Secondary Outcomes (end points)
Secondary Outcomes (explanation)

Experimental Design

Experimental Design
Physicians are randomly assigned to one of three conditions that differ only in the content of the initial inquiry. After the initial inquiry, GenAI-based standardized patients conduct follow-up conversations with physicians. All patient profiles are otherwise held constant, and physicians were randomly assigned to one of three conditions: a control group, an AI-diagnosis condition, and a friend of a friend's recommendation condition.
Experimental Design Details
Not available
Randomization Method
Randomization conducted in the office by a computer.
Randomization Unit
Physician
Was the treatment clustered?
No

Experiment Characteristics

Sample size: planned number of clusters
Between 3 and 10 multiples of 96 physicians (i.e., 288–960 physicians).
Sample size: planned number of observations
Between 3 and 10 multiples of 96 physicians (i.e., 288–960 physicians).
Sample size (or number of clusters) by treatment arms
Between 3 and 10 multiples of 32 physicians for each of the three groups: the control group, the AI-diagnosis disclosure group, and the human-source diagnosis disclosure group.
Minimum detectable effect size for main outcomes (accounting for sample design and clustering)
IRB

Institutional Review Boards (IRBs)

IRB Name
the CEIBS Research Committee and Institutional Review Board
IRB Approval Date
2025-09-12
IRB Approval Number
699-063