Abstract
Large Language Models (LLMs) have seen expanding application across domains, yet their effectiveness as assistive tools for scientific writing—an endeavor requiring precision, multimodal synthesis, and domain expertise—remains insufficiently evaluated. We examine the potential of LLMs to support domain experts in scientific writing, with a focus on abstract composition. We design an incentivized randomised trial with a hypothetical conference set up where participants with relevant expertise are segregated into authors and reviewers. Inspired by methods in behavioral science, our novel incentive structure encourages participants to produce high-quality outputs. Authors edit original (control) or their AI-generated (treatment) abstracts of published research from top-tier conferences. Reviewers evaluate if the edited abstract provides adequate justice to the research presented in the original abstract.