Primary Outcomes (end points)
Primary outcome 1: Test performance
Research Question 1: To what extent does direct instruction for performance (DIP) versus instruction for independent learning (IIL) in weekly R-studio practicals differ in effect on students’ test performance?
Hypothesis 1: Students in the DIP condition differ on average in their test performance from students in the IIL condition (two-sided).
We formulate the alternative hypothesis as any difference between conditions in performance on the test. We see support for contradicting mechanisms in the literature that to our knowledge have not been tested before as randomized controlled trial in the setting and timeframe of a real course. To our knowledge, no research on instruction for skill-acquisition has been replicated in a randomized experiment in the timeframe and setting of an entire course. Direct instruction for performance such as modelling of a task generally shows benefits for learning, presumably by providing more efficient instruction (Leppink et al., 2014; Van Gog & Rummel, 2010). This effect might however disappear or even reverse in the natural setting of an eight-week course when compared with instructions for independent learning. For instance, 1) if students compensate for a lack of modelling through their behavior during and outside practicals because of a lower illusion of understanding (Baars et al., 2020; Dunlosky & Rawson, 2012; Paik & Schraw, 2013), 2) if students develop skills and behavior to learn independently (Glogger-Frey et al., 2017; Roodenrys et al., 2012; Weijers et al., 2023), or 3) if some students may have obtained sufficient expertise in R in previous courses to learn new analyses with less guidance (e.g. Kalyuga et al., 2012). We aim to investigate such possible mechanisms in the other hypotheses, but we have no definite expectation which of the potential effects will have a stronger influence on test performance in our natural setting of an 8-week course.
Primary outcome 2: In-class performance and experiences
Research Question 2: To what extent does DIP versus IIL in weekly R-studio practicals differ in effect on students’ in-class performance and experiences?
Support for direct instruction for performance during such practicals is found in instructional research, but mostly in controlled settings over a shorter period of time (Sweller et al., 2019; Van Gog & Rummel, 2010).
Hypothesis 2a: Students receiving DIP perform higher during the practicals than students receiving IIL.
We base this expectation on research demonstrating that direct instruction for new skills generally results in more efficient execution of a task, however mostly in controlled settings over a shorter period of time (Dart, 2022; Sweller et al., 2019; Van Gog & Rummel, 2010). We measure performance as progress in the practical, as described in Outcomes (Explanation).
We compare conditions on the following learning experiences, measured through a survey at the end of practical.
Hypothesis 2b1: Regarding Cognitive Load, we expect lower ratings of extraneous cognitive load in the DIP condition.
This would be supported by cognitive and socio-cognitive research stressing modelling for performance as an efficient way for constructing relevant schemas for performing a new task (Sweller et al., 2019; Van Gog & Rummel, 2010), in our study this task is weekly the new type of analysis.
Hypothesis 2b2: Regarding Invested Mental Effort, we test an alternative hypothesis of any difference between conditions (two-sided).
This outcome is often used for cognitive load in research on instructional effects showing differences between instructional conditions (Ouwehand et al., 2021; Szulewski et al., 2017; Van Der Wel & Van Steenbergen, 2018). In the original publication using this measurement, Paas (1992) however found no difference between conditions in a statistics-education setting. Students’ actually invested effort was concluded to be more closely related to motivation to invest effort than to the requirements of a task. We adopt his original interpretation in our expectation that investment of effort may differ in any direction, or be the same between our conditions as well, even if actual cognitive load demands for learning to perform the analysis in our conditions may differ.
Hypothesis 2b3: We expect higher perceived learning in the DIP condition.
Higher performance and lower perceived effort are related to higher perceptions of learning and expectations of ease of learning in problem-solving tasks (Baars et al., 2020; Kirk-Johnson et al., 2019). Although we hypothesize no direction for invested mental effort in our setting, we do expect lower perceived required effort for performing the analysis in terms of cognitive load, and hence higher subjective perceptions of learning, with direct instruction for performance.
Primary outcome 3: use of resources and out-of-class engagement
Research Question 3: To what extent does IIL versus DIP in weekly R-studio practicals differ in effect on students’ engagement in terms of preparation, time-investment, and use of available sources for learning support?
Regarding preparation before the practical, we expect the following signs of engagement:
Hypothesis 3a1: More students in the IIL condition prepare by watching the lecture before the practical.
Hypothesis 3a2: More students in the IIL condition prepare by turning in their preparation assignments.
Regarding time investment before and after the practical, we expect the following signs of engagement:
Hypothesis 3b1: Students in the IIL condition report higher time investment in the week before a practical.
Hypothesis 3b2: Students in the IIL condition report higher planned time investment for the week after a practical.
We state the following hypotheses regarding use of support sources:
Hypothesis 3c1: Students in the IIL condition use more available sources of support.
In our blended course, students have access to many different sources of instruction and learning support (worked examples, lecture slides, opportunities for asking questions and collaboration, answer model etc.). Use of support sources is a subcomponent of different conceptualizations of engagement and self-regulation for learning (Hands & Limniou, 2023; Pintrich, 2004; Pintrich et al., 1993; Wang et al., 2024), that we expect to be enhanced when the teacher provides instructions for independent learning instead of direct instruction for performance.
Hypothesis 3c1 tests a general aggregate of use of resources, as patterns of preferred support sources may differ between students. To specifically address the role of worked examples in instruction for independent learning (Chen et al., 2023; Van Gog & Rummel, 2010; Van Harsel et al., 2022), the following hypotheses are separately tested as well.
Hypothesis 3c2: More students in the IIL condition access the example code in the week before and after the practical.
Hypothesis 3c3: In the IIL condition, more students access the example code during the practical.
We also hypothesize that students in the IIL condition seek earlier for the first time access to:
Hypothesis 3c4 the example code,
Hypothesis 3c5: the slides,
Hypothesis 3c6: the answer model after its availability.
Following Alhazbi et al. (2024), we investigate not only total use but also timing as measurement of engagement through hypotheses 3c4-3c6.
Primary outcome 4: Differential learning effects (expertise reversal)
Research Question 4: To what extent can expertise reveal differential learning effects of DIP and IIL in weekly R-studio practicals?
Hypothesis 4: We hypothesize that students with lower prior ability in R benefit more from DIP than students with higher prior ability, visible in:
Hypothesis 4a: exam performance,
Hypothesis 4b: performance during the practical,
Hypothesis 4c: perceived extraneous load,
Hypothesis 4d: invested mental effort, and
Hypothesis 4e: perceived learning.
The Expertise Reversal effect states that with higher prior ability, direct instruction may be less beneficial for learning and performance as compared to learners with lower prior ability (Castro-Alonso et al., 2021; Kalyuga, 2007; Tetzlaff et al., 2025; Van Gog & Rummel, 2010). Such a compound effect is generally attributed to learner specific levels of cognitive load for a particular task, and observed in different types of skill acquisition (Kalyuga et al., 2012; Stambaugh, 2011).