Skip to main content

How supervision and educational supports impact medical students’ preparation for future learning of endotracheal intubation skills: a non-inferiority experimental trial

Abstract

Background

Professional education cannot keep pace with the rapid advancements of knowledge in today’s society. But it can develop professionals who can. ‘Preparation for future learning’ (PFL) has been conceptualized as a form of transfer whereby learners use their previous knowledge to learn about and adaptively solve new problems. Improved PFL outcomes have been linked to instructional approaches targeting learning mechanisms similar to those associated with successful self-regulated learning (SRL). We expected training that includes evidence-based SRL-supports would be non-inferior to training with direct supervision using the outcomes of a ‘near transfer’ test, and a PFL assessment of simulated endotracheal intubation skills.

Method

This study took place at the University of Toronto from October 2014 to August 2015. We randomized medical students and residents (n = 54) into three groups: Unsupervised, Supported; Supervised, Supported; and Unsupervised, Unsupported. Two raters scored participants’ test performances using a Global Rating Scale with strong validity evidence. We analyzed participants’ near transfer and PFL outcomes using two separate mixed effects ANCOVAs.

Results

For the Unsupervised, Supported group versus the Supervised, Supported group, we found that the difference in mean scores was 0.20, with a 95% Confidence Interval (CI) of − 0.17 to 0.57, on the near transfer test, and was 0.09, with a 95% CI of − 0.28 to 0.46, on the PFL assessment. Neither mean score nor their 95% CIs exceeded the non-inferiority margin of 0.60 units. Compared to the two Supported groups, the Unsupervised, Unsupported group was non-inferior on the near transfer test (differences in mean scores were 0.02 and − 0.22). On the PFL assessment, however, the differences in mean scores were 0.38 and 0.29, and both 95% CIs crossed the non-inferiority margin.

Conclusions

Training with SRL-supports was non-inferior to training with a supervisor. Both interventions appeared to impact PFL assessment outcomes positively, yet inconclusively when compared to the Unsupervised and Unsupported group, By contrast, the Unsupervised, Supported group did not score well on the near transfer test. Based on the observed sensitivity of the PFL assessment, we recommend researchers continue to study how such assessments may measure learners’ SRL outcomes during structured learning experiences.

Peer Review reports

Background

Medical education cannot keep pace with the rapid advancements of knowledge in today’s society [1].But it can develop professionals who can. Rather than continually refining curricula to integrate more content [2], medical educators could consider the type of professionals they are aiming to develop, and then use established principles from the learning sciences to design and deliver content toward that aim [3].A concept that can inform the development of instruction and assessment methods for creating adaptive professionals has been termed “preparation for future learning” (PFL). PFL is understood as a learners’ ability to select and learn from new resources (e.g., updated guidelines, continuing education materials, colleagues, the internet) and to use that learning to facilitate solving novel problems [4].While studies have established links between assessments of learners’ PFL outcomes and the use of certain instructional designs, only recently have researchers started to investigate which specific learning mechanisms are associated with improved PFL outcomes [5].

Where PFL has been investigated in the learning of statistics [6,7], and in diagnostic reasoning [8], researchers have framed it as a construct that can be assessed as a type of learning transfer. That is, when learners exhibit PFL successfully, they are described as having ‘transferred in’ their previous knowledge to learn from novel resources, and as having ‘transferred out’ this new learning to solve new, related problems [9].By contrast, assessments of ‘near transfer’ involve measuring how well learners apply their knowledge acquired in one situation (e.g., diagnosing a common respiratory condition) when performing another task with familiar surface details (e.g., diagnosing other, less common respiratory conditions) [10,11].Studies suggest that near transfer tests (i.e., applying knowledge toimmediately perform一个任务)和PFL评估(例如,应用诺里居民区上空dge tolearn about and then performa novel task) represent distinct transfer outcomes [12], with PFL assessments offering greater potential to understand how instructional approaches impact learners’ capacity for future learning [6,7,8].

In capturing how well participants learn from new resources, studies using PFL assessments have revealed unique benefits of instructional designs such as integrated instruction [8], contrasting cases [13], and productive failure [14,15].An analysis of these instructional approaches suggests that most emphasize allowing learners to struggle while learning, to experiment with their own learning strategies, and to experience meaningful task variation [4,8,16,17].These characteristics align with ‘core processes’ of successful self-regulated learning (SRL): setting well-defined goals, persisting in challenging experiences requiring significant time and effort, and developing one’s own, idiosyncratic knowledge structures [18].Evidence from medical education suggests that instruction designed to support learners to enact SRL core processes (e.g., a ‘SRL-support’, such as a list of task-specific goals from which learners can choose to set and pursue) result in improved immediate performance and retention of clinical skills [19,20].However, most studies have yet to clarify whether instruction including such SRL-supports also benefits learning transfer [18,20].

By comparing different ways of supporting SRL using a PFL assessment as the primary outcome, we aim to inform the curricular mapping and assessment practices of organizations dedicated to health professions training. In particular, this may benefit schools with curricula emphasizing self-regulating, ‘master adaptive’ learners [21,22], and accrediting bodies, such as the Liaison Committee on Medical Education [23] and Accreditation Council for Graduate Medical Education [24),现在包括标准要求明确teaching of ‘self-directed learning’. Given the resource constraints facing most schools, we conceptualize SRL as a shared responsibility between learner and supervisor [20], which means supervisors can be present through their design of SRL-supports, rather than through direct instruction. With this perspective, we asked the question: how does a supervisor’s presence (i.e., physically present, or not), and the presence of SRL-supports (present or not) impact participants’ performance on a near transfer test and a PFL assessment?

We expected that the near transfer test and PFL assessment outcomes associated with training that includes evidence-based SRL-supports would be non-inferior to the outcomes of training that includes direct instruction from a physically present clinical supervisor. We also expected that training with either SRL-supports or Supervision would lead to improved outcomes, beyond the non-inferiority margin, compared to a training condition without either. For all comparisons, we expected that a PFL assessment would be more likely to detect larger mean group differences, relative to a near transfer test.

Methods

Study setting and design

This study took place at the University of Toronto from October 2014 to August 2015. The ‘humans in research’ Ethics Board approved the study; all participants provided informed consent and received a small honorarium.

In this randomized controlled non-inferiority trial [25], we considered a simulation-based training environment in which a supervisor is present as the most dominant, and likely resource-intensive, approach in health professions education (HPE). We argue that training which includes SRL-supports without requiring a supervisor’s time would be no less effective educationally, and potentially advantageous in cost- and resource-effectiveness. Our study design followed a modified double transfer protocol [6], depicted in Fig.1

Fig. 1
figure 1

Study protocol showing participant flow across all learning sessions and assessments

Participants

Via email, we recruited participants from a pool of approximately 1000 medical students and residents. We ensured participants had minimal experience performing endotracheal intubation, by setting a maximum of 10 previous successful intubations on either a patient or simulator, which our four clinician teachers came to consensus on, and is well below the 50 attempts reportedly needed for proficiency [26].We had low response rates when recruiting from just one learner population, and consequently recruited novices from three populations: pre-clerkship students, clerkship students, and post-graduate year 1 (PGY1) Internal Medicine residents. All participants met the inclusion criteria and were randomly assigned to one of three groups, balanced by their academic year.

Based on previous studies using similar global rating scales, we expected a standard deviation of 0.70 units and set a non-inferiority margin of 0.60 units on the 5-point Likert scale [27].We argue that 0.60 units on the GRS has been shown to be educationally meaningful in similar research studies [27], and can represent the difference between senior and junior postgraduate trainees in practice [28].Assuming no difference between the SRL-support and Supervised conditions, we calculated that 17 participants per group would be required to be 80% sure that the lower limit of a one-sided 95% confidence interval will be above the non-inferiority limit of 0.60 [29].

Simulated procedural skill

Research has consistently shown that many learning mechanisms generalize to both motor skill and verbal learning [30].Given most research on PFL has focused on learning statistical and diagnostic reasoning concepts, we chose to extend that work to the domain of invasive procedural skills. We asked participants to perform four different variations of endotracheal intubation on the Laerdal® Airway Management Trainer: the Table-top, Supine, Left-lateral Decubitus (LLD), and Straddling positions used in our prior research (Fig.2) [31,32].

Fig. 2
figure 2

The four clinical variations of endotracheal intubation used in our study design:aNormalbSupinecLeft Lateral DecubitusdStraddling

Outcome measures

Simulated endotracheal intubation performance

We used a Global Rating Scale (GRS) developed specifically for endotracheal intubation, consisting of four subscales and an overall rating (Additional file1). We developed this tool previously by modifying a pre-existing GRS, and we also collected favourable validity evidence in the form of strong inter-rater reliability and positive correlations with established performance metrics [31].In the present study, we video recorded all relevant participant performances and sent them to a resident and fellow in Anesthesiology, who rated the videos independently and in a blinded fashion (i.e., unaware of participant identity, and assigned group). For each rater, we calculated the average score across the five component scales of the GRS, and then calculated an intra-class correlation coefficient (ICC) to assess inter-rater reliability. We then averaged the two raters’ GRS score, which we used in all analyses.

Rater orientation

During a rater orientation session, we used 12 videos selected to represent the different intubation variations. We selected example videos to represent the key GRS verbal anchors of ‘poor’, ‘competent’, and ‘clearly superior’ for each variation. The raters stopped between each video to compare ratings, discuss any disagreements, and reach consensus. We did not use the consensus scores when calculating the ICC but did include them in the remaining analyses.

Educational interventions

We developed our three educational interventions using the proposed dimensions of SRL [20]: supervision (present or absent) and SRL-supports (present or absent). We chose to study three groups to increase the practical relevance of our research, given a fourth group (supervised, unsupported) would have been artificial (i.e., an instructor told to actively not support participants), and would likely have altered our results in favour of our hypotheses. We piloted and refined our approach for the three groups using three participants per group (data not collected during piloting).

Unsupervised, unsupported group

Participants assigned to this group did not receive supports beyond those provided to all three groups (i.e., an instructional video, anatomical model, notepad). Thus, participants received sufficient supports for content knowledge, but they did not receive supports for how to set learning goals, how to sequence their practice, or how to select learning strategies.

Unsupervised, supported group

Participants assigned to this group practiced using supports designed to help them self-regulate core-processes of SRL. The supports consisted of (all in Additional file2): an explanation about variable and random practice schedules (i.e., alternating tabletop and supine versions randomly) to highlight the benefits of using a challenging schedule for organizing one’s practice [33,34], a list of process goals they could set based on previous research showing that orienting learners to the processes of performance leads to better skill retention [35], and two brief interviews (conducted by author JM) that prompted participants to frame their practice in ways supportive of learning transfer [16].In the first interview, which followed the second intubation attempt, participants reflected on how they wouldreplicatetheir intubation approach in future experiences. In the second interview, which followed the seventh attempt, participants reflected on how they wouldapplytheir learning in future experiences where patient or contextual factors varied. We did not record their responses in either interview. We chose the timing of these interviews based on pilot data.

Supervised, supported group

Participants assigned to this group received one-on-one training with one of four university-affiliated clinician teachers. The lead author (JM) facilitated a meeting between the instructors, during which they developed a SRL-supportive teaching plan consisting of: (i) explaining key concepts for an initial 10-min, including the associated equipment, how to prepare for the procedure, and demonstrating the skill [36], (ii) organizing a blocked, variable practice schedule of the 16 attempts, with participants completing four successful attempts of Table-top variation, four successful attempts of Supine variation, and following that same sequence a second time [33], (iii) asking questions frequently, shifting from providing concurrent, hands on feedback for the first attempt to providing terminal, hands off feedback about multiple attempts, which aligns with motor-learning principles [37,38], and (iv) debriefing participants for 10-min, asking them to verbally repeat the steps of a successful intubation [39].We note that while some might consider this “external regulation of learning”, we consider it supportive of SRL because participants were free to practice independently within the design set by the instructors.

Study procedure

Session 1: initial session

After completing a demographic questionnaire, participants watched a six-minute instructional video outlining the steps for a successful Table-top intubation on a patient [40].They then completed a baseline test on the simulator, performing a Supine intubation. After the baseline test, participants were oriented to the simulator, and given unlimited access to content-related educational materials: an oropharynx anatomical model, the instructional video, and a notepad and pen.

All participants experienced variable practice [33,34], as we ensured they would perform 16 successful attempts of Table-top and Supine intubations (albeit in different sequences, depending on their assigned group). A successful attempt involved placing the endotracheal tube so both lungs could be inflated through bag-valve ventilation. The first session ended as each participant finished their 16th attempt (approximately 1–1.75 h per participant).

Session 2: assessing near transfer and the PFL assessment

All participants returned independently 2 weeks later. No instructors or SRL-supports were available, meaning participants experienced an unsupervised and unsupported second session, requiring them to utilize their previous knowledge and experience to regulate their learning. Participants immediately performed the left-lateral decubitus variation of intubation on the simulator. While performing this variation required participants to position their bodies differently relative to the simulator, the required technique to perform the skill was arguably familiar, which we believe fulfills the common definition of a ‘near transfer test’. [6,11].

Next, we implemented a ‘learn-then-perform’ PFL assessment, which involves participants studying a resource containing new information, and then using that information on a subsequent performance-based assessment [8,41].Our participants received 30 min to read an article explaining different variations of intubation (some they had practiced, some not) [32], and to then practice these variations on the simulator. The reading was succinct and provided illustrations for six different endotracheal intubation variations. Of these six, we used the Straddling variation for the PFL assessment because this technique required significant motor skill transformations compared to those the participants learned initially (Fig.2). By instigating such transformations, we expected that the skill would require new learning for the participants, and thus would require them to potentially demonstrate PFL: having to reconsider their equipment, to reconsider how it can be used to perform the novel approach to the procedure, and to use problem-solving processes they had not yet experienced in order to perform well in this related, yet novel condition.

Data analysis

For the baseline, near transfer and PFL assessments, participants completed two attempts, and we averaged their GRS scores across attempts and across raters to generate a more stable estimate of performance.

Participants’ performances on the two transfer tests would not be independent; thus, we conducted separate mixed effects analyses of covariance (ANCOVAs), with baseline test as the covariate, transfer test (near and PFL) as the within-subjects factor, and either Supervision or SRL-support as the between-subjects factor. We tested whether our data violated any assumptions underlying the ANCOVA model; our inspections of normality and homogeneity of variance suggested we could proceed with our planned analyses. Further, participants’ baseline scores correlated positively with both their near transfer test scores (r = 0.30,p = 0.03) and their PFL assessment scores (r = 0.20,p = 0.16). We calculated the 95% confidence interval on the difference between group means on each test using an established procedure [42].

Results

One participant from the Unsupervised, Supported group dropped out of the study. Participant demographic data are recorded in Table1。在所有的测试中,我们计算平均沉降问题res ICC = 0.67, representing acceptable inter-rater reliability. All GRS data are reported below as mean ± standard deviation.

Table 1 Participant demographics and baseline test data

When comparing the Unsupervised, Supported group (N = 17, near transfer = 3.34 ± 0.68, PFL = 3.31 ± 0.66) to the Supervised, Supported group (N = 18, near transfer = 3.54 ± 0.45, PFL = 3.40 ± 0.64), we found that the difference in mean scores was 0.20 (95% CI of − 0.17 to 0.57) on the near transfer test, and was 0.09 (95% CI of − 0.28 to 0.46) on the PFL assessment.

For the Unsupervised, Unsupported group, (N = 18, near transfer = 3.56 ± 0.53, PFL = 3.02 ± 0.74), we found that the difference in mean scores compared to the Supervised, Supported group was 0.02 (95% CI of − 0.35 to 0.39) on the near transfer test and was 0.38 (95% CI of 0.001 to 0.74) on the PFL assessment. When comparing the Unsupervised, Unsupported to the Unsupervised, Supported group, we found that the difference in mean scores was − 0.22 (95% CI of − 0.58 to 0.15) on the near transfer test and was 0.29 (95% CI of − 0.081 to 0.66) on the PFL assessment.

In summary, none of the observed differences in mean scores exceeded the 0.60 non-inferiority margin. When considering the 95% CIs, two observed comparisons did produce ranges that crossed the non-inferiority margin: both SRL-supported groups (SRL-supports alone, and with a Supervisor) compared to the Unsupervised, Unsupported group on the PFL assessment. Notably, for the near transfer test, the Unsupervised, Supported group scored lower than, and just within the non-inferiority margin, relative to the other two groups.

Discussion

We expected that the near transfer test and PFL assessment outcomes of simulated intubation skills training with SRL-supports present would be non-inferior to the outcomes of training with direct instruction from a clinical supervisor. Our analyses showed that none of the differences in mean scores exceeded the non-inferiority margin, confirming our hypothesis. Thus, we showed that training using evidence-based SRL-supports is not inferior to the standard training using direct instruction from a clinical supervisor.

Further, we expected that a PFL assessment would be more likely to detect any differences that exceeded our chosen non-inferiority margin, relative to a near transfer test. Strictly using the non-inferiority margin, only the PFL assessment produced scores aligned with our expectations, producing 95% CIs that suggest potential, yet inconclusive benefits of training with either SRL-supports or Supervision relative to training that did not include them. For the outcomes of the near transfer test, however, we did observe that the Unsupervised, Supported group scored lower than the other two groups, and just within the non-inferiority window.

We believe the magnitude of the observed mean group differences permit discussing the modest implications of our findings for how medical educators prepare learners for future learning, for how we might assess SRL skills in HPE, and for how to refine our working definition of ‘educational support' for SRL. Below we consider how our findings might inform additional replication and extension studies.

Sensitivity of PFL assessments to conditions supporting SRL

Similar to other studies in HPE [8], and statistics education [7], our PFL assessment produced outcomes in the direction of our hypothesized benefits of SRL-supports and Supervision compared to when they were absent. By contrast, the near transfer test produced mixed findings that misalign with our expectations, especially suggesting non-inferiority of training that includes no SRL supports when trainees are unsupervised. Medical educators currently do not systematically measure learners’ capacity to learn from and within daily practice, and our findings suggest they may benefit from using dynamic assessments, like PFL assessments, to measure this important outcome associated with adaptive expert clinicians [21,43,44].

The scholars who conceptualized PFL suggested learners become better prepared for future learning after experiencing instructional designs supporting their acquisition of conceptual knowledge (i.e., knowledge about facts and principles underlying the task). Most instructional approaches studied in previous research on PFL emphasized allowing learners to fail and to experience meaningful task variation [4,8,16,17].解释我们的研究结果与此镜头,SRL-supports may prompt learners to experiment with using strategies they may not have otherwise considered (i.e., a form of guided discovery), which may contribute to their ability to learn and perform on the PFL assessment. Practically, our findings suggest that medical educators at institutions that have formally scheduled time for SRL into curricula will want to ensure learners receive some supports on how to regulate their learning effectively during that time, rather than leaving them both unsupervised and unsupported [20,45].Notably, despite the expected sensitivity of the PFL assessment, the near transfer test produced scores implying the unsupervised, supported group performed lower than the other two groups. While this finding replicates other studies suggesting near transfer and PFL assessments measure different constructs [6,7,15], it also suggests the selected supports for SRL could be further optimized in future studies of near transfer outcomes.

Defining educational supports for SRL

When developing the supports for our unsupervised, supported group, we chose to use lists of process goals to focus goal-setting, multiple intubation scenarios to introduce variability during practice, interviews to focus learners on replicating and applying learned principles, and guides on how to organize practice sequences. In blending these various SRL-supports, we cannot discern how each contributed to participants’ learning outcomes. However, there is evidence in education research to suggest that a multi-dimensional approach is needed to more effectively support SRL [14,46].指导未来的工作,我们提出一个广泛的目前rization of SRL-supports: content-related supports (e.g., the instructional video), practice schedule supports (e.g., organizing practice in a random sequence), and supports of core SRL processes (e.g., such as goal-setting lists).

Based on this study, we define aneducational support for SRLas a human or material resource that guides learners to relate their learning strategies to the conceptual and/or procedural knowledge underlying the target task. Future work will inevitably refine this definition, as educators and researchers identify, implement, and test SRL-supports in their instructional designs [19].Presently, our findings suggest learners’ future learning might benefit most through a combination of SRL-supports targeting core processes, like goal-setting, and through targeted direct supervision (e.g., helping a learner to set a study schedule, and to consider how to vary content as they study). By contrast, our findings suggest our chosen complement of SRL supports did not benefit learners’ near transfer performance.

Limitations

Although each independent SRL-support we manipulated in the relevant experimental conditions has supporting evidence from previous studies, we acknowledge that using them in combination prevents us from clarifying the unique contributions of each to our findings. Future studies will benefit from collecting behavioural manipulation check [47] data to better understand how participants use available supports, or, alternatively, invent their own. Doing so would permit an understanding of whether results might be explained by participants’ adherence to their assigned intervention (or lack thereof) and would also permit a pragmatic understanding of which SRL-supports learners capitalize on and use.

在使用插管技能作为我们实验的重点erimental study, we acknowledge that some of the scenarios are uncommon clinically (e.g., intubating a patient in the straddling pose). However, we chose intubation as a case given the multiple possible realistic variations of this procedure, which lends to use in a double transfer protocol. Further, we note that for this experiment, our training protocol was not designed to ensure participants developed skill mastery, and consequently note that further research is needed on how to integrate the lessons learned into practical instructional designs.

Conclusion

我们的研究表明,提供SRL-supports (i.e., instructors or materials) during initial practice represents another educational intervention that leads to modest yet potentially educationally meaningful improvements in performance on a PFL assessment. These improvements were non-inferior compared to a Supervised training condition. Thus, our work represents the start of a conceptual link between SRL core processes and PFL assessments that we believe researchers can capitalize on further to understand the most effective ways to support learners’ SRL and to prepare them for future learning. When designing ‘self-directed learning’ time in many medical school curricula, we recommend that educators consider how they might integrate SRL-supports at the session level, using our definition above as a guide. We also challenge medical educators to consider how they can integrate PFL assessments into their programmatic assessments, which may provide a more complete picture of learners’ SRL outcomes in structured learning scenarios.

Availability of data and materials

The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request. They have not been shared on a public database.

Abbreviations

ANCOVA:

Analysis of Covariance

ICC:

内部类相关系数

GRS:

Global Rating Scale

HPE:

Health Professions Education

LLD:

Left-Lateral Decubitus

PGY:

Post-Graduate Year

PFL:

Preparation for Future Learning

SRL:

Self-Regulated Learning

References

  1. 塔基斯M,Scardamalia M. Doctors’ perspectives on their innovations in daily practice: implications for knowledge building in health care. Med Educ. 2008;42(10):975–81.

    ArticleGoogle Scholar

  2. Hawick L, Cleland J, Kitto S. Getting off the carousel: exploring the wicked problem of curriculum reform. Perspect Med Educ. 2017;6(5):337–43.

    ArticleGoogle Scholar

  3. Kulasegaram KM, Martimianakis MA, Mylopoulos M, Whitehead CR, Woods NN. Cognition before curriculum: rethinking the integration of basic science and clinical learning. Acad Med. 2013;88(10):1578–85.

    ArticleGoogle Scholar

  4. Bransford JD, Schwartz DL. Rethinking transfer: a simple proposal with multiple implications. In: Review of Research in Education; 1999. p. 1–42.

    Google Scholar

  5. Loibl K, Roll I, Rummel N. Towards a theory of when and how problem solving followed by instruction supports learning. Educ Psychol Rev. 2017;29(4):693–715.

    ArticleGoogle Scholar

  6. Schwartz DL, Bransford JD, Sears D. Efficiency and Innovation in Transfer. In: Transfer of Learning from a Modern Multidisciplinary Perspective. Greenwich: Information Age Publishing; 2005. p. 1–51.

  7. Schwartz DL, Martin T. Inventing to prepare for future learning: the hidden efficiency of encouraging original student production in statistics instruction. Cogn Instr. 2004;22(2):129–84.

    ArticleGoogle Scholar

  8. 塔基斯M,Woods N. Preparing medical students for future learning using basic science instruction. Med Educ. 2014;48(7):667–73.

    ArticleGoogle Scholar

  9. Lee H, Anderson J. Student learning: what has instruction got to do with it. Annu Rev Psychol. 2013;3(4):445–69.

    ArticleGoogle Scholar

  10. Salomon G, Perkins DN. Rocky roads to transfer: rethinking mechanism of a neglected phenomenon. Educ Psychol. 1989;24(2):113–42.

    ArticleGoogle Scholar

  11. Kulasegaram KM, Chaudhary Z, Woods N, Dore K, Neville A, Norman G. Contexts, concepts and cognition: principles for the transfer of basic science knowledge. Med Educ. 2017;51(2):184–95.

    ArticleGoogle Scholar

  12. Martin L, Schwartz DL. Prospective Adaptation in the Use of External Representations. Cogn Inst. 2009;27(4):370–400.

  13. Schwartz DL, Chase CC, Oppezzo MA, Chin DB. Practicing versus inventing with contrasting cases: The effects of telling first on learning and transfer. J Educ Psychol. 2011;103:759–75.

    ArticleGoogle Scholar

  14. Kapur M. Productive failure in learning math. Cogn Sci. 2014;38:1008–22.

    ArticleGoogle Scholar

  15. Steenhof N, Woods NN, Van Gerven PWM, Mylopoulos M. Productive failure as an instructional approach to promote future learning. Adv Heal Sci Educ. 2020;12:1–8.

    Google Scholar

  16. Broudy H. Types of knowledge and purposes of education. In: Schooling and the Acquisition of Knowledge. Abingdon: Routledge; 1977. p. 1–17.

  17. Lin X, Schwartz DL, Bransford J. Intercultural adaptive expertise: explicit and implicit lessons from Dr. hatano. Hum Dev. 2007;50(1):65–72.

    ArticleGoogle Scholar

  18. Sitzmann T, Ely K. A meta-analysis of self-regulated learning in work-related training and educational attainment: what we know and where we need to go. Psychol Bull. 2011;137(3):421–42.

    ArticleGoogle Scholar

  19. 库克哒,Aljamal Y, Pankratz说道VS, Farle Sedlack再保险y DR, Brydges R. Supporting self-regulation in simulation-based education: a randomized experiment of practice schedules and goals. Adv Heal Sci Educ. 2019;24(2):199–213.

    ArticleGoogle Scholar

  20. Brydges R, Manzone J, Shanks D, Hatala R, Zendejas B, Hamstra SJ, et al. Self-regulated learning in simulation-based training: a systematic review and meta-analysis. Med Educ. 2015;49(4):368–78.

    ArticleGoogle Scholar

  21. 塔基斯M,Brydges R, Woods NN, Manzone J, Schwartz DL. Preparation for future learning: a missing competency in health professions education? Med Educ. 2016;50(1):115–23.

    ArticleGoogle Scholar

  22. Cutrer WB, Miller B, Pusic MV, Mejicano G, Mangrulkar RS, Gruppen LD, et al. Fostering the development of master adaptive learners. Acad Med. 2017;92(1):70–5.

    ArticleGoogle Scholar

  23. Standards, Publications, & Notification Forms | LCME [Internet]. [cited 2019 Nov 19]. Available from:https://lcme.org/publications/

  24. Duval JF, Opas LM, Nasca TJ, Johnson PF, Weiss KB. Report of the SI2025 task force. J Grad Med Educ. 2017;9(6):11–57.

    ArticleGoogle Scholar

  25. Tolsgaard MG, Ringsted C. Using equivalence designs to improve methodological rigor in medical education trials. Med Educ. 2014;48(2):220–1.

    ArticleGoogle Scholar

  26. 中方通过ML, Maissan IM, hoek SE, Klimek M, Stolker RJ. Defining the learning curve for endotracheal intubation using direct laryngoscopy: a systematic review. Theatr Res Int. 2016;99:63–71.

    Google Scholar

  27. Brydges R, Nair P, Ma I, Shanks D, Hatala R. Directed self-regulated learning versus instructor-regulated learning in simulation training. Med Educ. 2012;46(7):648–56.

    ArticleGoogle Scholar

  28. Stroud L, Herold J, Tomlinson G, Cavalcanti RB. Who You know or what You know? Effect of examiner familiarity with residents on OSCE scores. Acad Med. 2011;86:S8–11.

    ArticleGoogle Scholar

  29. Sealed Envelope Ltd. Power calculator for continuous outcome non-inferiority trial. 2012 [cited 2020 Dec 11]. Available from:https://www.sealedenvelope.com/power/continuous-noninferior/

    Google Scholar

  30. Schmidt RA, Bjork RA. New conceptualizations of practice: common principles in three paradigms suggest new concepts for training. Psychol Sci. 1992;3(4):207–17.

    ArticleGoogle Scholar

  31. Manzone J, Tremblay L, You-Ten KE, Desai D, Brydges R. Task- versus ego-oriented feedback delivered as numbers or comments during intubation training. Med Educ. 2014;48(4):430–40.

    ArticleGoogle Scholar

  32. Tesler J, Rucker J, Sommer D, Vesely A, McClusky S, Koetter KP, et al. Rescuer position for tracheal intubation on the ground. Resuscitation. 2003;56(1):83–9.

    ArticleGoogle Scholar

  33. Merbah S, Meulemans T. Learning a motor skill: effects of blocked versus random practice a review. Psychol Belg. 2011;51(1):15–48.

    ArticleGoogle Scholar

  34. Schmidt RA, Lee TD. Motor Control And Learning: A Behavioral Emphasis [Internet]. 4th ed: Human Kinetics; 2005. p. 537. Available from:http://books.google.com/books?hl=en&lr=&id=z69gyDKroS0C&pgis=1

  35. Brydges R, Carnahan H, Safir O, Dubrowski A. How effective is self-guided learning of clinical technical skills? It’s all about process. Med Educ. 2009;43(6):507–15.

    ArticleGoogle Scholar

  36. Sawyer T, White M, Zaveri P, Chang T, Ades A, French H, Anderson J, Auerbach M, Johnston L, Kessler D. Learn, see, practice, prove, do, maintain: an evidence-based pedagogical framework for procedural skill training in medicine. Acad Med. 2015;90(8):1025–33.

    ArticleGoogle Scholar

  37. Walsh CM, Ling SC, Wang CS, Carnahan H. Concurrent versus terminal feedback: it may be better to wait. Acad Med. 2009;84(SUPPL. 10):S54–7.

    ArticleGoogle Scholar

  38. Xeroulis GJ, Park J, Moulton CA, Reznick RK, LeBlanc V, Dubrowski A. Teaching suturing and knot-tying skills to medical students: a randomized controlled study comparing computer-based video instruction and (concurrent and summary) expert feedback. Surgery. 2007;141(4):442–9.

    ArticleGoogle Scholar

  39. Levett-Jones T, Lapkin S. A systematic review of the effectiveness of simulation debriefing in health professional education. Nurse Educ Today. 2014;34(6):e58–63.

    ArticleGoogle Scholar

  40. Kabrhel C, Thomsen TW, Setnik GS, Walls RM. Videos in clinical medicine. Orotracheal intubation. N Engl J Med. 2007;356(17):e15.

    ArticleGoogle Scholar

  41. Chaudhary ZK, Mylopoulos M, Barnett R, Sockalingam S, Hawkins M, O’Brien JD, Woods NN. Reconsidering basic: integrating social and behavioral sciences to support learning. Acad Med. 2019;94(11S):S73–8.

    ArticleGoogle Scholar

  42. Lane DM. Confidence Interval on the Difference Between Means [Internet]. [cited 2020 Dec 11]. Available from:http://onlinestatbook.com/2/estimation/difference_means.html

  43. Mylopoulos RG. How student models of expertise and innovation impact the development of adaptive expertise in medicine. Med Educ. 2009;43(2):127–32.

    ArticleGoogle Scholar

  44. Mylopoulos RG. Putting the expert together again. Med Educ. 2011;45(9):920–6.

    ArticleGoogle Scholar

  45. Brydges R, Butler D. A reflective analysis of medical education research on self-regulation in learning and practice. Med Educ. 2012;46(1):71–9.

    ArticleGoogle Scholar

  46. Schwartz DL, Bransford JD. A time for telling. Cognition and instruction. 1998;16(4):475–522.

  47. Hauser DJ, Ellsworth PC, Gonzalez R. Are manipulation checks necessary?. Front Psychol. 2018;9:998.

Download references

Acknowledgements

The authors wish to thank the Royal College of Physicians and Surgeons of Canada for funding this project through their Medical Education Research Grant competition, and for continuing to fund medical education research. We wish to thank Drs. Stella Ng, Martin Tolsgaard, and Jeffrey Cheung for completing meaningful reviews of previous versions of this manuscript.

Funding

Royal College of Physicians and Surgeons of Canada 2014–2015 Medical Education Research Grant.

Author information

Authors and Affiliations

Authors

Contributions

JM, MM, CR, and RB all contributed to the conceptualization of the study design and research question. JM completed the data collection sessions. All authors contributed to the analysis and interpretation of the data. JM and RB co-wrote the first draft of the manuscript, and MM and CR provided significant edits. All authors read and approved the final manuscript.

Corresponding author

Correspondence toRyan Brydges

Ethics declarations

Ethics approval and consent to participate

We received ethical approval to conduct our research from the University of Toronto’s Office of Research Ethics (Protocol Reference #30343).

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1.

The Global Rating Scale (GRS) used to rate simulated endotracheal intubation performance on all tests.

Additional file 2.

Unsupervised, supported group’s SRL-supports during session 1.

Rights and permissions

Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visithttp://creativecommons.org/licenses/by/4.0/。The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and Permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Manzone, J.C., Mylopoulos, M., Ringsted, C.et al.How supervision and educational supports impact medical students’ preparation for future learning of endotracheal intubation skills: a non-inferiority experimental trial.BMC Med Educ21, 102 (2021). https://doi.org/10.1186/s12909-021-02514-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI:https://doi.org/10.1186/s12909-021-02514-0

Keywords

  • capaSelf-regulated learning
  • Self-directed learning
  • Lifelong learning
  • Theories of learning
  • Learning transfer
  • Simulation
  • Instructional design