Skip to main content

Digital learning of clinical skills and its impact on medical students’ academic performance: a systematic review

Abstract

Background

The constraints imposed by the COVID-19 pandemic has led to the rapid development and implementation of digital methods for teaching clinical skills in medical education. This systematic review presents both the benefits, challenges, and effectiveness of this transition.

Methods

A systematic search of six electronic databases (SCOPUS, Medline, CINAHL, PsycINFO, ERIC & Informit) was conducted on 1st October 2023 and updated on 1st April 2024 to identify peer- reviewed articles, from 2019 onwards, which used any type of digital tool (online or otherwise) to teach clinical skills to medical trainees (undergraduate or postgraduate) and were published in English language. The primary outcome synthesised was the reported effectiveness of these digital tools in the development of clinical skills. Risk of bias of included studies was assessed using the Quality Assessment With Diverse Studies (QuADS) tool.

Results

Twenty-seven studies involving 3,895 participants were eligible for inclusion in this review. The QuADS quality assessment scores ranged from 22 to 35, indicating medium quality and thirteen of the studies were randomized trials. Overall, digital teaching of clinical skills demonstrated improved or comparable outcomes to in-person teaching. There was a beneficial effect of digital learning tools on assessment outcomes, with meta-analysis showing a mean difference increase of 1.93 (95% CI 1.22 to 2.64), albeit with a high amount of statistical heterogeneity I2 97%, P < 0.001. Digital clinical skills teaching also resulted in improved student satisfaction scores in many situations and was also shown in one study to be cost effective.

Conclusion

Teaching of clinical skills using digital tools is an important alternative to the traditional format of face-to-face delivery, which is resource intensive and difficult to implement during a pandemic. This review demonstrates their potential efficacy in improving education outcomes, student satisfaction and potentially reducing costs. However, the integration of traditional and innovative digital teaching methods appeared to provide the most comprehensive learning experience. Future research could focus on longitudinal studies to assess the long-term impact and efficacy of different digital and blended learning modalities on the acquisition of clinical skills and professional competencies.

Peer Review reports

Background

The COVID-19 pandemic precipitated a paradigm shift in medical education, necessitating a global re-evaluation of pedagogical strategies to sustain the progression of clinical skills acquisition, which is vital for patient-centred care [1,2,3]. Historically, the foundation of medical training has been the integration of theoretical knowledge with practical, experiential learning, achieved through direct patient interactions and simulated clinical scenarios [4]. This model is especially crucial for students navigating the transition from theoretical pre-clinical studies to the hands-on clinical environment, ensuring they acquire essential skills such as history-taking, physical examination techniques, and procedural competencies requisite for effective patient care and hospital placements [5].

Prior to COVID-19, some universities used digital learning for various reasons, including increased accessibility and user preference. The advent of the COVID-19 pandemic, accompanied by stringent physical distancing mandates and lockdowns, significantly disrupted traditional clinical education pathways. In response, numerous educational institutions swiftly transitioned to remote and digital platforms for clinical training [2, 6,7,8]. This shift towards online clinical skills learning leverages electronic technologies to foster clinical reasoning, communication skills, and other core medical competencies, employing digital media to deliver consistent educational content, introduce innovative instructional methods, and facilitate the documentation of student engagement and performance assessments [9]. Nonetheless, the transition introduces several challenges, including diminished practical exercise opportunities, constraints on teaching physical examination techniques, and hurdles in the comprehensive evaluation of clinical competencies [4]. These challenges have spurred concerns about the efficacy of online and blended learning models in adequately preparing students for their clinical roles [4].

As the landscape of medical education continues to adapt to the challenges posed by the pandemic, the experiences and lessons learned from this period of enforced pedagogical innovation are poised to shape future educational strategies. Although most education has shifted back to in-person learning, there are some adaptations that may improve traditional teaching. The primary goal remains to ensure that all students attain the necessary clinical competencies, regardless of the educational formats employed. This period of accelerated adaptation may herald the development of more robust and flexible teaching methodologies, enhancing the acquisition and refinement of clinical skills in preparation for the demands of post-pandemic healthcare environments [10].

Given the evolving nature of medical education in response to the challenges posed by the pandemic, it is crucial to systematically review the experiences and lessons learned during this period of enforced pedagogical innovation. Previous systematic reviews in this area predate the COVID-19 pandemic [11, 12]. The current dearth of recent literature in this area necessitates a systematic review on this topic. Therefore, this systematic review aims to examine the impact of digital and blended learning environments on medical students’ academic performance in clinical skills training following the COVID-19 pandemic. We also aim to evaluate the effectiveness of educational adaptations, specifically digital media tools such as online modules, instructional videos, and lecture recordings, in maintaining high standards of clinical skills education.

Methods

This systematic review was reported according to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines [13].

Search strategy

Six electronic databases comprising MEDLINE, SCOPUS, PsychInfo, CINAHL, ERIC, and Informit were searched on 1st October 2023 and the search was updated on 1st April 2024. Reference lists of included studies and relevant systematic reviews were also searched to identify other eligible studies not captured by the search strategy. Search terms used related to the concepts of clinical skills, clinical competency, medical education, and online learning. The full search strategies for each database are outlined in Supplement 1. Search results were imported into Covidence systematic review software (Veritas Health Innovation, Melbourne, Australia. Available at www.covidence.org) for screening.

Eligibility criteria

To be eligible for inclusion, studies had to involve Population: medical trainees (either undergraduate or postgraduate) where Intervention: any type of digital media or tool (online or otherwise) was used to teach clinical skills compared to any other teaching format, and academic performance was measured or assessed. Studies involving other healthcare professions, such as chiropractic students, residents, resident physicians, or allied health students, were excluded. Digital media included electronic resources, platforms, applications, and other interactive digital media that used digital technology to create, deliver, or manage educational content, for example instructional videos, lecture recordings, and other digital resources used to facilitate learning were eligible for inclusion. Virtual reality (VR) technologies were not included in our systematic review because they typically require face-to-face delivery and specialised equipment, which differs from the digital media tools we aimed to evaluate. Our review focused on digital learning modalities that are accessible remotely and can be used by students without the need for physical presence or specialized hardware. Mixed interventions, e.g. both face-to-face and digital components, were also included. Clinical skills included competencies such as history-taking, physical examination, communication skills, and clinical reasoning. Procedural skills and technical competencies were included if they involved direct patient care. Clinical skills learning was defined as development of clinical reasoning, communication skills, and other medical competencies e.g. professionalism, ethical decision-making, teamwork, cultural competence, and evidence-based practice. Studies without a Comparison group were excluded. As the focus of the review was on COVID-19 pandemic’s effect, the search was limited to studies conducted from 2019 onwards. Additionally, included studies had to be peer-reviewed journal articles published in the English language. Commentaries and studies that only reported the Outcome of satisfaction levels (i.e. where academic performance was not assessed) were excluded. We defined assessment of academic performance as measuring a student’s competence in applying theoretical knowledge to practical clinical tasks [14].

Data extraction

Using COVIDENCE systematic review software, two authors independently screened studies for inclusion. Any discrepancies were resolved by a third author. After the studies were selected, a data extraction template was used to extract required information. Extracted variables included study authors, year published, number of participants, as well as review-specific outcomes such as type of digital teaching tool used, clinical skills taught, and academic performance measures.

Data analysis and synthesis

Numerical data was reported as mean and standard deviation (SD) or median and interquartile range, and categorical data as number and percentage. Meta-analyses on studies were conducted using Review Manager 5 software (Cochrane collaboration). The means and standard deviations from each trial were identified for inclusion in the meta-analysis. We used the Mantel-Haenszel random‐effects model to account for potential variability in participant conditions between studies and to calculate the pooled estimates (mean difference) and 95% confidence intervals. We assessed for apparent inconsistency in our results by examining methodological and statistical heterogeneity. We evaluated methodological heterogeneity by considering similarities amongst the included studies in terms of study design, participants, interventions, and outcomes, and used the data collected from the full‐text reports. We assessed statistical heterogeneity by calculating the ChI2 test or I2 statistic, judging an I2 value of 50% and a Chi2 P value of 0.05 or less as indicating substantial statistical heterogeneity. For thematic analysis we read each study in depth to identify study variables. Two investigators independently coded the data, and discrepancies were resolved through discussion and consensus. A third reviewer was consulted when necessary to resolve any remaining differences. Extracted data was summarised narratively, taking into consideration the interventions reported, and primary and secondary outcomes relevant to clinical skills teaching of medical students.

Risk of bias

Risk of bias was assessed using the ‘Quality Assessment with Diverse Studies’ (QuADS) tool [15]. Two authors independently assessed risk of bias using the QuADS tool, reaching consensus on final scores through discussion. Discrepancies were resolved by a third reviewer. The QuADS tool was deemed suitable as it allows for assessment of a broad range of methodologies and it has been reported as having good inter-rater reliability and validity [15]. The tool has 13 criteria to assess study quality, each of which are scored between zero (not stated at all) and three (explicitly described) (see Supplement 2). To ensure consistency in the assessment of quality, for each reviewed paper, the 13 criteria scores were summed and expressed as a percentage of the maximum possible score. This approach allowed for comparison of quality across the different papers. Interpretation of the quality evidence involved classification of total scores into low (< 60%), medium (60–80%) or high (> 80%). Studies were not excluded based on their quality rating, though the significance of their findings were considered when reporting the results and drawing conclusions based on the findings of all the included studies.

Results

Search results and study characteristics

Of 1092 unique records identified from the search strategy, 1035 (94.8%) were excluded through title and abstract screening, leaving 57 for full-text review. Of these, 30 were excluded for a variety of reasons, see Fig. 1 for details. Twenty-seven studies which involved 3,895 participants were included in the final analysis [16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42]. Figure 1 illustrates the screening and study selection process.

Fig. 1
figure 1

PRISMA flow diagram of study screening and selection

The studies were published from a wide range of countries, five from the United Kingdom (UK) [17, 27, 30, 35, 37]; four from Germany [20,21,22, 36]; and two from the United States of America [23, 26]. Of the studies, twelve were randomized trials [17, 18, 20,21,22, 25, 27, 30,31,32, 35, 36]; four were prospective observational studies [23, 24, 28, 34]; four were mixed methods studies [29, 39, 41, 42] and the remainder were a variety of study designs. The interventions were also varied, the most common formats used were videos (n = 12) and online modules/ courses (n = 6). Table 1 summarises the characteristics of the included studies.

Table 1 Characteristics and outcomes of included studies

Effectiveness of digital learning tools

The included studies demonstrated that online and digital learning interventions could enhance acquisition and enhancement of clinical knowledge and skills across various domains, see Fig. 2. Online modules and videos significantly improved clinical knowledge and specific skill sets such as diagnostic imaging and cardiac auscultation [16, 18, 24, 26]. Video-based learning emerged as a particularly effective tool in enhancing clinical knowledge, though it did not significantly impact anatomical knowledge acquisition. Its perceived advantages in presentation comprehensibility and engagement highlight the potential for multimedia resources to enrich the learning experience [20]. Similarly, animated media and video-based interventions showed improved adherence to correct algorithms and competency in handling e-patient scenarios, demonstrating the effectiveness of dynamic visual content in medical education [21, 22]. Overall video tools showed a beneficial effect with a mean difference of 1.64 (95% CI 0.22 to 3.06), while interactive modules showed a beneficial effect with a mean difference of 2.27 (95% CI 1.28 to 3.25). These results are summarised in Fig. 2, which also demonstrates a high amount of statistical heterogeneity in the results.

Fig. 2
figure 2

Forest plot of digital learning tools effectiveness by intervention

Studies also explored the impact of learning modalities on specific skills such as cardiac auscultation and practical skills like suturing, finding that while certain interventions led to improvements in proficiency and confidence, the overall effectiveness varied across different competencies and learning outcomes [26, 36]. High dropout rates in some studies, posed challenges in accurately measuring the effectiveness of the interventions [37]. Plackett [27] reported that ‘eCREST’, an electronic clinical reasoning simulation tool, improved students’ ability to gather essential patient information and learn clinical reasoning skills. Comparatively, suturing and knot-tying proficiencies were found to be comparable across virtual classroom training and traditional face-to-face training, suggesting that certain practical skills can be effectively taught through virtual methods [17, 32]. However, both these training modalities were superior to computer-based learning in producing better outcomes, albeit with varying costs per attendee, highlighting the economic considerations in choosing the optimal training approach [17].In the area of CPR training, computer-based learning groups demonstrated a less comprehensive understanding of procedures such as calling for assistance and using a defibrillator, when compared to their counterparts in face-to-face training sessions [18]. This points towards the limitations of computer-based learning in fostering practical critical skills in emergency scenarios. Figure 3 summarises the effect of digital learning tools by skill area and shows a high amount of statistical heterogeneity in the results.

Fig. 3
figure 3

Forest plot by of digital learning tools effectiveness by skill area

Student performance and knowledge retention

Studies consistently showed that digital interventions positively affect student performance and knowledge retention. Confidence levels and diagnostic abilities significantly increased following interventions and remained elevated four months post-intervention in some studies, indicating sustainable improvements in these crucial areas [19]. Videos and animated media enhanced the comprehension and procedural adherence in clinical skills, indicating superior performance compared to traditional methods [20,21,22]. Studies using online modules or courses demonstrated an improvement in median final exam scores when compared to traditional lecture-based learning methods alone [16]. This is further supported by evidence showing a significant correlation between the frequency of visits to online modules and the extent of knowledge gained, indicating the value of engagement with digital resources.

Gong et al. [38] found that a blended learning platform enhanced student-centred learning and clinical practice, demonstrating higher theoretical and practical assessment scores. Blended learning approaches, incorporating web applications and simulated electronic health records, were associated with better immediate and delayed post-intervention test scores, enhanced confidence in electrocardiogram analysis, and improved clinical practice skills [28, 31]. This suggests that integrating online resources with traditional teaching methods can elevate learning outcomes significantly. Moreover, virtual microscopy was favoured over optical microscopy for its higher scores in subjective impressions, indicating a preference for digital tools in certain areas of study [41]. However, no significant differences were observed in academic performance between different groups [41].

Kasai et al. [29], highlighted that simulated electronic health records and online problem-based learning improved multiple clinical skills, including medical interviewing and counselling; while Huang et al. [34] reported that an online course enhanced competency in basic ocular examination, though students preferred using it as an additional tool rather than a replacement for traditional methods. Saeed et al. [39] and Vincent et al. [31] highlighted improvements in examination skills and breaking bad news skills, with significantly improved self-efficacy and OSCE scores, showcasing the effectiveness of hybridised video-based learning.

Figure 4 summarises the effect of digital learning tools on assessment outcomes. The mean difference favours digital teaching, mean difference 1.93 (95% CI 1.22 to 2.64), with a high amount of statistical heterogeneity I2 97%, P < 0.001.

Fig. 4
figure 4

Forest plot by of digital learning tools effectiveness by outcome

Student satisfaction and engagement

Despite mixed efficacy for improving clinical skills there was a consensus of better engagement and satisfaction levels with digital clinical skills teaching. This increase in satisfaction did not appear specific to any one modality. For example, digital modules [34], videos [20, 25], web applications [28, 29], and even virtual meeting platforms (Microsoft Teams) [39, 42] all showed increased satisfaction scores by students. Nazari et al. [25] reported that step-by-step video demonstrations were perceived to have lower cognitive load and were preferred over continuous video demonstrations. Saeed et al. [39, 42] indicated that students were generally satisfied with their learning experience. Somera et al. [41] noted positive subjective impressions and engagement with blended learning and virtual microscopy environments. Only one study showed reduced satisfaction from students when it was described that learning visuospatial concepts remotely online was less optimal than in person training [32]. Unit of analysis issues precluded the use of proportional meta-analysis to create a pooled, overall proportion.

Cost effectiveness of digital learning tools

The review also highlighted the potential cost-effectiveness of online learning modalities and noted improvements in long-term retention and confidence among learners. In 2022, Nathan et al. showed that computer based learning and virtual classroom training for teaching suturing skills resulted in a cost saving of 58.9% and 44.2% respectively compared to face-to-face training [17]. The other included studies did not include detailed cost breakdowns.

Risk of bias of included studies

The QuADS results indicated that the scores ranged from 56 to 90%. There were more medium quality studies (n = 14) compared to low (n = 2) and high-quality studies (n = 12). Only two studies scored below 60%. Most studies had very low scores on stakeholder engagement in the research. The studies with higher scores were randomised trials and they were judged to be appropriate in their statistical analyses and study designs. The risk of bias assessment is detailed in Table 2.

Table 2 Quality assessment outcome using the QuADS tool

Discussion

This systematic review highlights the growing role and potential of digital learning interventions in medical education. The broad geographical distribution and diverse study designs of the reviewed studies underscores the widespread interest and applicability of digital tools in enhancing clinical knowledge and skills. The study findings suggest that online learning platforms can be a valuable complement to traditional clinical education, offering flexible, engaging, and potentially more cost-effective training options.

Our analysis demonstrated that online modules and courses may improve exam scores and knowledge retention, particularly when compared to traditional lecture-based methods. This aligns with previous studies that have highlighted the benefits of e-learning in medical education, particularly in providing flexible, scalable, and accessible learning opportunities [11, 12]. The effectiveness of these tools in improving specific skills, such as diagnostic imaging [16] and cardiac auscultation [26], underscores their potential to supplement or even replace traditional teaching methods in certain contexts. Furthermore, virtual training in practical skills like suturing and knot-tying proved to be as effective as face-to-face training, though computer-based learning fell short in critical areas such as CPR training. This indicates that while digital interventions can be highly effective, they must be carefully matched to the skills being taught.

The results suggest that digital and blended learning interventions lead to sustained improvements in student performance and knowledge retention. Studies by Nathan et al. [17] and Azman et al. [19] showed significant gains in suturing skills and diagnostic abilities, consistent with literature suggesting that interactive and multimedia-enhanced learning can lead to better retention and application of knowledge [43, 44]. The findings from Kasai et al. [29] and Huang et al. [34] further support the notion that digital tools can enhance competency in various clinical tasks, although a balanced integration with traditional methods is often preferred.

High levels of student satisfaction and engagement with digital learning approaches were evident across multiple studies [20, 34, 39]. The studies reported that interactive and video-based learning environments were well-received by students, enhancing their engagement and satisfaction. This is in line with research that highlights the importance of interactivity and multimedia in maintaining student interest and motivation [45, 46], although the review by Ulum was not specific to medical education [45]. However, some studies noted that while digital methods were effective, students still valued the personal interaction and hands-on experience provided by traditional face-to-face training [31, 39]. The integration of multimedia resources, such as video-based learning, showed significant benefits in terms of engagement and comprehension, particularly in clinical contexts, though anatomical knowledge gains were less pronounced [20]. The success of blended learning approaches and virtual microscopy further supports the advantage of combining traditional and digital methods for optimal learning outcomes.

However, our review also identified a large amount of statistical and individual variability in the effectiveness of different digital interventions, with high dropout rates posing a challenge in some studies. Despite these challenges, the overall satisfaction and engagement levels were higher with digital learning modalities, suggesting that they can enhance the learning experience significantly.

Moreover, the potential cost savings associated with digital learning, as evidenced by the studies on suturing skills [17], present a compelling case for their broader implementation, especially in resource-limited settings. This finding is particularly relevant in the context of the COVID-19 pandemic, which necessitated rapid shifts to online learning and highlighted the need for cost-effective, scalable educational solutions [47].

Implications for practice

The integration of digital and blended learning into medical education has profound implications for the future of the field. The adaptability and scalability of these tools can address the challenges posed by increasing student numbers and limited clinical training opportunities [11]. Furthermore, the ability to provide consistent and standardised training through digital platforms can potentially enhance the overall quality of medical education, ensuring that all students receive comprehensive and equitable training.

However, the preference for traditional methods in certain scenarios, as noted in the study by Huang et al. [34], suggests that a hybrid approach may be most effective. Combining the strengths of digital tools with the hands-on, interactive nature of traditional training can create a more holistic and effective educational experience. This hybrid model can leverage the flexibility and accessibility of digital learning while preserving the essential elements of face-to-face interaction and practical skill development.

Future research should focus on longitudinal studies to assess the long-term impact of digital and blended learning on clinical skills and professional competencies. Exploring the optimal balance between digital and traditional methods and understanding the factors influencing student preferences and learning outcomes, will be critical in shaping the future of medical education.

Strengths and limitations

Strengths of our review included the diverse range of studies from various countries and studied interventions. This diversity enhances the generalisability of the findings across different educational contexts and healthcare systems. Additionally, the use of rigorous inclusion criteria and detailed risk of bias assessment, using the QuADS tool, ensures the quality and reliability of the synthesized evidence. However, the review also has limitations. The heterogeneity of the included studies, in terms of interventions, outcomes, and assessment methods, complicated direct comparisons and synthesis of results. The variability in study quality, as indicated by the wide range of QuADS tool scores, suggests that some findings should be interpreted with caution. Additionally, high dropout rates in several studies may have introduced bias, affecting the robustness of the conclusions. The lack of detailed cost analyses in most studies limits the ability to fully assess the economic impact of digital learning interventions. Despite these limitations, the review provides a comprehensive overview of the current landscape of digital learning in medical education and identifies key areas for future research and improvement.

Conclusions

The findings of our review suggest that digital and blended learning methodologies may offer benefits in medical education, particularly in terms of knowledge acquisition, confidence building, and engagement. The effectiveness of these approaches varies depending on the skill being taught and the specific educational context. The integration of traditional and innovative teaching methods appears to offer the most comprehensive learning experience, underscoring the importance of a multifaceted approach to medical education. Future studies should look at the relative efficacy of the different digital modalities.

Data availability

Data is provided within the manuscript or supplementary information files.

Abbreviations

QuADS:

Quality Assessment With Diverse Studies

PRISMA:

Preferred Reporting Items for Systematic Reviews and Meta-Analyses

References

  1. Frenk J, Chen LC, Chandran L, Groff EO, King R, Meleis A, et al. Challenges and opportunities for educating health professionals after the COVID-19 pandemic. Lancet. 2022;400:1539–56.

    Article  Google Scholar 

  2. Khamees D, Peterson W, Patricio M, Pawlikowska T, Commissaris C, Austin A, et al. Remote learning developments in postgraduate medical education in response to the COVID-19 pandemic – A BEME systematic review: BEME Guide 71. Med Teach. 2022;44:466–85.

    Article  Google Scholar 

  3. Daniel M, Gordon M, Patricio M, Hider A, Pawlik C, Bhagdev R, et al. An update on developments in medical education in response to the COVID-19 pandemic: a BEME scoping review: BEME Guide 64. Med Teach. 2021;43:253–71.

    Article  Google Scholar 

  4. Cairney-Hill J, Edwards AE, Jaafar N, Gunganah K, Macavei VM, Khanji MY. Challenges and opportunities for undergraduate clinical teaching during and beyond the COVID-19 pandemic. J R Soc Med. 2021;114:113–6.

    Article  Google Scholar 

  5. Nyoni CN, Dyk LH-V, Botma Y. Clinical placement models for undergraduate health professions students: a scoping review. BMC Med Educ. 2021;21:598.

    Article  Google Scholar 

  6. Tay YX, Chow HC, Ooi CC. Impact on clinical learning during coronavirus disease 2019 pandemic: from disruption to recovery and the new norms. Korean J Med Educ. 2020;32:297–305.

    Article  Google Scholar 

  7. Saad S, Richmond C, King D, Jones C, Malau-Aduli B. The impact of pandemic disruptions on clinical skills learning for pre-clinical medical students: implications for future educational designs. BMC Med Educ. 2023;23:364.

    Article  Google Scholar 

  8. Australian Rural Health Education Network. The Ongoing Impact of the COVID-19 Pandemic on Student Placements facilitated by University Departments of Rural Health. 2023.

  9. Abbasi K. How to unlock the technology and wisdom of tomorrow? J R Soc Med. 2021;114:107–107.

    Article  Google Scholar 

  10. Grafton-Clarke C, Uraiby H, Gordon M, Clarke N, Rees E, Park S, et al. Pivot to online learning for adapting or continuing workplace-based clinical learning in medical education following the COVID-19 pandemic: a BEME systematic review: BEME Guide 70. Med Teach. 2022;44:227–43.

    Article  Google Scholar 

  11. George PP, Papachristou N, Belisario JM, Wang W, Wark PA, Cotic Z, et al. Online eLearning for undergraduates in health professions: a systematic review of the impact on knowledge, skills, attitudes and satisfaction. J Glob Health. 2014;4:010406.

    Article  Google Scholar 

  12. Cook DA, Garside S, Levinson AJ, Dupras DM, Montori VM. What do we mean by web-based learning? A systematic review of the variability of interventions. Med Educ. 2010;44:765–74.

    Article  Google Scholar 

  13. Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. PLoS Med. 2021;18:e1003583.

    Article  Google Scholar 

  14. Hawkins RE, Welcher CM, Holmboe ES, Kirk LM, Norcini JJ, Simons KB, et al. Implementation of competency-based medical education: are we addressing the concerns and challenges? Med Educ. 2015;49:1086–102.

    Article  Google Scholar 

  15. Harrison R, Jones B, Gardner P, Lawton R. Quality assessment with diverse studies (QuADS): an appraisal tool for methodological and reporting quality in systematic reviews of mixed- or multi-method studies. BMC Health Serv Res. 2021;21:144.

    Article  Google Scholar 

  16. Durán-Guerrero JA, Ulloa-Guerrero LH, Salazar-Díaz LC. Blended learning: an effective methodology for teaching radiology to medical students. Rev Fac Med. 2019;67:273–7.

    Article  Google Scholar 

  17. Nathan A, Fricker M, Georgi M, Patel S, Hang MK, Asif A, et al. Virtual interactive Surgical skills Classroom: a Parallel-group, non-inferiority, Adjudicator-blinded, randomised controlled trial (VIRTUAL). J Surg Educ. 2022;79:791–801.

    Article  Google Scholar 

  18. Cho K, Foo YM, Dalziel B, Hu W. Chatbot-mediated learning of cardiac auscultation. Intern Med J. 2022;52:2176–80.

    Article  Google Scholar 

  19. Azman M, Rajasingam D, Mohammad Mursal Mokhtar RH, Hamzah AT, Iskandarrani II, Roslan AA, et al. Supplementary web-based learning platform for improving diagnostic ability and confidence in Otoscopy among Final Year Medical Undergraduate. Brunei Int Med J (BIMJ). 2021;17:116–23.

    Google Scholar 

  20. Grosser J, Bientzle M, Shiozawa T, Hirt B, Kimmerle J. Acquiring clinical knowledge from an online video platform: a randomized controlled experiment on the Relevance of Integrating Anatomical Information and clinical practice. Anat Sci Educ. 2019;12:478–84.

    Article  Google Scholar 

  21. Lehmann R, Lutz T, Helling-Bakki A, Kummer S, Huwendiek S, Bosse HM. Animation and interactivity facilitate acquisition of pediatric life support skills: a randomized controlled trial using virtual patients versus video instruction. BMC Med Educ. 2019;19:7.

    Article  Google Scholar 

  22. Herrmann-Werner A, Weber H, Loda T, Keifenheim KE, Erschens R, Molbert SC, et al. But Dr Google said… training medical students how to communicate with E-patients. Med Teach. 2019;41:1434–40.

    Article  Google Scholar 

  23. Berland N, Lugassy D, Fox A, Goldfeld K, Oh S-Y, Tofighi B, et al. Use of online opioid overdose prevention training for first-year medical students: a comparative analysis of online versus in-person training. Substance Abuse. 2019;40:240–6.

    Article  Google Scholar 

  24. Hansen JR, Gefke M, Hemmingsen R, Fog-Petersen C, Hoegh EB, Wang A, et al. E-Library of Authentic Patient Videos Improves Medical Students’ Mental Status Examination. Acad Psychiatry. 2020;44:192–5.

    Article  Google Scholar 

  25. Nazari T, van de Graaf FW, Dankbaar MEW, Lange JF, van Merrienboer JJG, Wiggers T. One step at a time: step by step Versus continuous video-based learning to prepare medical students for Performing Surgical procedures. J Surg Educ. 2020;77:779–87.

    Article  Google Scholar 

  26. Power JE, Toft LEB, Barrett M. The Murmur Online Learning Experience (MOLE) curriculum improves medical students’ ability to correctly identify Cardiac Murmurs. MedEdPORTAL: J Teach Learn Resour. 2020;16:10904.

    Article  Google Scholar 

  27. Plackett R, Kassianos AP, Kambouri M, Kay N, Mylan S, Hopwood J, et al. Online patient simulation training to improve clinical reasoning: a feasibility randomised controlled trial. BMC Med Educ. 2020;20:245.

    Article  Google Scholar 

  28. Viljoen CA, Millar RS, Manning K, Burch VC. Effectiveness of blended learning versus lectures alone on ECG analysis and interpretation by medical students. BMC Med Educ. 2020;20:488.

    Article  Google Scholar 

  29. Kasai H, Shikino K, Saito G, Tsukamoto T, Takahashi Y, Kuriyama A, et al. Alternative approaches for clinical clerkship during the COVID-19 pandemic: online simulated clinical practice for inpatients and outpatients-A mixed method. BMC Med Educ. 2021;21:149.

    Article  Google Scholar 

  30. Brewer PE, Racy M, Hampton M, Mushtaq F, Tomlinson JE, Ali FM. A three-arm single blind randomised control trial of naive medical students performing a shoulder joint clinical examination. BMC Med Educ. 2021;21:390.

    Article  Google Scholar 

  31. Vincent A, Urben T, Becker C, Beck K, Daetwyler C, Wilde M, et al. Breaking bad news: a randomized controlled trial to test a novel interactive course for medical students using blended learning. Patient Educ Couns. 2022;105:105–13.

    Article  Google Scholar 

  32. Zaghal A, Marley C, Rahhal S, Hassanieh J, Saadeh R, El-Rifai A, et al. Face-to-face versus distance learning of basic suturing skills in novice learners: a quantitative prospective randomized trial. BMC Med Educ. 2022;22:290.

    Article  Google Scholar 

  33. Enoch LC, Abraham RM, Singaram VS. A comparative analysis of the impact of online, blended, and face-to-face learning on medical students’ clinical competency in the affective, cognitive, and psychomotor domains. BMC Med Educ. 2022;22:753.

    Article  Google Scholar 

  34. Huang Z, Yang J, Wang H, Chen B, Zheng D, Chen H. Integration of massive Open Online Course (MOOC) in Ophthalmic skills Training for Medical students: outcomes and perspectives. Asia-Pacific J Ophthalmol (Philadelphia Pa). 2022;11:543–8.

    Article  Google Scholar 

  35. Flatt E, Brewer P, Racy M, Mushtaq F, Ashworth R, Ali F, et al. Can educational video resources improve learning when used to augment traditional teaching of clinical examination? A randomized control trial of novice medical students. BMC Med Educ. 2023;23:21.

    Article  Google Scholar 

  36. Lang F, Gerhauser AS, Wild C, Wennberg E, Schmidt MW, Wagner M, et al. Video-based learning of coping strategies for common errors improves laparoscopy training-a randomized study. Surg Endosc. 2023;37:4054–64.

    Article  Google Scholar 

  37. Rajendran K, Walters B, Kemball B, McKinley RK, Khan N, Melville CA. Pilot evaluation of an online resource for learning paediatric chest Radiograph Interpretation. Cureus. 2021;13:e12762.

    Google Scholar 

  38. Gong J, Ruan M, Yang W, Peng M, Wang Z, Ouyang L, et al. Application of blended learning approach in clinical skills to stimulate active learning attitudes and improve clinical practice among medical students. PeerJ. 2021;9:e11690.

    Article  Google Scholar 

  39. Saeed S, Kashif U, Zaki S, Samad K, Yousuf MZ, Raza M, et al. Teaching clinical skills using online modality through modified Peyton’s Framework: an experience from a Medical University in Pakistan. J Adv Med Educ Professionalism. 2023;11:15–23.

    Google Scholar 

  40. Heriwardito A, Ramlan AAW, Basith A, Aristya L. Effectiveness of endotracheal intubation and mask ventilation procedural skills training on second-year student using modified Peyton’s four-step approach during COVID-19 pandemic. Med Educ Online. 2023;28:2256540.

    Article  Google Scholar 

  41. Somera Dos Santos F, Osako MK, da Perdona G, Alves SC, Sales MGKU. Virtual Microscopy as a Learning Tool in Brazilian Medical Education. Anatomical Sci Edu. 2021;14:408–16.

    Article  Google Scholar 

  42. Saeed S, Khan MH, Siddiqui MMU, Dhanwani A, Hussain A, Ali MM. Hybridizing video-based learning with simulation for flipping the clinical skills learning at a university hospital in Pakistan. BMC Med Educ. 2023;23:595.

    Article  Google Scholar 

  43. Ruiz JG, Mintzer MJ, Leipzig RM. The impact of E-Learning in Medical Education. Acad Med. 2006;81:207.

    Article  Google Scholar 

  44. Ellaway R, Masters K. AMEE Guide 32: e-Learning in medical education part 1: learning, teaching and assessment. Med Teach. 2008;30:455–73.

    Article  Google Scholar 

  45. Ulum H. The effects of online education on academic success: a meta-analysis study. Educ Inf Technol (Dordr). 2022;27:429–50.

    Article  Google Scholar 

  46. Mayer RE. Incorporating motivation into multimedia learning. Learn Instruction. 2014;29:171–3.

    Article  Google Scholar 

  47. Dhawan S. Online learning: a panacea in the Time of COVID-19 Crisis. J Educational Technol Syst. 2020;49:5–22.

    Article  Google Scholar 

Download references

Acknowledgements

Not applicable.

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Consortia

Contributions

Conceptualization: AD, FA, SW, RM, BMA; Data curation: AD, FM, FA, SW, RM, BMA; Formal analysis: RM, FM, BMA; Investigation: all authors; Methodology: all authors; Project administration: RM, BMA; Resources: all authors; Software: all authors; Supervision: BMA; Validation: RM; Writing – original draft: RM; Writing – review and editing: all authors. All authors reviewed and approved the final manuscript.

Corresponding author

Correspondence to Richard G. McGee.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

McGee, R.G., Wark, S., Mwangi, F. et al. Digital learning of clinical skills and its impact on medical students’ academic performance: a systematic review. BMC Med Educ 24, 1477 (2024). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12909-024-06471-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12909-024-06471-2

Keywords