- Research
- Open access
- Published:
Enhancing feedback mechanisms in medical education: insights from clinical instructors in a Taiwanese medical center
BMC Medical Education volume 25, Article number: 575 (2025)
Abstract
Objective
This study aimed to evaluate clinical instructors’ agreement with and the effective implementation of the Objective Structured Clinical Examination (OSCE) in a medical center in Taiwan.
Methods
Clinical instructors qualified as examiners at the designated medical center were surveyed using a stratified random sampling technique. A total of 287 valid responses out of 350 questionnaires distributed were collected, yielding an 82% response rate. Descriptive statistics, independent sample t-tests, one-way ANOVA, and Pearson correlation analysis were employed for data analysis.
Results
The findings indicated that clinical instructors generally supported and demonstrated proficiency in the OSCE. Moreover, those who had participated in more scoring activities and received faculty training in the past three years showed stronger alignment with the OSCE and greater effectiveness in its implementation. A moderate to high positive correlation was found between instructors’ endorsement of the OSCE and their successful execution of it.
Conclusions
The results underscore the importance of clinical instructors’ endorsement and proficiency in OSCE implementation. These findings hold significance for medical education administrators, clinical instructors, and future research endeavors in medical education, aiming to enhance instructional quality.
Introduction
In the evolving healthcare landscape, patient needs are constantly changing, necessitating improvements in medical education to enhance healthcare quality and patient safety. Medical education assessment now emphasizes a patient-centered approach, focusing not only on knowledge but also on skill training and attitude development. To move beyond rigid classroom teaching and written exams, the Objective Structured Clinical Examination (OSCE) has emerged as a key reform in Taiwanese medical centers and a global trend in licensing examinations for healthcare professionals [1]. Introduced by Harden and Gleeson in 1975, OSCE utilizes standardized patients (SPs) in structured clinical scenarios to evaluate essential competencies such as history-taking, physical examination, communication, and patient education. This method ensures competency-based learning outcomes, clinical proficiency, and serves as a reference for instructional improvement [2]. Recognizing its effectiveness, Taiwan incorporated OSCE into the National Medical Licensing Examination. Since 2003, the Taiwan Association of Medical Education has collaborated with national agencies to implement the “Clinical Skill Assessment Test Program,” establishing OSCE sites in medical schools and teaching hospitals. By 2019, 26 hospitals were actively engaged in this initiative [3].
In recent years, various training hospitals entrusted with implementing the “Clinical Skill Assessment Test Program” by the Taiwan Association of Medical Education have been facing the impact of lack of funding. Hospitals need to maintain the quality of OSCE execution using self-generated resources. As a result, it has been observed that the number of participants in examiner training courses organized by hospitals has been decreasing year by year. Although OSCE implementation has been widely studied in Western countries, there remains a gap in understanding its acceptance, effectiveness, and sustainability in Taiwan, particularly in the context of diverse medical professions beyond physicians. This study aims to address this gap by exploring the perspectives of clinical educators across multiple disciplines on the acceptance and implementation of OSCE.
To investigate this, it is necessary to first understand the current level of recognition and implementation of OSCE. Beyond the practical implementation and execution of training, there is a need for investment in terms of financial resources, time, and manpower. Properly utilizing these limited resources to identify issues will help transform training from being seen as an unnecessary expenditure to an important investment. This investment can lead to improved student learning outcomes and faculty development, ultimately enhancing the quality of teaching. A key strength of this study is its comprehensive examination of clinical educators’ perspectives across different professional backgrounds, which has not been extensively explored in previous OSCE-related research. By identifying factors that influence OSCE acceptance, this study provides valuable insights into optimizing faculty training and resource allocation to enhance medical education effectiveness.
As medical education evolves, educators must continually enhance their teaching and professional skills to integrate new knowledge into practice [4]. Expanding the pool of OSCE examiners across various professions requires regular faculty training to strengthen understanding of the assessment method. A medical center in Taiwan has implemented OSCE in over ten medical technology fields beyond Western and traditional Chinese medicine and dentistry. However, limited awareness of curriculum quality and faculty training effectiveness necessitates further investigation. This study examines how clinical educators from different backgrounds perceive and accept OSCE, focusing on factors influencing examiner training participation and sustainable implementation in Taiwan.
Literature review
Objective Structured Clinical Examination (OSCE) in Taiwan
The Objective Structured Clinical Examination (OSCE), introduced by Ronald Harden in 1975, is an assessment method in medical education using multiple thematic stations. Examinees rotate through stations, demonstrating clinical skills within a set time, assessed by an examiner. OSCE evaluates history taking, physical examination, diagnosis, communication, and other competencies beyond written exams [5]. As OSCE became integral to national medical licensing exams worldwide, Taiwan established a promotion team in 2004. After a phased implementation involving government and medical institutions, OSCE was officially incorporated into Taiwan’s national medical licensing examination in 2013 [6].
The OSCE is an assessment tool based on objective and standardized principles. Candidates go through a series of timed stations in a circuit, where they undergo professional performance evaluations in simulated environments. Trained instructors serve as scoring examiners at each station, following standardized scoring sheets for assessment [7]. Based on Miller’s Pyramid theory, OSCE effectively measures learners’ ability to perform clinical tasks, making it a widely used clinical skills assessment method [8]. It integrates summative and formative assessment approaches, offering standardization, objectivity, and bidirectional feedback [9]. Summative OSCEs assess competency for certification or promotion, while formative OSCEs identify learning gaps and provide feedback to enhance instruction. Instructors frequently use the formative approach to evaluate interns and new staff, refining training based on observed progress [10].
The OSCE assesses learners’ clinical skills in simulated scenarios. Its advantages lie in the ability of the examiners to standardize the content of the assessment and control the clinical situations of the test questions. It is suitable for critical conditions, rare cases, complex clinical situations, and interdisciplinary cooperation, unlike methods such as mini-CEX and DOPS. In clinical medical education, a patient-centered approach emphasizing skill training and attitude cultivation is gaining importance. This shift seeks to overcome the limitations of traditional didactic methods and paper-based assessments, which cannot fully measure clinical abilities. OSCE offers key benefits: (1) controlled, standardized environments prevent medical errors, (2) it aligns curriculum and assessment for efficient teaching and evaluation, and (3) standardized patients can be reused for assessment and feedback [11].
Designing and planning an OSCE assessment requires extensive preparation, including arranging clinical skills center equipment, preparing instructional materials, training examiners and standardized patients, conducting the OSCE assessment, and conducting post-assessment reviews. This process demands a significant amount of human and material resources. It also involves personnel costs, including training fees for OSCE examiners and standardized patients, as well as attendance fees [12]. While the implementation of OSCE in Taiwan has been supported by government agencies, economic costs and human resource challenges remain, necessitating further improvements.
Agreement and related studies
Agreement refers to the psychological process of aligning one’s thoughts or actions with a subject or concept, resulting in a harmonious convergence of vision and behavior [13]. It reflects a level of affiliation with the subject, manifested in seriousness, emotions, and actions. Some studies suggest that agreement involves the internalization of a subject’s values, influencing an individual’s thoughts, feelings, and behavior towards that subject [14]. Scholars also equate agreement with acceptance, describing it as a positive attitude and behavioral intention towards curriculum changes, as evidenced by teachers’ research outcomes [15]. Taking into account the discussions of the scholars above, “agreement” is a process through which an individual, after social learning or socialization, gains awareness and understanding of things or entities, generating a subjective sense of approval within themselves. In this study, “agreement” refers to the subjective level of approval an individual holds after perceiving and comprehending a matter.
Based on the collection of relevant literature regarding agreement in education, instructional projects, or policy implementation, this study has summarized and analyzed the research and opinions of various experts and scholars. Reviewing the literature, “instructional administrative efficiency” was the most frequently addressed aspect, followed by “teacher professional development” and “teaching curriculum policies,” with “student learning achievement” sharing the third position. Therefore, this study has chosen these four dimensions as the basis for designing the agreement questionnaire. Additionally, the revised scales on OSCE agreement for clinical instructors as formulated by Shun (2008) and Chen (2015) were referenced [13, 16], as explained below:
-
(1)
Instructional Administrative Efficiency:
Instructional administrative efficiency refers to a hospital’s ability to integrate internal and external resources, employ technological methods, systematic management, and effective leadership to support teaching. This enhances the overall medical quality and patient safety of the hospital, thus achieving educational objectives.
-
(2)
Teacher Professional Development:
Teacher professional development refers to teachers actively and proactively engaging in various forms of learning and research throughout their teaching careers. This involves enhancing their teaching abilities, engaging in self-reflection and inquiry to elevate professional knowledge (cognition), strengthening professional attitudes, and improving teaching quality (skills). The aim is to meet students’ learning needs and enhance their learning outcomes.
-
(3)
Teaching Curriculum Policies:
Teaching curriculum policies involve the hospital designing specialized courses based on the clinical quality and patient safety of trained students or learners. These courses are grounded in the resources of students’ clinical skills, hospital context, and occupational characteristics. The design aligns with the hospital’s vision and curriculum philosophy, resulting in distinct courses tailored to different professions.
-
(4)
Student Learning Achievement:
Student learning achievement refers to the growth or changes displayed by students during and after their learning processes. This includes an increase in knowledge (cognition), proficiency in skills (skills), and alterations in learning attitudes and motivation (affective). The goal is to enhance the level of enthusiasm for learning.
Implementation efficacy and related studies
OSCE has become a standard assessment in medical education, extending beyond medical doctors to other healthcare professionals. Teaching-oriented OSCE involves direct observation by clinical instructors, peer review via video recording, and real-time feedback on clinical skills. This feedback helps students identify strengths and areas for improvement, enhancing the educational value of OSCE [17]. Effective OSCE planning requires meticulous preparation, addressing potential challenges like inadequate content, unclear instructions, or poorly trained standardized patients. It is essential to align assessment content with curriculum objectives to ensure consistency [18]. Clinical instructors must possess the necessary knowledge and skills to design and manage OSCE materials, balancing teaching, clinical duties, and OSCE-related workload [12].
This study, based on expert opinions and relevant literature on the application of OSCE, synthesizes and identifies the implementation effectiveness variables suitable for this research:
-
(1)
Student Performance:
Research has shown that medical students who underwent external clinical experience before a second OSCE assessment showed significant improvement in treatment recommendations, with better accuracy and presentation compared to their first assessment without such experience [19]. An evaluation of doctor-patient communication skills in seventh-year medical students revealed commendable etiquette and attitudes, though there is room for improvement in professional knowledge and verbal precision [20]. A comparative study between conventional skill assessments and OSCE examinations found that while students more easily passed conventional assessments, OSCE was more comprehensive, appropriate, and impartial [21].
-
(2)
Teaching Professionalism:
Research indicates that essential behaviors for clinical instructors include role modeling as exemplary learners, a sense of responsibility towards the profession, the ability to apply theory in clinical settings, and adept interpersonal relationships and communication skills. Being capable of providing correct and clear demonstrations to students, a clinical healthcare professional should embody the necessary professionalism, assisting newly entering healthcare personnel in delivering optimal patient care and addressing their queries. This involves applying theory to real clinical scenarios [22].
-
(3)
Professional Dedication:
In education, teachers should uphold professionalism, adhere to ethical norms, and invest time and effort in students’ learning and growth. A dedicated attitude is essential for overcoming challenges and fulfilling the mission of education [23]. Clinical educators, in particular, must adapt to the evolving clinical landscape and engage in self-directed learning for their professional development. The professional competence of clinical nursing educators includes a genuine interest in patient care, the integration of theory with practice, and the ability to serve as role models for students during internships [24].
Methodology
Research structure
This study aims to investigate the level of agreement and implementation effectiveness of OSCE among clinical instructors in medical centers. Through a review of relevant literature both domestically and internationally, the relationships among various variables were identified. Clinical instructor background variables are considered independent variables, and “agreement level” and “implementation effective-ness” are the dependent variables, forming the two main components. The research framework is depicted in Fig. 1.
Participates
This study targeted a total of 430 clinical instructors employed at a medical center in the year 2023, who met the criteria for examiner certification. The research utilized a questionnaire survey method, which was conducted in two stages: a pilot questionnaire and an official questionnaire. After undergoing validation through expert scrutiny and necessary revisions, the pilot questionnaire was used for the survey. To ensure the questionnaire’s accuracy, a sample size of over 150 participants was necessary. During the factor analysis, the ratio of items to participants should be 1:5, and the total sample size should not be less than 100 to ensure the reliability of the results [25, 26]. Therefore, the pilot questionnaire of this study aimed to include a sample size of 150 participants. For the pilot questionnaire, the researchers employed a stratified sampling approach based on the size of the target population’s occupational categories. The sample size for each occupational category was determined based on the percentage of clinical instructors in each category at a certain medical center. The sampling method used was purposive sampling.
Following item analysis and reliability analysis, the pilot questionnaire was refined to create the official questionnaire. To ensure the convenience of the study and consider sample representativeness and response rate, the researchers used a “stratified random sampling” approach for the official questionnaire. Strata were defined based on the occupational categories’ size, with the sample size for each category determined by the percentage of clinical instructors in each category. Gay and Airasian (2003) recommended that the formal sample size for questionnaire surveys should be at least 10% of the target population [27]. When the target population is relatively small, a formal sample size of over 20% of the target population is appropriate [25, 26]. Based on these considerations, this study set the formal questionnaire sample size at 350 participants, and the sampling would be conducted using proportional random sampling.
Research tools
This study employed a questionnaire survey method to collect relevant demographic information. Using the literature review and research framework as references, a preliminary draft of the pilot questionnaire was created. To enhance the construct validity and reliability of the research instrument, item analysis and Cronbach’s α coefficient testing were conducted using the statistical software SPSS after implementing the pilot questionnaire. This aimed to evaluate the reliability and validity of the research tool.
After the questionnaire content was reviewed and revised by experts in the field of medical education, it was refined and compiled into the official questionnaire. The content of the official questionnaire consisted of three parts: the first part covered “Basic Information of Clinical Instructors,” the second part included the " Clinical Instructors’ Level of Agreement with OSCE " with a total of 23 items, and the third part encompassed the “Clinical Instructors’ Perception of the Implementation Effectiveness of OSCE” with a total of 14 items. The survey questionnaire used a Likert-type five-point scale for responses. Clinical instructors were asked to provide their own real-life responses to the questions related to their perceived acceptance and implementation effectiveness of OSCE. Respondents selected the option that closely matched their feelings and experiences from the five response categories: “Strongly Agree,” “Agree,” “Neutral,” “Disagree,” and “Strongly Disagree.” These options were assigned scores from 5 to 1, respectively.
Data processing
After conducting the questionnaire survey in this study, the collected valid questionnaire data were archived and coded. Statistical analysis of the data was performed using the SPSS software. Initially, descriptive statistics were used to analyze the distribution of background information within the research sample, including frequency distribution, percentages, means, and standard deviations. Additionally, the acceptance level and implementation effectiveness of Objective Structured Clinical Examination (OSCE) among clinical instructors at the medical center were examined. Independent samples t-tests and one-way analysis of variance (ANOVA) were employed to explore the differences in acceptance level and implementation effectiveness of OSCE among clinical instructors with different backgrounds. In cases where one-way ANOVA yielded significant results, the Scheffé method was utilized for post hoc comparisons to understand the differences between groups for that particular variable. Pearson’s correlation coefficient was employed to analyze the relationship between clinical instructors’ acceptance level and implementation effectiveness of OSCE across various dimensions. The statistical significance level for all tests in this study was set at α = 0.05.
Results
Analysis of clinical instructors’ personal background variables
This study focused on clinical instructors from a certain medical center who were certified as examiners. The survey was conducted using a stratified random sampling method. A total of 350 research questionnaires were distributed, and 287 valid questionnaires were collected (82.0%). The background variables in this questionnaire included gender, age, education level, job category, teaching experience, teacher rank, Examiner for medical/nurse exams (past 3 years), Participation in assessment sessions (past 3 years), OSCE instructor training (past 3 years). Descriptive statistical methods such as frequency distribution and percentages were used to present the basic data of the questionnaire samples, aiming to understand the distribution of basic information. Based on the Central Limit Theorem sample size guidelines, each group should have at least 30 samples to ensure that the distribution of means tends toward a normal distribution, and the larger the sample size, the closer it approaches a normal distribution with smaller variability. Therefore, in cases where the group size did not reach 30 participants, those groups were combined.
Among the respondents, 71.4% were female and 28.6% male. Age distribution showed 33.4% were 40 years or younger, 44.9% were 41–50 years, and 21.6% were above 50. Educational attainment included 54.0% with a college/university degree and 46.0% with a graduate degree. Regarding job categories, 65.5% worked in institutions with over 101 medical professionals, 13.2% in facilities with 51–100, and 21.3% in smaller institutions with 50 or fewer personnel. Teaching experience varied, with 15.3% having ≤ 5 years, 17.8% with 6–10 years, 20.9% with 11–25 years, and 46.0% with over 26 years. Faculty ranks included 41.1% ordinary faculty, 32.8% advanced faculty, and 26.1% senior faculty. Professional assessment participation in the past three years showed 20.2% served as medical/nursing examiners. Additionally, 30.3% attended ≥ 3 assessment sessions, 27.9% attended 1–2, and 41.8% had no participation. OSCE instructor training was completed by 66.9%, while 33.1% had no training.
Clinical instructors’ level of agreement and implementation effectiveness of OSCE
To understand the current status of clinical instructors’ level of agreement and implementation effectiveness regarding the OSCE at a certain medical center, statistical methods such as mean and standard deviation were employed to analyze participants’ responses in various dimensions and the overall scale of the questionnaire. This study utilized a Likert-type five-point scale for differentiation, with average scores ranging from 0 to 5 for each dimension. An average score above 4.50 indicates “strongly agree,” while a score between 3.50 and 4.49 indicates “agree.” A score between 2.50 and 3.49 represents “neutral,” whereas a score between 1.50 and 2.49 denotes “disagree.” An average score below 1.50 signifies “strongly disagree.”
The average scores of the level of agreement in various dimensions of the acceptance scale are ranked as follows: “Teaching Professional Growth” (M = 4.35), “Teaching Resources and Administrative Efficiency” (M = 4.33), “Teaching Curriculum Policies” (M = 4.26), and " Student Learning Achievements” (M = 4.23). From the average scores, it can be inferred that clinical instructors’ level of agreement regarding the OSCE tends to be in the “Agree” range for dimensions such as “Teaching Resources and Administrative Efficiency,” “Teaching Professional Growth,” “Teaching Curriculum Policies,” and " Student Learning Achievements.”
In the analysis of questionnaire items, the highest average scores were observed in statements such as No. 9 (M = 4.38), No. 7 (M = 4.37), No. 2 and No. 3 (M = 4.36). The lowest average scores were observed in statements such as No. 20 (M = 4.18), No. 21 (M = 4.12), and No. 16 (M = 4.06). These three statements had relatively lower levels of agreement, as shown in Table 1.
The average scores of the different dimensions in the implementation effectiveness scale are ranked as follows: “Teaching Expertise” (M = 4.32), “Professional Attitude” (M = 4.20), and “Student Performance” (M = 4.19). From the above average scores, it can be seen that clinical instructors’ performance in the implementation effectiveness of the objective structured clinical examination (OSCE) tends to be “Agree” in dimensions such as “Student Performance,” “Teaching Expertise,” and “Professional Attitude”.
In the analysis of questionnaire items, the highest average scores are found in statements such as No. 29 (M = 4.35), No. 31 (M = 4.34), and No. 32 (M = 4.32). The lowest average scores are observed in statements like No. 24 (M = 4.17), No. 26 (M = 4.13), and No. 37 (M = 4.12), as shown in Table 2.
Analysis of clinical instructors’ agreement levels of OSCE across different background variables
The analysis revealed significant differences in agreement levels based on instructors’ prior experience as national examination assessors, participation in assessment sessions, and engagement in OSCE-related faculty training. Instructors with national examination assessor experience demonstrated significantly higher agreement levels compared to those without such experience (t = 2.018, p = 0.045), as presented in Table 3.
Furthermore, the frequency of assessment participation had a notable impact on agreement levels (F = 12.270, p < 0.001). Post hoc comparisons indicated that instructors who had conducted three or more assessment sessions exhibited significantly higher agreement levels than those with no assessment experience (p < 0.001) or with only one to two sessions (p < 0.001), as shown in Table 4.
Additionally, faculty training in the past three years was associated with significantly higher agreement levels (t = 3.247, p = 0.001), underscoring the importance of continuous professional development in fostering positive perceptions of OSCE (Table 5).
Conversely, no significant differences in agreement levels were observed across gender, age, educational background, job category, teaching experience, or faculty rank (p > 0.05). These findings suggest that demographic and professional background factors do not substantially influence instructors’ agreement levels regarding OSCE.
Analysis of clinical instructors’ implementation efficacy of OSCE across different background variables
The findings indicate that clinical instructors’ implementation efficacy of OSCE was significantly influenced by their level of participation in assessment sessions and faculty training. Instructors who had conducted three or more assessment sessions demonstrated significantly higher implementation efficacy compared to those with no assessment experience (F = 5.900, p = 0.003). Post hoc comparisons confirmed that instructors with extensive assessment experience exhibited notably stronger implementation effectiveness (p < 0.01) than those with limited or no experience (Table 6).
Similarly, faculty training in the past three years was positively correlated with implementation efficacy (t = 2.225, p = 0.027). Instructors who had undergone training reported significantly greater effectiveness in applying OSCE methodologies in clinical education (Table 7).
In contrast, no significant differences in implementation efficacy were observed based on gender, age, educational background, job category, teaching experience, faculty rank, or experience as national examination assessors (p > 0.05). These results suggest that while structured training and assessment participation enhance OSCE implementation efficacy, general demographic and professional factors do not exert a notable influence.
Analysis of correlation between agreement and implementation effectiveness of OSCE
To investigate the correlation between the agreement levels and implementation effectiveness of clinical instructors at a medical center, Pearson’s product-moment correlation coefficient was employed to analyze whether there is a significant correlation between the agreement levels and implementation effectiveness of clinical instructors. A positive correlation coefficient indicates a positive relationship between the two variables, while a negative correlation coefficient indicates a negative relationship. The magnitude of the correlation coefficient reflects the degree of association between the variables. In this study, the absolute values of the correlation coefficients were used, and when the absolute value of the correlation coefficient (denoted as γ) is less than 0.4, it indicates a “low correlation.” If γ is between (inclusive) 0.4 and 0.7, it indicates a “moderate correlation.” When γ is greater than 0.7, it indicates a “high correlation [25, 26].”
The research findings demonstrate a significant correlation between the overall agreement levels and implementation effectiveness of clinical instructors, with each aspect showing a significantly high positive correlation (p < 0.01). The “Teaching Resources and Administrative Efficiency” aspect of clinical instructor agreement levels and implementation effectiveness also shows a significant moderate positive correlation (p < 0.01). The “Teacher Professional Growth” aspect of clinical instructor agreement levels and implementation effectiveness reveals a significant moderately high positive correlation (p < 0.01). The “Teaching Curriculum Policies” aspect of clinical instructor agreement levels and implementation effectiveness displays a significant moderately high positive correlation (p < 0.01). The " Student Learner Achievement” aspect of clinical instructor agreement levels and implementation effectiveness indicates a significant high positive correlation (p < 0.01).
Summing up the results, the absolute values of the correlation coefficients between the various aspects of agreement levels and implementation effectiveness range from 0.524 to 0.852. This indicates a moderately high positive correlation between the different dimensions of agreement levels and implementation effectiveness. It suggests that higher agreement levels among clinical instructors regarding OSCE correspond to higher implementation effectiveness, and higher implementation effectiveness in OSCE leads to better agreement levels, as shown in Table 8.
Discussion
Analysis of the clinical instructors’ agreement levels and implementation effectiveness in the medical center
The results of this study indicate that among the dimensions of agreement levels towards OSCE, the aspect of “Teacher Professional Growth” holds the highest degree of agreement within the medical center. The items with the highest and second-highest mean scores are “Being able to periodically self-assess and improve the content of OSCE teaching training to enhance teaching effectiveness and quality” (M = 4.38) and “Participating in internal and external OSCE-related training activities to enhance personal professional growth” (M = 4.37). Regarding the dimensions of OSCE implementation effectiveness, the aspect of “Teaching Professionalism” demonstrates the highest level of agreement. The items with the highest and second-highest mean scores within this aspect are “Participating in OSCE training helps to better understand the direction of curriculum planning and instructional material design” (M = 4.35) and “Participating in OSCE training enhances feedback skills towards students” (M = 4.34).
Differences in clinical instructors’ agreement levels towards OSCE based on various background variables
The research findings indicate that clinical instructors with different genders, ages, educational backgrounds, occupational categories, teaching experience, and teaching ranks exhibit no significant differences in their agreement levels towards the OSCE on both specific aspects and the overall scale. The agreement levels of clinical instructors towards OSCE are not influenced by these background variables. Originally, this study considered that the number of instructors in different occupational categories might impact their agreement levels due to potential variations in material preparation and resource allocation for clinical activities. However, the results suggest that even with varying numbers of instructors in different categories, there are no significant differences in their agreement levels towards OSCE. On the other hand, the experience of clinical instructors in serving as national examination assessors, the number of actual assessment sessions conducted in the past three years, and whether they have received teacher training within the past three years significantly impact their agreement levels towards OSCE. This indicates that clinical instructors who have experience as examiners, have conducted more assessment sessions, and have undergone recent teacher training exhibit higher levels of agreement towards the Objective Structured Clinical Examination.
Differences in clinical instructors’ implementation effectiveness of OSCE based on various background variables
The research findings reveal that clinical instructors with different genders, ages, educational backgrounds, occupational categories, teaching experience, and teaching ranks exhibit no significant differences in the implementation effectiveness of the OSCE on specific aspects as well as the overall scale. The implementation effectiveness of OSCE is not influenced by these background variables. Similar findings have been reported in previous research, which showed that examiners from diverse backgrounds can achieve good agreement in OSCE assessments when using standardized criteria, despite individual differences in scoring approaches [28].
However, the experience of clinical instructors in serving as national examination assessors does not show significant differences in the overall implementation effectiveness. Yet, it does show significant differences in the “Learner Performance” and “Teaching Professionalism” aspects of implementation effectiveness. Clinical instructors with experience as national examination assessors exhibit significantly higher implementation effectiveness compared to those without such experience, aligning with the hypothesis of this study. The number of actual assessment sessions conducted in the past three years significantly impacts the implementation effectiveness of the OSCE as a whole. Post hoc comparisons indicate that clinical instructors who conducted “3 or more assessment sessions” in the past three years have significantly higher implementation effectiveness compared to clinical instructors with “no actual assessment sessions.” Whether clinical instructors received teacher training within the past three years results in a significant overall difference in the implementation effectiveness of OSCE. However, this difference is not significant in the “Learner Performance” aspect, suggesting that learner performance is not affected by whether clinical instructors have received teacher training.
Correlation between clinical instructors’ agreement level and implementation effectiveness of OSCE
Significant differences in the correlation between clinical instructors’ agreement level and the implementation effectiveness of OSCE are observed across various aspects. The correlation coefficients (γ values) in descending order are “Overall Implementation Effectiveness” (γ = 0.852), “Teaching Professionalism” (γ = 0.812), “Professional Attitude” (γ = 0.793), and “Learner Performance” (γ = 0.761). These correlations exhibit a significant and strong positive relationship, with all significance levels being p < 0.01. The absolute values of correlation coefficients between agreement levels and implementation effectiveness range from 0.524 to 0.852, indicating a moderate to high positive correlation across different dimensions. These findings indicate that higher agreement levels among clinical instructors correspond to greater OSCE implementation effectiveness. This aligns with Hawezy et al. (2024), which demonstrated that examiners with prior OSCE experience perceive it as a fair and comprehensive assessment tool, enhancing confidence in its use. Additionally, structured OSCE frameworks, such as the Grand OSCE (GOSCE), further support its acceptance and effectiveness in medical education, reinforcing the observed correlation in this study [29].
Conclusions
This study demonstrates that clinical instructors at the medical center exhibit high agreement with and effective implementation of the Objective Structured Clinical Examination (OSCE). Key findings are summarized as follows:
-
(1)
High Agreement Levels with OSCE: Clinical instructors achieved an overall high agreement score (4.29), particularly in “Teacher Professional Growth,” indicating a strong commitment to improving their understanding and application of OSCE despite clinical responsibilities.
-
(2)
Effective OSCE Implementation: The overall implementation effectiveness score (4.24) underscores the positive impact of OSCE on “Teaching Expertise,” reinforcing its role in enhancing instructional quality and professional development.
-
(3)
Impact of Assessment Frequency: Instructors who participated in OSCE assessments more frequently over the past three years demonstrated significantly higher agreement levels, suggesting that hands-on experience fosters greater acceptance and integration of OSCE in teaching practices.
-
(4)
Influence of Teacher Training: Those who received training within the past three years exhibited higher implementation effectiveness, emphasizing the value of structured training in improving instructional strategies and assessment proficiency.
-
(5)
Positive Correlation Between Agreement and Implementation: A significant correlation (ranging from 0.524 to 0.852) between agreement levels and implementation effectiveness indicates that greater acceptance of OSCE translates into more effective application.
These findings suggest that increasing participation in OSCE assessments and structured faculty training can further enhance instructional quality. Hospital administrative education units should facilitate interdisciplinary collaborations and targeted training to optimize OSCE adoption and improve both instructor competencies and student learning outcomes.
Data availability
Data available on request due to restrictions regarding privacy, legal and ethical concerns. The data presented in this study are available on request from the corresponding author.
Change history
14 May 2025
A Correction to this paper has been published: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12909-025-07276-7
References
Tsai SL, Chen CH, Fang JT, Tsai JJ, Chang CC. A guideline for the implementation of objective structured clinical examination. J Med Educ. 2008;12(2):118–32. https://doiorg.publicaciones.saludcastillayleon.es/10.6145/jme.200806_12(2).0007.
Cheng CY, Wu MF, Liou WS. Pharmacy objective structured clinical examination (OSCE) scenario design practice. Formosa J Clin Pharm. 2018;26(4):251–7. https://doiorg.publicaciones.saludcastillayleon.es/10.6168/FJCP.201810_26(4).0001.
Taiwan Association of Medical Education. 2022 Medical Clinical Skills Examination (OSCE) Implementation Plan Results Report. 2022. Available online: https://reurl.cc/015a4M (accessed on 22 March 2025).
Lu CHA, March Study of Prefessional Growth and Teaching Effectiveness of Teachers’ Participation on the Team Teaching with Industry Experts in. Senior High Schools. Master’s Thesis, National Changhua University of Education, Changhua County, Taiwan, 2018. Available online: https://hdl.handle.net/11296/avuzd5 (accessed on 10 2025).
Dung BT. National examination Innovations - The objective structured clinical examination (OSCE) is included in the medical examination. Sharp Innov Experimental Politics. 2013;3:3–15. https://reurl.cc/2L2ajX.
Liu KM, Gao MM. The history, status and prospect of the inclusion of objective structured clinical examination in the National examination for Taiwan medical Doctors. Natl Elite Magazine. 2015;11(4):45–70. https://reurl.cc/eDeKLK.
Khan KZ, Ramachandran S, Gaunt K, Pushkar P. The objective structured clinical examination (OSCE): AMEE guide 81. Part I: an historical and theoretical perspective. Med Teach. 2013;35(9):e1437–46. https://doiorg.publicaciones.saludcastillayleon.es/10.3109/0142159X.2013.818634.
Luang HM. The Reliability and Validity of OSCE Instrument: A Preliminary Analysis. Master’s Thesis, Central Taiwan University of Science and Technology, Taichung City, Taiwan, 2019. Available online: https://hdl.handle.net/11296/qjgv64 (accessed on 18 March 2025).
Clarke S, Rainey D, Traynor M. Using the objective structured clinical examination (OSCE) to assess orthopaedic clinical skills for the registered nurse. Int J Orthop Trauma Nurs. 2011;15(2):92–101. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.ijotn.2010.11.003.
Boulet JR, Smee SM, Dillon GF, Gimpel JR. The use of standardized patient assessments for certification and licensure decisions. Simul Healthc. 2009;4(1):35–42. https://doiorg.publicaciones.saludcastillayleon.es/10.1097/SIH.0b013e318182fc6c.
Chen YL, March An Action Research on Constructing Questions for Objective Structured Clinical Examinations of Traditional Chinese Medicine. Master’s Thesis, National Changhua University of Education, Changhua County, Taiwan, 2017. Available online: https://hdl.handle.net/11296/cj843d (accessed on 16 2025).
Yang CI, Lee LH, Yang HC, Chen SL. The OSCE and Five-Year junior nursing colleges in Taiwan: development and challenges. Cheng Ching Med J. 2016;12(1):40–7. https://www.airitilibrary.com/Publication/alDetailedMesh?docid=18136702-201601-201601080015-201601080015-40-47.
Chen YC. A Study of Teacher Recognition and the Effect of Program of Transformating of High Quality for Elementary Schools in Yunlin County. Master’s Thesis, National Yunlin University of Science and Technology, Yunlin County, Taiwan, 2015. Available online: https://hdl.handle.net/11296/sa2e3c (accessed on 19 March 2025).
Cin HCA, Study on Elementary School Teachers’s Professional Development Evaluation Identification. Master’s Thesis, National Taipei University of Education, Taipei City, Taiwan, 2008. Available online: https://hdl.handle.net/11296/hn6t33 (accessed on 07 March 2025).
Lee CK, Yin HB. Teacher receptivity to and concerns about curriculum change: an exploration of research on curriculum implementation. J Educational Res Dev. 2005;1(1):107–28. https://journal.naer.edu.tw/periodical_detail.asp?DID=vol001_06.
Shun CM, March A Study on the Level of Understanding and Recognition of School Marketing Strategies of National Elementary School Teachers. in Yunlin County. Master’s Thesis, National Chiayi University, Chiayi County, Taiwan, 2008. Available online: https://hdl.handle.net/11296/q73f49 (accessed on 14 2025).
Brazeau C, Boyd L, Crosson J. Changing an existing OSCE to a teaching tool: the making of a teaching OSCE. Acad Med. 2002;77(9):932. https://journals.lww.com/academicmedicine/Fulltext/2002/09000/Changing_an_Existing_OSCE_to_a_Teaching_Tool__The.36.aspx.
McWilliam P, Botwinski C. Developing a successful nursing objective structured clinical examination. J Nurs Educ. 2010;49(1):36–41. https://doiorg.publicaciones.saludcastillayleon.es/10.3928/01484834-20090915-01.
Martin RD, Ngo N, Silva H, Coyle WR. An objective structured clinical examination to assess competency acquired during an introductory pharmacy practice experience. Am J Pharm Educ. 2020;84(4). https://doiorg.publicaciones.saludcastillayleon.es/10.5688/ajpe7625.
Laio SC, Chen JD, Yeh JH, Yip PK. Assessment of communication skills in a group of Seventh-Year medical students with the objective structured clinical examination (OSCE). Formos J Med. 2011;15(4):356–63. http://lawdata.com.tw/tw/detail.aspx?no=230130.
Sholadoye TT, Tolani MA, Aminu MB, Maitama HY. Clinical examination among medical students: assessment and comparison of the strengths and weaknesses of objective structured clinical examination and conventional examination. Nigerian J Surg. 2019;25(2):208–12. https://www.ajol.info/index.php/njs/article/view/215873.
Chang CC, Laio MN, Dou SJ, Hu ST. An evaluation of preceptors in nursing administration practice. J Nurs. 2000;47(6):32–42. https://doiorg.publicaciones.saludcastillayleon.es/10.6224/JN.47.6.32.
Lin CF. The research on inquiring Pingtung County elementary school teachers’ views about the tentative teacher professional development evaluation. Master’s Thesis, National Pingtung University of Education, Pingtung County, Taiwan, 2010. Available online: https://hdl.handle.net/11296/6k8n4s (accessed on 28 March 2025).
Huang SY. March Relationships Between Personality Traits and Teaching Self-Efficacy in Clinical Nursing Preceptors. Master’s Thesis, Chang Jung Christian University, Tainan City, Taiwan, 2016. Available online: https://hdl.handle.net/11296/uxw83d (accessed on 30 2025).
Chang SH. Research Methods: Social Science and Biomedical Methodology; Wu-Nan Book: Taipei, Taiwan, 2020; Available online: https://reurl.cc/QLVRdo (accessed on 03 March (2025).
Babbie ER. The Practice of Social Research, 15th ed.; Cengage: Boston, MA, USA, 2021; Available online: https://reurl.cc/9Gb78O (accessed on 05 March 2025).
Gay LR, Airasian P. Educational research: Consequences for analysis and applications. 2003. https://reurl.cc/M8alLp
Purnajati OG, Rachmadya NH, Rahayu GR. Does objective structured clinical examination examiners’ backgrounds influence the score agreement? Asia Pac Scholar. 2021;6(2):48. https://doiorg.publicaciones.saludcastillayleon.es/10.29060/TAPS.2021-6-2/OA2367.
Hawezy DJ, Barzinjy ST, Saadadin G, Hawezy GA, Baqer W. Perception of undergraduate medical students and examiners towards grand objective structured clinical examination. BMC Med Educ. 2024;24(1):1392. https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12909-024-06388-w.
Acknowledgements
The authors would like to express their sincere gratitude to the clinical instructors for their valuable feedback and contributions to the test. We also extend our heartfelt thanks to all colleagues in the Department of Medical Education at Changhua Christian Hospital. Additionally, we would like to acknowledge the member of the Editorial Board, Associate Editor Dr. Kirsi Forsberg, Assistant Editor Dr. Sridhar Thopella, Editor Dr. Irene Pala, and the anonymous reviewers for their meticulous review and insightful suggestions, which have greatly enhanced the quality of our manuscript.
Funding
The authors received no financial support for the research, authorship, and/or publication of this article.
Author information
Authors and Affiliations
Contributions
C.-W.L., C.-N.C., W.-S.H., W.-L.H. and Y.-J.T. contributed meaningfully to this study. All authors have read and agreed to the published version of the manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
The procedures employed in this study complied with the ethical standards of the relevant institutional and national research committees and adhered to the principles of the 1964 Declaration of Helsinki and its subsequent amendments or equivalent ethical guidelines. The study was approved by the Institutional Review Board (IRB) of Changhua Christian Hospital (CCH) under reference number 220418, with approval granted on May 4, 2022. All participants were fully informed about the study’s objectives, procedures, and their rights, and written informed consent was obtained prior to participation.
Consent for publication
Not applicable.
Conflicts of interest
The authors declare no potential conflict of interest with respect to the research, authorship, and/or publication of this article.
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The original version of this article was revised: the authors identified a spelling error in the name of author Yi-Jung Tsai.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Liao, CW., Cheng, CN., Ho, WS. et al. Enhancing feedback mechanisms in medical education: insights from clinical instructors in a Taiwanese medical center. BMC Med Educ 25, 575 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12909-025-07128-4
Received:
Accepted:
Published:
DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12909-025-07128-4