Abdominal Ultrasound Performance Assessment: A Comparison of Generic and Extended OSCE Rating Scales
Authors:
Simon
Schneider
1
Daniel
Stricker
2
Christoph
Berendonk
2
Joanna
Baron-Stefaniak
3
Dr. med.
Robin
Walter
1✉,3
Emailrobin.walter@unibe.ch
1
Institute of Primary Health Care, Faculty of Medicine
University of Bern
Mittelstrasse 43
3012
Bern
Switzerland
2A
Institute for Medical Education, Faculty of Medicine
University of Bern
Switzerland
3
Department of Anaesthesiology and Intensive Care Medicine
Barmherzige Schwestern Krankenhaus
Vienna
Austria
Simon Schneider1
Daniel Stricker2
Christoph Berendonk2
Joanna Baron-Stefaniak3
Robin Walter1, 3
Affiliations
1 Institute of Primary Health Care, Faculty of Medicine, University of Bern, Switzerland
2 Institute for Medical Education, Faculty of Medicine, University of Bern, Switzerland
3 Department of Anaesthesiology and Intensive Care Medicine, Barmherzige Schwestern Krankenhaus, Vienna, Austria
Correspondence to:
Dr. med. Robin Walter
Department of Anaesthesiology and Intensive Care Medicine, Barmherzige Schwestern Krankenhaus, Vienna, Austria
and
Institute of Primary Health Care, Faculty of Medicine, University of Bern, Switzerland
Mittelstrasse 43
3012 Bern
robin.walter@unibe.ch
A
Abstract
Objective Structured Clinical Examinations (OSCE) are a widely used tool for assessing ultrasound competence, yet the optimal format for its rating scales remains debated. Extended, task-specific rating scales may offer detailed guidance, while more generic scales offer broader applicability and ease of use. This study compares whether and how ratings differed when using an extended versus a generic rating scale for assessing performance in basic abdominal ultrasound skills.
Methods
A
In this single-centre cohort study, 80 medical students participated in an OSCE for abdominal ultrasound rated by both an extended and generic rating scale in parallel. Five domains were evaluated: image settings, transducer handling, examination technique, image explanation and overall performance. Each OSCE station was rated by two assessors simultaneously, one using each scale.
Results
The generic rating scale demonstrated significantly higher internal consistency (mean Cronbach’s α generic 0.803 vs. extended 0.699; p = 0.011) and a higher generalisability (Phi generic 0.529 vs. extended 0.466). Absolute ratings on the generic rating scale were significantly lower (performance P generic 75.38% vs. extended 77.99%, p < 0.001), though in-between stations difficulty was only significantly different for two stations.
Conclusion
In total, the generic rating scale showed a more stringent rating with a higher internal consistency and a higher variance due to participants’ performance.
Therefore, a well-designed generic rating scale can reliably and efficiently assess basic abdominal ultrasound performance, possibly offering greater generalisability and simpler implementation than extended task specific rating scales.
Key Words
Ultrasound education
clinical evaluation
assessment techniques
rating scales
near-peer teaching
near-peer assessment
A
A
A
A
Background
Objective structured clinical examinations (OSCE) are a common examination format for ultrasound courses as it captures its manifold facets best [1–5]. Their organisation, administration, and execution demand substantial knowledge, experience, and planning [6, 7]. The creation of the OSCE rating scale is a key component of this process [7]. OSCE assessment requires rating scales to capture as much information as possible within limited observation time. While generic rating scales have shown advantages for less technical skills [8–10], it remains debated whether this also applies to technical skills such as ultrasound [11, 12].
Accordingly, recent literature on ultrasound exams generally distinguishes into two options: task-specific (i.e. extended) rating scales [3, 13] and generic rating scales [1, 5, 14]. Both types of rating scales, generic and extended, seem to provide advantages [15]. Supporters of extended rating scales argue they offer greater reliability and validity [3, 13, 16]. In contrast, supporters of a generic rating scale contend that the use of elaborate and procedure-specific tools may not always provide a better estimate of performance than scales relying on general competencies [5, 15, 17, 18]. Furthermore, generic rating scales are easier to develop and can be quickly adapted to new OSCE stations, thereby simplifying OSCE organisation and coordination. Despite those facts, there has never been a head-to-head comparison of a generic and a task-specific extended rating scale for abdominal ultrasound; they have all been proposed and validated individually.
In this study, we aimed to compare a generic and an extended rating scale evaluating the same domains. We investigated whether and how ratings differed in terms of internal consistency, generalisability and performance scores when using an extended versus a generic rating scale for assessing performance in basic abdominal ultrasound skills.
Methods
Context
The Young Sonographers, a subsection of the Swiss Society for Ultrasound in Medicine (SGUM/SSUM), have provided courses on basic abdominal ultrasound since 2019, comprising five hours of e-learning and 16 hours of practical training taught by near-peer tutors [19, 20]. Following a revision of this course on abdominal ultrasound [21], existing OSCE stations and their task-specific extended rating scales [13] were updated accordingly (unpublished thesis, available on request). Based on this revision, two rating scales were compared: one extended, task-specific version and one generic scale using the same domains.
Study design
This single-centre cohort study was conducted at the University of Bern to compare both rating scales in the abdominal ultrasound OSCE. Eighty medical students (3rd − 6th year; 65% female; mean age 23 years) who had completed the course and consented to participation were included. The study aimed to examine how both rating scales (extended, task-specific vs. generic) performed in terms of internal consistency and generalisability. Performance scores P (expressed as percentages out of a maximum of 30 points) were analysed across the rating scales, series and stations to better understand the sources of variability.
OSCE organisation
Twelve OSCE stations each assessed one course’s learning objective through a 4-minute practical ultrasound task. Stations were combined into eight six-station series (A1–D2; Appendix 1). Participants alternated between examiner and model roles.
Each series was completed by two groups (n = 11/12) of participants, except for Series D1 and D2, which were completed by only one group each (n = 6).
Rating scales
Both rating scales used identical domains - image settings, transducer handling, examination technique, image interpretation, and overall performance - with a total of 30 points. The extended scale contained task-specific descriptors (Appendix 2), whereas the generic scale provided brief, uniform prompts (Fig. 1).
A Figure 1 Generic rating scale
|
|
Criterion
|
Never/Very bad
|
Poor/much support
|
medium
|
good
|
Always/very good
|
|
Image settings (focus, gain, penetration depth)
|
0
|
1
|
2
|
3
|
4
|
5
|
|
Transducer handling (orientation, positioning, coupling, examination speed)
|
0
|
1
|
2
|
3
|
|
Examination (depending on the task: speed, examination in several planes, measurement)
|
0
|
1
|
2
|
3
|
4
|
5
|
6
|
7
|
8
|
9
|
10
|
|
Image interpretation (showing/circling, naming, explanations)
|
0
|
1
|
2
|
3
|
4
|
5
|
6
|
7
|
8
|
|
Overall performance
|
0
|
1
|
2
|
3
|
4
|
|
Total score
|
/30
|
| Integer point ratings are possible. |
Data collection
Each OSCE station was rated by two assessors simultaneously, one using each scale. Assessors were randomly assigned and included two SGUM-accredited physicians and 23 trained near-peer tutors with ≥ 1 year of teaching experience. Separate information sessions were held for each group, which conveyed basics for the OSCE itself and the study format, as well as instructions on how to use the rating scales, followed by two exemplary training videos to ensure a common frame of reference [22]. For each video existed an optimal rating score proposed by experienced ultrasound assessors, which was used to lay ground for the discussion among the raters.
Statistical Analysis
To evaluate reliability, internal consistency was quantified using Cronbach's α, as a number between 0 and 1 [23]. We considered a Cronbach’s α of 0.7 and greater suitable for summative examinations such as ours [3, 12]. Generalisability Theory (GT) was used to assess the influence of non-rating scale-related factors on the final score (i.e. how much of variance was explained by the participants performance). Phi coefficients were calculated, ranging from 0 to 1, with 1 meaning that 100% of the variance in scores is due to the respondent's performance [24]. Phi coefficients were calculated for the following model: “(examinee : series) x station”. A Phi coefficient of 0.5 or higher was deemed adequate for this OSCE.
A multivariate split-plot ANOVA was used to examine how rating scale (within-subjects), series and OSCE station (between-subjects) contributed to variability in performance scores (α ≤ 0.05). Effect sizes were reported using partial η2, interpreted as small (< 0.07), medium (0.07–0.14) and strong (> 0.14). The level of significance was adapted using a post-hoc Bonferroni correction. If relevant differences between rating scales were found, paired-sample t-tests were performed to identify responsible OSCE stations.
All statistical analyses were performed using SPSS for Windows, version 28 (IBM, Armonk, NY, USA). For the Generalizability Theory (GT), G_String, version 2019 was used.
Ethical considerations
A
The ethics committee of Bern, Switzerland declared this study was exempt from the Human Research Act and the need for a comprehensive ethical appraisal (BASEC number Req-2023-00254). Collected data was treated after being anonymised.
A
All participants were informed about the study, then provided informed consent to participate or could withdraw without consequences on them receiving the course certificate.
A
A
This study was conducted in accordance with the ethical principles outlined in the Declaration of Helsinki.
Results
Internal consistency and generalisability of rating scales
Our results indicate that internal consistency was higher for the generic scale than for the extended rating scale (Cronbach’s α 0.803 vs 0.699; p = 0.011, Table 1). Across stations, Cronbach’s α ranged from 0.50–0.83 for the extended and 0.74–0.87 for the generic scale (Appendix 4). The generalisability analysis (GT) revealed a Phi coefficient of 0.466 for the extended and 0.529 for the generic scale (Table 1), suggesting that generic scales show greater performance-related variance than extended scales. The main results for both rating scales are indicated in Table 1.
Table 1
Participants’ performance P, Cronbach’s, and Generalisability Phi per rating scale
|
α
|
| |
|
Extended rating scale
|
Generic rating scale
|
Significance
|
|
Cronbach’s α
|
Mean
|
0.699
|
0.803
|
p = 0.011*
|
| |
Min.
|
0.500
|
0.739
|
|
| |
Max.
|
0.827
|
0.873
|
|
|
Generalisability (Phi)
|
|
0.466
|
0.529
|
|
|
Performance P [%]
|
Mean
|
77.99
|
75.38
|
p < 0.001*
|
| |
SD
|
7.13
|
7.67
|
|
|
*Indicated p-values derive from ANOVA or t-test analyses between rating scales, all significant
|
Influence of rating scale on examinees’ performance
In multivariate testing, using a split-plot ANOVA we found a significant interaction between type of rating scale and performance P (F(1, 461) = 22.95, p < 0.001, η² = 0.047; small effect), in other words the two rating scales rated participants’ performance significantly different with the generic rating scale attributing lower scores than the extended scale.
Additionally, “series” and “station” showed significant effects (F(7, 461) = 2.22, p = 0.032, η² = 0.033; F(11, 461) = 3.95, p < 0.001, η² = 0.086; small–medium effects) on participants’ performance P. Though, a significant interaction between type of rating scale and station was found (F(11, 461) = 2.782, p = 0.002, η2 = 0.062, small effect size), the interaction between type of rating scales and series was not significant (p = 0.108), suggesting the difference between scales was balanced across the 12 series.
Finally, we tested individual OSCE stations between rating scales to identify whether all stations are responsible for the significant overall difference using paired samples t-tests. Analyses yielded statistically significant differences between participants’ performance P only for the stations “retroperitoneum” and “hepatic veins” (p < 0.001, Appendix 3).
Discussion
In this single-centre cohort study, we evaluated whether a generic and an extended rating scale for abdominal ultrasound skills yield comparable ratings on examinees’ performance in OSCE. We found that a generic rating scale rated participants’ performance with a higher internal consistency and generalisability, while attributing lower rating scores overall.
Our findings are aligned with broader medical education literature questioning the superiority of task-specific rating scales in performance-based exams. Early on, Regehr et al. showed that global rating scales demonstrate higher reliability and validity than checklists [10], mirroring our observation that the generic scale produced more consistent ratings. Studies comparing generic domain-based versus extended checklist-based OSCE assessments likewise report that generic ratings better discriminate between levels of competence [8, 9]. Although some authors argue that generic ratings scales may be less appropriate for technical skills [11, 12], our results show that generic rating scales can also yield reliable assessments in a highly technical domain such as ultrasound.
Within ultrasound education, our results complement the ongoing debate on the usefulness of detailed, task-specific rating scales. Hofer et al. advocated for extended scales to capture procedural complexity [3]. However, in our study, extended scales produced inflated scores particularly in anatomically complex stations (hepatic veins, retroperitoneum), suggesting that extensive rating scales may inadvertently mask performance differences. More recent work supports moving away from highly granular rating scales: Ma et al. identified critical point-of-care ultrasound rating items but recommended restricting assessment to essential behaviours [15], and Tolsgaard et al. emphasised evaluation of core domains, such as image optimisation, systematic examination, and interpretation, rather than long lists of procedural steps [17]. Our direct comparison provides empirical support for this shift, indicating that generic scales may better capture ultrasound performance by focusing on broader domains that reflect the integrated nature of the skill.
Taken together, our findings suggest that generic rating scales may be a suitable choice when designing OSCE assessments for basic abdominal ultrasound. They may simplify the development of new stations and reduce examiner training burden by avoiding the creation of detailed, station-specific rating scales. Further research should explore whether these findings extend to less technical areas of medical education and whether faculty and near-peer assessors perform differently when using generic scales.
Strength and limitations
This study is the first to directly compare two rating scale formats for abdominal ultrasound OSCE assessment. To minimise cross-contamination of rating styles between assessors, we separated examiners into different break rooms and rotated them regularly. Our use of a diverse examiner group, including faculty and trained near-peer tutors, reflects real-world practice and supports generalisability for technical skills examination such as ultrasound. However, because only a small proportion of examiners were faculty, subgroup analysis was not feasible; results might differ in an assessor cohort composed entirely of faculty. Transferability to other OSCE contexts may also be limited by the narrow and well-defined task of abdominal ultrasound, where assessors share a relatively uniform mental model of optimal performance; conditions that may not hold in broader clinical domains.
Conclusion
For abdominal ultrasound OSCEs, generic rating scales appear to offer more stringent rating behaviour, higher internal consistency, and greater performance-related variance than extended scales. A well-designed generic rating scale can reliably and efficiently assess basic abdominal ultrasound competence, potentially offering greater generalisability and simpler implementation than extended, task-specific rating scales.
Abbreviations
OSCE Objective structured clinical examinations
SGUM/SSUM Swiss society for ultrasound in medicine
ANOVA Analysis of variance
GT Generalizability theory
Declarations
Ethical considerations and consent to participation
The ethics committee of Bern, Switzerland declared this study was exempt from the Human Research Act and the need for a comprehensive ethical appraisal (BASEC number Req-2023-00254). Collected data was treated after being anonymised. All participants were informed about the study, then provided informed consent to participate or could withdraw without consequences on them receiving the course certificate. This study was conducted in accordance with the ethical principles outlined in the Declaration of Helsinki.
A
Author Contribution
SS, CB, DS and RW contributed to the conception of the work. RW, SS, CB and DS contributed to the design of the work. SS and RW did the acquisition, and DS performed the primary analysis. SS, DS, CB and RW interpreted the data. SS wrote the main manuscript and CB, DS, JB and RW substantively revised it. All authors have read and approved the final manuscript.
Manuscript funding
This work did not receive external funding.
A
Data Availability
The used data are available from the corresponding author upon request.
Electronic Supplementary Material
Below is the link to the electronic supplementary material
A
Acknowledgement
We would like to thank all the students, near-peer tutors and faculty tutors who participated in this study.
Bibliography
1.Bailitz J, O’Brien J, McCauley M, et al. Development of an expert consensus checklist for emergency ultrasound. AEM Educ Train. 2022;6:e10783. https://doi.org/10.1002/aet2.10783.
2.Bell C, Hall AK, Wagner N, et al. The Ultrasound Competency Assessment Tool (UCAT): Development and Evaluation of a Novel Competency-based Assessment Tool for Point‐of‐care Ultrasound. AEM Educ Train. 2021;5:e10520. https://doi.org/10.1002/aet2.10520.
3.Hofer M, Kamper L, Sadlo M, et al. Evaluation of an OSCE Assessment Tool for Abdominal Ultrasound Courses. Ultraschall Med - Eur J Ultrasound. 2011;32:184–90. https://doi.org/10.1055/s-0029-1246049.
4.Norcini J, Anderson B, Bollela V et al. Criteria for good assessment: Consensus statement and recommendations from the Ottawa 2010 Conference. Med Teach. 2011;33:206–14. https://doi.org/10.3109/0142159X.2011.551559
5.Teichgräber U, Ingwersen M, Ehlers C, et al. Integration of ultrasonography training into undergraduate medical education: catch up with professional needs. Insights Imaging. 2022;13:150. https://doi.org/10.1186/s13244-022-01296-3.
6.Khan KZ, Gaunt K, Ramachandran S, Organisation, et al. Adm Med Teach. 2013;35:e1447–63. https://doi.org/10.3109/0142159X.2013.818635.
7.Melcher P, Zajonz D, Roth A et al. Peer-assisted teaching student tutors as examiners in an orthopedic surgery OSCE station – pros and cons. GMS Interdiscip Plast Reconstr Surg DGPW 5Doc17 2016. https://doi.org/10.3205/IPRS000096
8.Giemsa P, Wübbolding C, Fischer MR, et al. What works best in a general practice specific OSCE for medical students: Mini-CEX or content-related checklists? Med Teach. 2020;42:578–84. https://doi.org/10.1080/0142159X.2020.1721449.
9.Pinilla S, Lerch S, Lüdi R, et al. Entrustment versus performance scale in high-stakes OSCEs: Rater insights and psychometric properties. Med Teach. 2023;45:885–92. https://doi.org/10.1080/0142159X.2023.2187683.
10.Regehr G, MacRae H, Reznick RK, et al. Comparing the psychometric properties of checklists and global rating scales for assessing performance on an OSCE-format examination. Acad Med. 1998;73:993–7. https://doi.org/10.1097/00001888-199809000-00020.
11.Mahmoud A. A Comparison of Checklist and Domain-Based Ratings in the Assessment of Objective Structured Clinical Examination (OSCE) Performance. Cureus 2023. https://doi.org/10.7759/cureus.40220
12.Daniels VJ, Pugh D. Twelve tips for developing an OSCE that measures what you want. Med Teach. 2018;40:1208–13. https://doi.org/10.1080/0142159X.2017.1390214.
13.Vetsch A, Berendonk C, Hari R. Reliabilität und Durchführbarkeit einer Ultraschallprüfung für Studierende in der Schweiz. Prim Hosp Care Médecine Interne Générale. 2020. https://doi.org/10.4414/phc-f.2020.10247.
14.Walzak A, Bacchus M, Schaefer JP, et al. Diagnosing Technical Competence in Six Bedside Procedures: Comparing Checklists and a Global Rating Scale in the Assessment of Resident Performance. Acad Med. 2015;90:1100–8. https://doi.org/10.1097/ACM.0000000000000704.
15.Ma IWY, Desy J, Woo MY, et al. Consensus-Based Expert Development of Critical Items for Direct Observation of Point-of-Care Ultrasound Skills. J Grad Med Educ. 2020;12:176–84. https://doi.org/10.4300/JGME-D-19-00531.1.
16.Lockyer J, Carraccio C, Chan M-K, et al. Core principles of assessment in competency-based medical education. Med Teach. 2017;39:609–16. https://doi.org/10.1080/0142159X.2017.1315082.
17.Tolsgaard MG, Todsen T, Sorensen JL, et al. International Multispecialty Consensus on How to Evaluate Ultrasound Competence: A Delphi Consensus Survey. PLoS ONE. 2013;8:e57687. https://doi.org/10.1371/journal.pone.0057687.
18.Chenot J-F, Simmenroth-Nayda A, Koch A, et al. Can student tutors act as examiners in an objective structured clinical examination? Med Educ. 2007;41:1032–8. https://doi.org/10.1111/j.1365-2923.2007.02895.x.
19.Hari R, Kälin K, Birrenbach T, et al. Near-peer compared to faculty teaching of abdominal ultrasound for medical students – A randomized-controlled trial. Ultraschall Med - Eur J Ultrasound. 2024;45:77–83. https://doi.org/10.1055/a-2103-4787.
20.Räschle N, Hari R. «Blended Learning»-Basiskurs Sonografie: Im Peer-Tutoring zu einem SGUM-akkreditierten Ultraschall-Kurs. Praxis. 2018;107:1255–9. https://doi.org/10.1024/1661-8157/a003116.
21.Walter R, Räschle N, Blunier A-S, et al. Revision des e-Learnings zum Grundkurs Abdomen der Young Sonographers. Praxis. 2022;111:495–501. https://doi.org/10.1024/1661-8157/a003907.
22.Woehr DJ, Huffcutt AI. Rater training for performance appraisal: A quantitative review. J Occup Organ Psychol. 1994;67:189–205. https://doi.org/10.1111/j.2044-8325.1994.tb00562.x.
23.Tavakol M, Dennick R. Making sense of Cronbach’s alpha. Int J Med Educ. 2011;2:53–5. https://doi.org/10.5116/ijme.4dfb.8dfd.
24.Li M, Shavelson RJ, Yin Y, et al. Generalizability Theory. In: Cautin RL, Lilienfeld SO, editors. Encycl. Clin. Psychol. 1st ed. Wiley; 2015. pp. 1–19. https://doi.org/10.1002/9781118625392.wbecp352.