Next Article in Journal
Evaluating Chat GPT-4o’s Comparative Performance over GPT-4 in Japanese Medical Licensing Examination and Its Clinical Partnership Potential
Previous Article in Journal
From Written Tests to OSCE: A Study on the Perceptions of Assessment Reform by Students and Faculty in the French Dental Curriculum
error_outline You can access the new MDPI.com website here. Explore and share your feedback with us.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feasibility of Script Concordance Test Development: A Qualitative Study of Medical Educators’ Experiences

by
Reem M. Alhossaini
1,2,*,
Anthony Richard Cox
1 and
Sarah Katie Pontefract
1
1
Department of Pharmacy, School of Health Sciences, College of Medicine and Health, University of Birmingham, Birmingham B15 2TT, UK
2
Department of Pharmacy Practice, College of Pharmacy, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
*
Author to whom correspondence should be addressed.
Int. Med. Educ. 2026, 5(1), 8; https://doi.org/10.3390/ime5010008
Submission received: 11 October 2025 / Revised: 15 December 2025 / Accepted: 22 December 2025 / Published: 7 January 2026

Abstract

Script Concordance Tests (SCTs) assess clinical reasoning under uncertainty. While construction guidelines exist, the feasibility of collaborative development approaches and educators’ real-time experiences remain underreported. This feasibility study explores how medical educators construct SCTs collaboratively and their perceptions of the process. Four UK-based medical educators developed SCTs for prescribing in older adults during a three-hour workshop involving observation with a think-aloud approach, followed by a post-workshop focus group. Data were analysed using Braun and Clarke’s thematic analysis, Tuckman’s group development informed observation analysis, and an inductive approach for the focus group. Educators created seven vignettes (30 items) in 127 min, averaging 18 min per vignette. Observation revealed small-team role specification (lead, scribe, challenger) and three themes: content development, quality checks, and team dynamics. The team progressed rapidly through Tuckman’s stages, spending most of the time in the performing stage. Focus group analysis revealed four themes: design features, perceived utility, group dynamics and best-practice recommendations. This study demonstrates the feasibility of collaborative SCT development through structured teamwork. Educators perceived SCTs as practical, as valuing effective team dynamics and clear role distribution. Findings can provide practical insights for institutions implementing SCT development, emphasising practice sessions and appropriately sized collaborative teams.

1. Introduction

In medical practice, clinical reasoning and decision-making in uncertain situations are foundation skills [1]. Prescribing medication is a complex task that requires clinical reasoning and decision-making abilities [2]. Reflecting these competencies, existing research suggests that inadequate prescribing skills, such as inappropriate prescribing, irrational prescribing, or other forms of prescribing competency issues, can compromise patient safety [3,4]. Thus, there is a necessity for effective prescribing skills remedies through training and assessment in medical education and continuing professional development.
Every training or intervention aimed at improvement should be paired with appropriate assessment methods. Assessment utility depends on multiple factors including validity, reliability, educational impact, acceptability, feasibility, and cost-effectiveness [5].
Despite this framework, research has prioritised developing training tools over creating tailored assessments for prescribing competencies, particularly in uncertain clinical situations. A systematic review by Kamarudin and Penm (2013) reported various interventions but noted limited information on the validity and reliability of the associated assessment tools [4]. Similarly, Shi and Qin’s recent review (2023) highlighted the need for validated and reliable assessment measures in future studies [6].
Traditional assessment methods often fail to capture these nuanced cognitive processes involved in prescribing. Script Concordance Tests (SCTs) have emerged as a promising tool designed to evaluate reasoning skills under uncertainty [7].
Developed by Bernard Charlin and colleagues, building upon earlier work with Diagnostic Script Questionnaires [7,8] SCTs are validated assessment tools widely used across healthcare education [9,10,11]. A concise outline of the SCT structure is available in Supplementary File S1.
However, guidelines exist [12,13] and systematic reviews [10] provide theoretical frameworks without capturing actual construction processes. While many SCT studies examine validity and learner performance [9,10,11] or provide construction principles [7,12,13]; the practical experiences of educators during SCT development, i.e., their real-time decisions, collaborative negotiations, quality considerations, and perceptions, remain undocumented. This created three knowledge gaps: First, process transparency: no studies capture how educators actually translate guidelines into practice, including time requirements, decision points, and problem-solving strategies.
Second, collaborative evidence: guidelines recommend single or paired construction [12,13]. However, there is no evidence to support team-based approaches that institutions may need to adopt.
Third, practitioner insights: many studies validate SCT effectiveness [9,11], educators’ experiences, challenges, and lessons learned during construction remain absent from the literature.
Prior to allocating resources for widespread SCT implementation, evidence is required regarding the practical feasibility of collaborative development approaches. Feasibility studies evaluate whether an intervention or process can be successfully delivered with standard institutional limitations, assessing factors such as time, resource requirements, stakeholders’ acceptability, and preliminary evidence of utility [14]. This feasibility evidence is essential for informing decisions regarding their implementation on a large scale and identifying potential barriers or facilitators to adoption in educational settings.
This two-part qualitative feasibility study explores how one UK team of medical educators (developing an SCT for subsequent use in a randomised controlled trial RCT with final-year medical students) constructed SCTs for prescribing and experienced that process. Specifically, the study examined whether collaborative SCT development is feasible, acceptable, and efficient. We observed a live in-person SCT-building workshop using a think-aloud approach; followed by a post-workshop focus group. The study aims to provide practical design insights into team dynamics and design processes that can inform effective collaborative SCT development for assessing clinical reasoning under uncertainty.

2. Methods

This study employed a two-part qualitative feasibility design using methodological triangulation, consisting of an observational phase followed by a focus group interview. The primary objective was to observe real-time SCT construction processes, while the secondary objectives included exploring educators’ perceptions of the process, perceptions about SCTs and providing best practices for SCT development based on their experiences.
Reporting follows the Consolidated Criteria for Reporting Qualitative Research (COREQ) guideline [15].

2.1. Methodological Triangulation

This study employed methodological triangulation, systematically integrating three data sources: observational data (real-time behaviours), think-aloud protocols (cognitive processes), and focus group discussions (reflective perceptions). Following individual analysis, findings were compared using thematic mapping matrices. Convergent findings strengthened confidence, while divergent perspectives revealed process complexity. This approach captured the ‘what,’ ‘how,’ and ‘why’ of collaborative SCT development.

2.2. Participants and Sampling

Purposive sampling was used, as described by Palinkas et al. [16]. The inclusion criteria were: a minimum of 5 years of experience in both medical education and clinical experience relevant to medication prescribing, and previous experience in creating various types of assessment content both competencies were essential for developing educationally sound SCTs focused on prescribing decisions. Eligible educators and doctors from UK higher education and healthcare institutions were identified through internal networks (S.K.P., A.R.C.) as well as through professional recommendations (including limited snowball sampling).
Six individuals from the two universities and hospitals were invited by email; two withdrew due to scheduling conflicts, and four completed both parts. The four participants were all affiliated with the University of Birmingham, which resulted in pre-existing professional relationships among participants. As in a feasibility study, sample size was determined by the need to test whether a process can be implemented, rather than to achieve thematic saturation [14]. The final sample size of four was determined by practical feasibility constraints (two withdrawals due to scheduling conflicts) rather than priori sample size calculation. However, four participants forming a single collaborative team proved appropriate for examining the feasibility of team-based SCT development, which allows for intensive observation of the collaborative process, time requirements and assessment of acceptability. According to Vasileiou et al. [17] sample size adequacy should be evaluated according to study-specific factors such as research aims, scope and depth of data collected. The feasibility study was designed for intensive process observation rather than achieving thematic saturation or making population inference, prioritising depth of observation within practical constraints.
Professional background, experience, and prior SCT exposure are summarised in Supplementary File S2 Table S1. Participants’ sex, gender, race and ethnicity were not collected as they are not considered relevant to the study’s focus on SCT construction and medical educators’ perceptions of SCT and its development. Participants were purposefully selected from medical and pharmacy disciplines to reflect the interprofessional nature of prescribing practice. This diverse professional representation ensured that SCT development incorporated perspectives from key stakeholders involved in medication management decisions.

2.3. Part 1: The Observational Phase

2.3.1. Aim

To capture a deeper understanding of participants’ experiences and thought processes while formulating SCT, following an in-person workshop introducing core SCT principles and construction guidelines.

2.3.2. Data Collection

  • Location and setting: In-person, a three-hour observational session was conducted at the University of Birmingham’s (UoB) medical school in a seminar room that had movable furniture arranged to facilitate participants’ interactions, with only the participants and the researcher present in that room.
  • Context: The SCTs were developed for subsequent use in an RCT with final-year medical students following an eLearning module; no trial data are reported here.
  • The workshop: The workshop was designed to facilitate the understanding of SCT principles essential for creating good test items. R.M.A. designed and facilitated the workshop following the guidelines described by Fournier, Demeester [12] and Lubarsky, Dory [13]
Structure: 30 min orientation, then around 120 min item writing (main-block), 10 min wrap-up. See Supplementary File S3 Table S2 for a phase-by-phase breakdown (purpose, activities, outputs, roles, and materials).
During the session, participants focused on developing SCTs related to older adult prescribing. They were already familiar with the intended learning outcomes for the module, which were also included among the workshop materials. Participants were tasked to create three vignettes with three items each within a three-hour time frame. Materials included SCT examples, a quality matrix, and templates. Observations were recorded, detailing the process and time spent on each SCT component.
  • Think-Aloud Protocols as described by Ericsson [18], were utilised to understand the cognitive processes behind expert performance, where participants vocalised their thoughts during task execution. This method was chosen for its ability to provide real-time insights and facilitate observation of interactions, in contrast with retrospective protocols, where thoughts are reported after reviewing a performance recording. The researcher (R.M.A.) adopted an observer role, remaining seated to the side, seeking to observe without interacting or influencing the setting, taking field notes. The notes are to document the physical setting, participants’ interactions, communication patterns, team workflow, and group dynamics, complementing the audio recording.

2.3.3. Data Analysis

Data Preparation
The audio recording from the observation section was transcribed verbatim by R.M.A. and uploaded to NVivo 12 Plus (QSR International Pty Ltd., Burlington, MA, USA, 2018) for efficient, organised data coding and analysis. Field notes were documented by R.M.A., including the setting, participants’ interactions, and the workflow, which were integrated with the transcript data during analysis. The identities of participants were anonymised (PE1–PE4) in the transcripts to maintain confidentiality.
Thematic Analysis Approach
Analysis was performed by following Braun and Clarke’s thematic analysis framework [19,20], utilising deductive and inductive approaches to identify patterns within the observation data. Tuckman’s framework [21] was applied deductively, while additional themes related to SCTs construction process emerged inductively from the data.
Phase 1 data familiarisation: R.M.A. read the observation transcript multiple times to familiarise with the data. During the initial phase, preliminary notes were taken regarding recurring ideas, interactions, behaviours, vocalised thoughts and potential patterns relevant to the research focus.
Phase 2 initial coding: R.M.A. conducted initial coding on the complete dataset. S.K.P. independently coded a subset of 20% of the transcripts to enhance credibility; disagreements were resolved through discussion and codebook refinement. The codebook was updated systematically across all analytical phases, with coding decisions, disagreements, and resolutions documented in a decision log serving as our audit trail. This approach ensured framework evolution while maintaining consistency and transparency [22].
Tuckman’s stages of group development—forming, storming, norming, and performing—[21] provided a framework and was applied deductively to observational data. Additional inductive coding was conducted to capture emergent themes related to SCT development processes not accounted for Tuckman’s framework.
See Table 1 For a brief description of Tuckman’s Four stages of group development.
Phase 3 theme development: codes were collected and examined for patterns that aligned with Tuckman’s stages, as well as themes emerging inductively from the data related to SCTs item creation process and workflow within the team. All relevant codes extracted were then gathered within each optional theme, and relationships between themes were mapped to understand group dynamics and workflow throughout the observation period.
Phase 4 theme refinement: Each candidate theme was reviewed against the coded extracts and the entire dataset to ensure internal coherence and a clear distinction between them. Themes were assessed for their fit within Tuckman’s framework and their relevance to the research questions.
Phase 5 theme definition and naming: Each theme was refined through detailed analysis to identify its essence and scope. Clear definitions articulated what each theme captured regarding observed group behaviours, interaction and SCT development strategies, and how it related to the research objectives. Themes and subthemes were assigned descriptive names that conveyed their central concepts.
Phase 6 report production: the final analysis integrated analytical narrative with illustrative quotes from the observation transcript to demonstrate each theme, providing a coherent account of both group development patterns and the process involved in SCT construction.

2.4. Part 2: Post-Workshop Focus Group Interview

2.4.1. Aim

To gather a deeper insight from the perspectives of the same medical educators on the feasibility, acceptability, and experience of creating SCTs collaboratively and their experiences of the development process.

2.4.2. Data Collection

  • Location and setting: The focus group was scheduled shortly after the workshop to ensure participants’ recollections and insights remained fresh and accurate, with only the participants and the researcher present during the focus group.
  • Interview Guide: (R.M.A.) facilitated the focus group, adhering to the guidelines outlined by Jordan, Clarke [23] for conducting interviews in medical education research; a semi-structured interview format was employed, covering topics such as participants’ experiences, SCT comparisons with other assessments, best practices, and recommendations for future SCT development. The interview guide was reviewed by an expert and underwent refinement for content validity and clarity [24].

2.4.3. Data Analysis

Data Preparation
The 55 min session was recorded and transcribed by R.M.A. verbatim into a Microsoft Word document, which was then imported to NVivo 12 Plus. Participants reviewed the transcript and confirmed the accuracy, and comments were reviewed and clarified. Participants’ identities were anonymised, consistent with the observational transcript.
Thematic Analysis Approach
Data were analysed following Braun and Clarke’s thematic analysis [19,20], as recommended by Maguire and Delahunt [22]. This approach involves an inductive approach to identify common themes emerging from participants’ perspectives and experiences. Notably, specific themes was anticipated to align with Van Der Vleuten Formula [5]. Moreover, themes related to best practices and construction methods are expected to be identified.
Phase 1 data familiarisation: R.M.A. read the focus group transcript multiple times to familiarise with the data. Having conducted and transcribed the interview, the R.M.A. was already familiar with the data. To ensure accuracy, the transcript was compared against the original recording multiple times, with R.M.A. re-listening while reading on several occasions to verify transcription fidelity.
During the initial phase, preliminary notes were taken regarding recurring ideas, participants’ reflections on the development process and their impression about SCT [20].
Phase 2 initial coding: R.M.A. conducted initial coding of the complete transcript, systematically identifying features of the data relevant to the research aims. The code captured participants’ perspectives on SCT characteristics, designs, and their experiences in construction. S.K.P. independently coded a subset of 20% of the focus group transcripts to enhance credibility; disagreements were resolved through discussion and iterative codebook refinement [22]. Consistent with the observational analysis, coding decisions, disagreements, and resolutions were documented in our decision log audit trail. The approach ensured consistency between data sources while allowing appropriate framework evolution.
Phase 3 theme development: codes were collected and examined for patterns that could form potential themes. Related codes were grouped, and all relevant coded extracts were gathered within each provisional theme. Particular considerations for SCT development and group collaboration.
Phase 4 theme refinement: Each candidate theme was reviewed against the coded extracts and the entire dataset to ensure internal coherence and a clear distinction between them. Themes were assessed for their relevance to the research objectives concerning feasibility, acceptability and stakeholder experiences. Themes lacking sufficient data support were merged with related themes or discarded.
Phase 5 theme definition and naming: Each theme was refined through detailed analysis to identify its essence and scope. Clear definitions articulated what each theme captured regarding participants’ perceptions and experiences of SCT development. Themes and subthemes were assigned descriptive names that conveyed their central concepts.
Phase 6 report production: the final analysis integrated analytical narrative with illustrative quotes from the focus group transcript to demonstrate each theme, providing a coherent account of medical educators’ perspectives on the perceived feasibility, acceptability, and practical considerations of SCT development.

2.5. Researcher Positioning, Reflexivity and Trustworthiness

The lead author (R.M.A.), a PhD researcher, facilitated the workshop, conducted non-participant observation, and carried out the interviews. The researcher’s dual roles as both facilitator and observer created potential for influencing group dynamics and participants responses. To minimise observer effects, the researcher maintained a peripheral seating position during development activities and avoided directive interventions. Participants were explicitly informed that the researcher’s role was observational rather than evaluative. We acknowledge that complete elimination of observer influence is impossible, and to mitigate potential bias from R.M.A.’s dual role as facilitator and analyst, S.K.P. independently coded 20% of transcripts. Disagreements were resolved through consensus meetings. R.M.A. has completed training in qualitative research methods and holds a certification in Good Clinical Practice. The researcher had a mixed pre-existing relationship with participants: one established relationship, two minimal with professional contact, and one with no prior relationship. Participants in the focus group interview reviewed the transcript excerpts for accuracy; minor clarifications were incorporated.

2.6. Ethics

The study adhered to ethical principles, obtaining written informed consent and ensuring data confidentiality. Participation was voluntary; all participants provided written informed consent for participation, audio recording, and use of anonymised data. The University of Birmingham Ethics Committee approved the study protocol [ERN_21-1350] before data collection began.

3. Results

3.1. Sampling

Six individuals met the inclusion criteria and agreed to participate; two withdrew due to scheduling conflicts. Four completed both the SCT construction and the focus group session. The participants represented diverse disciplines: medicine (n = 1), clinical pharmacy (n = 2), and pharmacy (n = 1). Primary practice settings included university settings (n = 2) and tertiary hospital settings (n = 2). All participants had ≥10 years of clinical practice. Three had ≥ 10 years of education experience; one had 6–9 years. Prior SCT exposure was limited, with one having no previous exposure and three having limited exposure. Pre-existing relationships varied: three were departmental colleagues, and one was a result of a prior collaboration. (Supplementary File S2 Table S1).

3.2. Part 1: Observational Study Findings

3.2.1. Feasibility Outcomes

Task Achievement and Efficiency
Participants produced seven SCT vignettes within a three-hour time frame. Each vignette included three to five items, with a total of 30 items. The total construction time was 127 min, giving an average of 18 min per vignette and four items per vignette. Table 2 summarises the timings and item counts during the development. A sample of one SCT that was produced by participants is used in Supplementary File S1 SCT Structure Example.
Beyond the development timings, Table 3 summarises key feasibility indicators observed in our study. The evidence suggests that collaborative SCT development demonstrated feasibility across multiple dimensions within our study context. Time efficiency was evidenced through consistent development rates, while participant acceptance was observed through sustained engagement and positive feedback. Workload appeared manageable through collaborative processes, and resource requirements remained minimal.

3.2.2. Content Coverage

Cases addressed common and challenging issues in prescribing for older adults. Topics included delirium, falls, acute kidney injury, urinary tract infections, adverse drug reactions, drug interactions, postural hypotension, hospital discharges, medication reviews, and the assessment of treatment risks versus benefits.

3.2.3. Group Development

The team progressed rapidly through the group development process. Overview of key behaviours, interactions and decision-making observed at each of Tuckman’s stages of group development during SCT construction, Table 4, highlights how role allocation, conflict management and collaborative problem-solving contribute to team function. Figure 1 represents the group’s trajectory progression through Tuckman’s stages, with the performing stage occupying the largest proportion of the timeline observed, followed by shorter periods of forming and storming.

3.2.4. Key Behavioural Patterns Observed

  • Leadership and Roles:
All team members actively participated and played important roles in generating ideas. Roles emerged naturally and without conflict. Certain members assumed prominent responsibilities:
  • Participant 2 emerged as a content expert and primary idea generator, often leading discussions and providing clinical insights.
    Participant 1 facilitated group discussions and served as the scribe, summarising ideas to ensure clarity and focus.
    Participant 3 and Participant 4 provided validation and alternative perspectives.
  • Communication patterns:
  • The team exhibited effective communication strategies, including:
    Actively listening to ensure all viewpoints were considered.
    Supportive responses to uncertainty
    Building on each other’s ideas
    Constructively challenging ideas to refine SCT development.
    Using humour to maintain a positive atmosphere.

3.2.5. Think-Aloud Analysis

Transcript analysis yielded three key themes related to SCT development and group dynamics:
1.
Content Development: This theme captures the foundational aspects of SCT item creation, which include two subthemes:
A.
Learning/Mastering SCT Format: This refers to checking question format and requirements. Participant 2 responds to the question about case linkages by stating, “Well, it doesn’t matter, because the format is that you’re not going to link them” (PE2). Another way the participants mastered the SCT was by working backwards, writing items starting from the outcomes and ending with writing the case.
B.
Developing Clinically Authentic Cases: included illustrations of incorporating realistic clinical scenarios, discussing the cases based on their clinical experience and mapping these cases to the task of older adult prescribing. Participant 1 reminded the others about the cases’ alignment with the module “With a past medical history of COPD, which is not in the older adult module.” (PE1).
2.
Assessment of Quality and Efficiency: This theme emphasises the refinement of SCT items in which participants, during the development process, aim to ensure high-quality and effective items. This theme is divided into three subthemes.
A.
Iterative Question Drafting Process: This process is marked by a dynamic, rapid generation of draft questions through continuous brainstorming within the team. Participants actively seek input from each other, where they specifically ask for feedback on various components, such as question scoring and continuous re-evaluation of the cases and items in anticipation of the expert panel review. As Participant 3 articulated a concern about unanimous expert responses, noting, “If all the experts answer the same, then it [the items] all goes.” (PE3), suggesting the potential impact of unanimous expert opinions.
B.
Balancing Question Difficulty and Relevance: This involves avoiding simple questions and incorporating nuanced clinical details to ensure clinical judgement is required. Participant 1 rejected an easy question by stating, “That’s too easy.” (PE1) The team discussed the complexities of a vitamin D deficiency scenario, highlighting the importance of understanding the implications of prescribing calciferol for calcium and vitamin D levels, thus preventing direct, predictable questions in a clinical context.
C.
Crafting Challenging Judgement-based Options: This subtheme focuses on achieving a balance of ambiguity to elicit a spread of responses. One recommends “make it a bit more ambiguous, not ambiguous a little less obvious,” (PE3), which is believed to lead to a variety of responses “that’s going to lead to a nice spread across experts.” (PE2). This approach aims to elicit nuanced and subtle judgments, as highlighted by observations such as “It’s quite a nuanced sort of decision” and “That’s very subtle, and it’s a judge.” (PE2)
3.
Team Dynamics
A.
Differing Perspectives: This subtheme captures instances where participants expressed opposing viewpoints, such as question difficulty, relevance, and scoring.
B.
Collaborative Process: This involves participants building on each other’s ideas and combining their clinical expertise to refine SCT items collaboratively. An example of this dynamic is when (PE2) suggests considering prior actions and implies resuming simvastatin, currently on hold due to clarithromycin treatment, (PE1) built on this idea considering re-prescription, which showcases effective teamwork in leveraging collective knowledge.
C.
Insightful Discussions: The team engaged in profound discussions about various topics. For example, they debated students’ understanding of questions, with Participant 1 noting that “Students may find a question obvious” (PE1) while Participant 3 suggested using STOPP/START criteria to clarify medication review questions. Additionally, they aimed to improve scenario relevance by proposing changes such as referencing “recurrent falls” in the case to justify medication reviews.

3.3. Part 2:Focus Group Interview Findings—Stakeholder Perceptions

Four major themes emerged from the focus group: (1) SCT Perceived Utility, (2) Design Features, (3) Group Dynamics, and (4) Best Practice Recommendations. Thematic analysis results are visually represented in Supplementary File S4 Figure S1.
(1)
SCT Perceived Utility
Participants viewed SCTs as feasible and practical for assessing clinical decision-making skills authentically. They highlighted perceived efficiencies in grading and the potential for online delivery and automation in high-stakes contexts. Some also felt that constructing SCT items could be less demanding than writing high-quality MCQs: “In some ways it’s easier than MCQs, because with an MCQ you have to find four distractors, and with these you do not, so I quite liked it personally.” (PE3) These reflections represent participants’ perceptions due to the workshop and insight while developing the SCT rather than outcomes measured in the present study. As one participant noted, “The marking can be really difficult…when you’re doing high stakes exams, you’re marking criteria can mean the difference between pass or fail” (PE1). Anticipated educational impact was also highlighted from participants’ perceptions, particularly for prescribing judgments, with SCTs viewed as prompting deeper engagement with reasoning: “They certainly will have an educational impact as people start to think of it [SCT] more about prescribing judgments.” (PE2).
(2)
Design Features
Participants highlighted two distinguishing design features of SCTs and compared them with their expertise with other assessment forms. First, design flexibility; unlike MCQs, SCT was perceived to simplify the construction and modification of items by eliminating the need for multiple distractors. They perceived that this allows the task to centre around nuanced clinical judgement rather than factual recall. Second, structure and format; while SCTs share some similarities with EMQs (i.e., core stems), participants noted that SCTs were distinguished by their focus on decision-making under uncertainty.
(3)
Group Dynamics
From their experience in the SCT development, participants described effective collaboration characteristics by natural role division of labour (e.g., assigning leaders, scribe and challenger). As one participant observed, “What I noticed about that session…people fell into roles.” (PE3). They also reflected on how team size shaped contribution and dynamics: “I think four is a perfect amount; with more than four you start to get people who do not contribute.” (PE2). The process was iterative, with ideas triggered from other ideas “What happened quite a lot. You say something that sparks? … I was thinking about something else. …, and then you go.” (PE3) and was supported by engaging in constructive debates to refine items. A “challenge culture” within the team, inviting peers to question assumptions and provide different input “, That’s what [multidisciplinary working] is about.” (PE2).
(4)
Best Practice Recommendations
Several best practice suggestions arose after the participants’ new experience. These suggestions included leveraging established tools, such as STOPP-START, building SCT item banks to support reuse and broader applicability, conducting practice sessions for new developers, and maintaining small team sizes (four to six members) to improve efficiency and productivity: “If you had 12 people, split down into two or three groups, then get back together.” (PE2).

4. Discussion

This feasibility study explored four medical educators’ experiences of developing SCTs for prescribing decisions in older adults for final-year medical students. Demonstrating that collaborative SCT development is feasible and acceptable within typical institutional constraints. The study provides insights into the test construction process, their perceptions of SCT utility, comparisons to other assessments, and recommendations for best practices.

4.1. Reflections on the SCT Development Process

Observations and focus groups revealed several critical aspects of the development. The efficiency and collaborative dynamics observed in the workshop suggest that SCTs can be developed in a time-effective manner when supported by a well-coordinated team. The collaborative dynamics seen during the workshop sparked a variety of ideas. Furthermore, the educators acknowledged the collaborative idea-building among themselves.
The iterative process of drafting, discussing both panel and test-taker perspectives, and refining SCT items accordingly facilitated the tailoring to common scenarios encountered in older adults. This approach is consistent with Johnson, Constantinou [25] findings, which emphasise the importance of incorporating examinee perspectives in question development to ensure clarity and relevance in test items.
Additionally, the participants engaged in the practice of creating items prior to the scenarios, a technique Johnson, Constantinou [25] describe as “reverse logic,” a process in which one starts with the intended outcome or answer and works backwards, which aids in crafting more effective test items.
Moreover, the novelty of SCTs necessitates that developers possess a thorough understanding of both the subject matter and the SCT methodology to create items that accurately evaluate clinical reasoning in ambiguous situations. Introducing practice sessions has proven essential in supporting the development of clear and well-aligned SCT components, as evidenced by think-aloud protocols and the focus interview; these sessions have significantly contributed to content development by enhancing understanding of the SCT format and the ability to design challenging judgement-based options and clinically authentic cases. The participants acknowledged this positive impact during the group interview.
The pace of SCT development during the observation period was rapid, which participants acknowledged was faster than anticipated, suggesting that the structured collaborative approaches combined with clear role distribution can enhance the efficiency in SCT construction. These findings support the feasibility of the collaborative development within typical time and resource constraints. The observed momentum gain indicated that the initial familiarisation with the SCT format, though requiring upfront time investment, accelerates subsequent item development once the participants master the assessment methodology.

4.2. Comparison of SCT with Other Assessments

Participants perceived SCTs as providing distinct advantages over MCQs and EMQs for assessing clinical decision-making.
Participants value SCTs for their focus on decision-making under uncertainty rather than factual recall [26,27]. One participant noted that SCTs “assesses something different to multiple choice question” (PE1), underscoring their perceived uniqueness in capturing nuanced aspects of decision-making, which aligns with contemporary demands for comprehensive and authentic assessment frameworks [28].
Participants acknowledged the structural similarities between SCTs and EMQs (both use a central stem or scenario to anchor questions). However, they distinguished SCTs by their emphasis on reasoning in ambiguous situations that expert clinicians recognise may reasonably differ in their approaches, whereas EMQs effectively evaluate pattern recognition and differential diagnosis skills through their thematic structure [29,30].
This position suggests that SCTs are more appropriate in cases where the educational goal is to assess clinical reasoning in ambiguous situations where multiple possible answers exist. This mirrors real-world clinical practices and aligns with recent trends that advocate for authentic assessments that reflect the complexity of clinical decision-making processes [31].

4.3. SCTs’ Utility

According to Van der Vleuten, C.P.M. (1996), the utility of an assessment tool is crucial to ensure that the assessments are not only accurate and reliable but also support learning, are resource-efficient, and acceptable to all stakeholders, thus optimising the educational impact [5].
SCTs have been demonstrated in previous studies to be a valid and reliable tool for assessing clinical reasoning in different fields [32,33]. In this study, participants have reported their perceptions of practicality and educational impact of SCTs, such as the practicality of construction, scoring, and delivery, which affirm their value in assessing clinical reasoning in prescribing.
Practicality in design and grading: SCTs have more straightforward grading criteria than other open-ended assessments, making them more suitable for high-stakes testing. The analysis revealed that participants valued SCTs for the design flexibility and simplicity in terms of the elimination of distractor constructions, which are required in MCQs. It was perceived as enhancing efficiency in item development.
Participants, based on their expertise in constructing different assessments, noted the possibilities of automated scoring and online delivery that could enhance SCT’s feasibility for large-scale assessments. While in this current study, these features were not evaluated, previous research has demonstrated the successful implementation of automated SCT scoring systems [34,35,36,37], suggesting this represents a practical direction for future development to promote broader SCT adoption.
The empirically demonstrated time efficiency of collaborative development, coupled with participants’ perceptions of SCT advantages based on their expertise in various assessment development, supports the potential institutional feasibility of SCT implementation.
Educational impact: The differentiation between SCTs and other assessments, such as MCQs and EMQs, further illustrates SCTs’ unique position in medical education. By demanding an evaluation of clinical scenarios against a backdrop of uncertainty, SCTs mimic the complexities of real-world clinical decision-making, thereby offering a more authentic assessment of clinical reasoning skills. These findings accord with Charlin, Roy [7] citing SCTs’ alignment with uncertain clinical reasoning. Unlike MCQs, SCTs offer a more nuanced assessment of decision-making, focusing on Judgment rather than recall. These findings agree with Lubarsky, Dory [13], who cited that SCTs are more of a knowledge application manifested in a form of decision-making. SCTs were seen as having potential educational impacts, particularly in influencing how prescribers’ approach clinical judgments. This suggests SCTs could play a valuable role in training. Moreover, SCT has proven to exert an educational impact on other fields. It is a reliable and standardised test of genetic counselling trainees’ clinical reasoning skills [38]. In a paediatric clerkship, SCTs were used for formative feedback and performance evaluation, indicating a correlation between SCT scores and other measures [39]. Anaesthesiologists employ SCTs to evaluate residents’ clinical reasoning for airway management, showing they can distinguish between clinical experience levels [40].

4.4. Development Guidelines and Recommendations

Recommendations were provided to optimise SCT development, including using existing resources as templates, building banks of items, having practice sessions, and managing team sizes. These suggestions could make development more efficient and also offer practical guidance for programmes that wish to move from the single four-person feasibility team towards broader implementation.

4.4.1. Using Existing Resources as Templates

Participants recommended leveraging established clinical tools and guidelines as templates to enhance scenario authenticity and clinical relevance, such as STOPP/START criteria (Screening Tool of Older Persons’ Prescriptions/Screening Tool to Alert to Right Treatment) to clarify medication review questions. STOPP/START represents evidence-based criteria for identifying potentially inappropriate prescribing patterns in older adults [41,42], which could inform future SCT development by providing structured frameworks for medication review scenarios. Such existing clinical resources serve as valuable templates for ensuring that SCT scenarios reflect real-world clinical challenges and support the development of clinically authentic assessment content.

4.4.2. Building SCT Bank

Participants recommended developing an SCT bank, as one of them felt that some questions constructed during the workshop could be adapted for use with populations beyond older adults. Establishing such a bank would be a practical step, particularly to support diverse contexts. We consider this a valuable recommendation, provided that the SCTs undergo a rigorous quality assurance process and are validated. This aligns with Aldekhayel, Alselaim [43], who highlighted the critical importance of ensuring the validity of SCT components to establish them as readily available resources for clinical reasoning assessments.

4.4.3. Practice Sessions

Training sections are significant for novice educators. In this study, the training session was conducted as a workshop, providing participants with the opportunity to address misconceptions regarding the SCT construction process. The session was found to be successful, with participants acknowledging its benefits. Similarly, a study utilised a workshop to train nursing educators in SCT design further emphasised the importance of such sessions in developing the necessary skills and understanding [44].

4.4.4. Optimising Team Sizes for SCT Development

The role of team size has been a topic of debate. Fournier, Demeester [12] advocate for smaller teams, suggesting that single-person efforts or two can boost efficiency and inspire creativity, a view also supported by Lubarsky, Dory [13]. However, findings from this study indicate that teams with more than two members can further accelerate the development process and enhance creativity. Focus group reflections and observations revealed that interactions within larger groups effectively balanced collaboration and productivity.
During the focus group, participants speculated that teams of four or up to six would foster productive collaboration. At the same time, they believe that teams larger than six risked reduced efficiency due to non-productive discussions, a concern also witnessed by Fournier, Demeester [12].
However, these recommendations about team size of more than four should be interpreted with caution, as they represent participants’ predictions rather than empirical observation. The data from this study support that a team size of four, with pre-existing professional relationships, worked well for this specific task. Their suggestion of a team of six members might be optimal warrants testing for future studies.
The participants recommended dividing larger groups into subgroups of four to six members, which could later reconvene to collaboratively refine SCT items. These reflections derive from one four-person team and warrant exploration in other settings.

4.4.5. Institutional Implementation Infrastructure

Translating these recommendations into routine practice requires basic institutional infrastructure. First, programmes need access to exemplar SCTs, clearly formatted templates for vignettes and items, and a shared repository for storing, versioning, and curating SCT vignettes, scoring keys, and supporting evidence (for example, an institutional SCT bank). Secondly, local leadership and facilitation capacity are needed to plan and deliver workshops, support development teams, and oversee quality assurance of newly created items. This includes ongoing faculty development to maintain expertise and support new team members as programmes expand. Thirdly, cross-institutional collaboration could enable partner institutions to co-develop and share SCT banks while still undertaking local review to ensure contextual and curricular fit. Together, these elements provide a pragmatic framework for programmes that wish to adopt and scale SCT development beyond the single four-person feasibility team described in this study.

4.5. The Role of Group Dynamics in SCT Development

The success of team collaboration plays a pivotal role in the effectiveness of SCT development. The team’s progression through Tuckman’s stages was evident, though the influence of the team’s pre-existing professional relationships and shared expertise [21].
Pre-existing professional relationships appeared to accelerate the team development. The compressed early forming–storming phases may be explained by reduced initial uncertainty and established trust, enabling participants to focus on task completion rather than relationship building [45,46]. This familiarity allowed for the natural adoption of roles without prolonged negotiation, facilitating a direct focus on the task. Evidence from healthcare teams indicates that increased interprofessional familiarity can improve collaborative performance [45]. Repeated experiences working together predict faster learning [46]. The observed rapid progression through stages may not be representative of newly formed teams without established working relationships. Teams composed of unfamiliar members would likely require extended forming and storming phase, potentially affecting development efficiency. Future development initiatives should consider whether to prioritise existing professional relationships for time efficiency or deliberately construct diverse unfamiliar teams to test broader applicability.
The study observation shows that most activity occurred in the performing stage, which demonstrates that well-coordinated teams with a shared professional background can maximise productive collaboration. However, newly formed teams without pre-existing relationships may require a longer forming and storming period [47] but could achieve effective collaboration through structured facilitation.
Several limitations should be acknowledged. First, while the study observed high SCT production rates, the quality of the SCT items has not yet been comprehensively evaluated. Second, all participants were from the same institution and had pre-existing relationships, which may have influenced group dynamics. This familiarity likely facilitated a shorter Forming and Storming stage in Tuckman’s model, potentially differing from dynamics observed in newly formed teams. Third, the final sample size was determined by practical constraints rather than planned design, as two initially invited participants withdrew due to scheduling conflicts. While the resulting team size of four proved effective for intensive process observation, this represents a feasibility constraint rather than an optimal, simple size determination. Fourth, this feasibility study used a small purposive sample (n = 4) to intensively observe in real-time in a single collaboration workshop as recommended [14], rather than for broader generalisation across multiple settings. While this approach facilitated the detailed capture of the process, it limits transferability to other institutional contexts. Fifth, methodological limitations include potential observer bias from the researcher’s presence during activities and dual facilitator-observer role, which may have influenced participant behaviour. These may affect interpretation of observed behaviours and reported perceptions. Future research should investigate SCT development with more diverse teams and incorporate robust quality assurance measures to evaluate the clarity and effectiveness of SCTs produced after team collaboration, ensuring their overall success.
This study provides evidence of the process feasibility of collaborative SCT development and illuminates participants’ perceptions of utility and acceptability. Our findings represent observed collaborative behaviours and participant-reported experiences, rather than validated performance outcomes or empirically tested effectiveness measures. The feasibility evidence reflects successful task completion within resource constraints and participants’ experience; however, long-term sustainability and effectiveness across diverse institutional contexts require further investigation.
In conclusion, the present study explored how one team, consisting of four medical educators with previous experience in developing various assessments, developed SCTs for prescribing assessments to be utilised in an RCT. This was executed through observation and focus group methods. Main findings suggest that SCTs can be efficiently created when teams have a structured process, allocated roles, and have established professional relationships. Regarding the perceived findings from the participants, they perceived SCTs as practical for the assessment of clinical reasoning under uncertainty and noted the SCTs’ perceived design advantage over MCQs.
However, this study has several limitations: the quality of SCT items was not evaluated, all participants were from a single institution with pre-existing relationships among them, and the perceptions of educational impact represent anticipated rather than measured outcomes.
Best practices recommendations that emerged from the single-team experience, such as conducting practice sessions, team size, leveraging existing clinical tools as templates and developing validated item banks. However, these recommendation needs empirical testing across different contexts and settings. Practical implication: For institutional consideration, this study suggests that collaborative approaches supported by appropriate facilitation and basic infrastructure showed potential for efficient development of SCT items in our study context. Future research should examine the quality of collaborative SCT development and compare outcomes across different team configurations, and compare SCTs produced by collaborative teams versus individual item writers as well as empirically test optimal development processes in various settings.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ime5010008/s1, Supplementary File S1. SCT Structure Example; Supplementary File S2. Participant Characteristics Table S1; Supplementary File S3. Workshop Structure Table S2; Supplementary File S4. Thematic Map Figure S1.

Author Contributions

Conceptualization, R.M.A.; methodology, R.M.A.; validation, R.M.A. and S.K.P.; formal analysis, R.M.A.; investigation, R.M.A.; resources, R.M.A.; data curation, R.M.A.; writing—original draft preparation, R.M.A.; writing—review and editing, R.M.A., S.K.P. and A.R.C.; visualization, R.M.A.; supervision, S.K.P. and A.R.C.; project administration, S.K.P.; funding acquisition, R.M.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of Higher Education in Saudi Arabia and Princess Nourah bint Abdulrahman University.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and was approved by the University of Birmingham Ethics Committee [ERN_21-1350]. Informed consent was obtained from all participants involved in the study. All participants provided written consent for participation, audio recording and use of anonymised data.

Data Availability Statement

The datasets used during the current study are available from the corresponding author on reasonable request, subject to privacy and ethical restrictions.

Acknowledgments

The authors gratefully acknowledge the medical educators who participated in this study for sharing their time, insights, and experiences.

Conflicts of Interest

The authors report there are no competing interests to declare.

References

  1. Berger, S.; Krug, K.; Goetz, K. Encountering uncertainty and complexity in decision-making: An observational study of clinical reasoning among medical and interprofessional groups of health care students. J. Interprofessional Care 2023, 37, 262–271. [Google Scholar] [CrossRef] [PubMed]
  2. Maxwell, S.R. Rational prescribing: The principles of drug selection. Clin. Med. 2016, 16, 459–464. [Google Scholar] [CrossRef] [PubMed]
  3. Aronson, J.K. A prescription for better prescribing. Br. J. Clin. Pharmacol. 2006, 61, 487–491. [Google Scholar] [CrossRef]
  4. Kamarudin, G.; Penm, J.; Chaar, B.; Moles, R. Educational interventions to improve prescribing competency: A systematic review. BMJ Open 2013, 3, e003291. [Google Scholar] [CrossRef]
  5. Van Der Vleuten, C.P.M. The assessment of professional competence: Developments, research and practical implications. Adv. Health Sci. Educ. Theory Pract. 1996, 1, 41–67. [Google Scholar] [CrossRef]
  6. Shi, W.; Qin, H.; Vaughan, B.; Ng, L. Educational Interventions for Medical Students to Improve Pharmacological Knowledge and Prescribing Skills: A Scoping Review. Perspect. Med. Educ. 2023, 12, 348–360. [Google Scholar] [CrossRef]
  7. Charlin, B.; Roy, L.; Brailovsky, C.; Goulet, F.; van der Vleuten, C. The Script Concordance test: A tool to assess the reflective clinician. Teach. Learn. Med. 2000, 12, 189–195. [Google Scholar] [CrossRef] [PubMed]
  8. Charlin, B.; Brailovsky, C.; Leduc, C.; Blouin, D. The Diagnosis Script Questionnaire: A New Tool to Assess a Specific Dimension of Clinical Competence. Adv. Health Sci. Educ. Theory Pract. 1998, 3, 51–58. [Google Scholar] [CrossRef]
  9. Humbert, A.J.; Johnson, M.T.; Miech, E.; Friedberg, F.; Grackin, J.A.; Seidman, P.A. Assessment of clinical reasoning: A Script Concordance test designed for pre-clinical medical students. Med. Teach. 2011, 33, 472–477. [Google Scholar] [CrossRef] [PubMed]
  10. Dory, V.; Gagnon, R.; Vanpee, D.; Charlin, B. How to construct and implement script concordance tests: Insights from a systematic review. Med. Educ. 2012, 46, 552–563. [Google Scholar] [CrossRef] [PubMed]
  11. Iglesias Gómez, C.; González Sequeros, O.; Salmerón Martínez, D. Clinical reasoning evaluation using script concordance test in primary care residents. Pediatría (Engl. Ed.) 2022, 97, 87–94. [Google Scholar] [CrossRef]
  12. Fournier, J.P.; Demeester, A.; Charlin, B. Script concordance tests: Guidelines for construction. BMC Med. Inform. Decis. Mak. 2008, 8, 18. [Google Scholar] [CrossRef] [PubMed]
  13. Lubarsky, S.; Dory, V.; Duggan, P.; Gagnon, R.; Charlin, B. Script concordance testing: From theory to practice: AMEE guide no. 75. Med. Teach. 2013, 35, 184–193. [Google Scholar] [CrossRef] [PubMed]
  14. Bowen, D.J.; Kreuter, M.; Spring, B.; Cofta-Woerpel, L.; Linnan, L.; Weiner, D.; Ballen, S.; Patrick, C.; Squiers, L.; Fabrizio, C.; et al. How We Design Feasibility Studies. Am. J. Prev. Med. 2009, 36, 452–457. [Google Scholar] [CrossRef]
  15. Tong, A.; Sainsbury, P.; Craig, J. Consolidated criteria for reporting qualitative research (COREQ): A 32-item checklist for interviews and focus groups. Int. J. Qual. Health Care 2007, 19, 349–357. [Google Scholar] [CrossRef] [PubMed]
  16. Palinkas, L.A.; Horwitz, S.M.; Green, C.A.; Wisdom, J.P.; Duan, N.; Hoagwood, K. Purposeful Sampling for Qualitative Data Collection and Analysis in Mixed Method Implementation Research. Adm. Policy Ment. Health 2015, 42, 533–544. [Google Scholar] [CrossRef]
  17. Vasileiou, K.; Barnett, J.; Thorpe, S.; Young, T. Characterising and justifying sample size sufficiency in interview-based studies: Systematic analysis of qualitative health research over a 15-year period. BMC Med. Res. Methodol. 2018, 18, 148. [Google Scholar] [CrossRef]
  18. Ericsson, K.A. Capturing Expert Thought with Protocol Analysis: Concurrent Verbalizations of Thinking during Experts’ Performance on Representative Tasks. In The Cambridge Handbook of Expertise and Expert Performance, 2nd ed.; Williams, A.M., Kozbelt, A., Ericsson, K.A., Hoffman, R.R., Eds.; Cambridge University Press: Cambridge, UK, 2018; pp. 192–212. [Google Scholar]
  19. Braun, V.; Clarke, V. Using thematic analysis in psychology. Qual. Res. Psychol. 2006, 3, 77–101. [Google Scholar] [CrossRef]
  20. Coates, W.C.; Jordan, J.; Clarke, S.O. A practical guide for conducting qualitative research in medical education: Part 2—Coding and thematic analysis. AEM Educ. Train. 2021, 5, e10645. [Google Scholar] [CrossRef]
  21. Tuckman, B.W.; Jensen, M.A.C. Stages of Small-Group Development Revisited1. Group. Facil. 2010, Volume 10, 43–48. [Google Scholar]
  22. Maguire, M.; Delahunt, B. Doing a Thematic Analysis: A Practical, Step-by-Step Guide for Learning and Teaching Scholars. All. Irel. J. Teach. Learn. High. Educ. 2017, 9, 3351. [Google Scholar]
  23. Jordan, J.; Clarke, S.O.; Coates, W.C. A practical guide for conducting qualitative research in medical education: Part 1—How to interview. AEM Educ. Train. 2021, 5, e10646. [Google Scholar] [CrossRef]
  24. McGrath, C.; Palmgren, P.J.; Liljedahl, M. Twelve tips for conducting qualitative research interviews. Med. Teach. 2019, 41, 1002–1006. [Google Scholar] [CrossRef] [PubMed]
  25. Johnson, M.; Constantinou, F.; Crisp, V. How do question writers compose external examination questions? Question writing as a socio-cognitive process. Br. Educ. Res. J. 2017, 43, 700–719. [Google Scholar] [CrossRef]
  26. Al-Rukban, M. Guidelines for the Construction of Multiple Choice Questions Tests. J Family Community Med 2006, 13, 125. [Google Scholar] [CrossRef]
  27. Murphy, J.F.A. Assessment in medical education. Ir. Med. J. 2007, 100, 387–396. [Google Scholar]
  28. Ajjawi, R.; Tai, J.; Dollinger, M.; Dawson, P.; Boud, D.; Bearman, M. From authentic assessment to authenticity in assessment: Broadening perspectives. Assess. Eval. High. Educ. 2024, 49, 499–510. [Google Scholar] [CrossRef]
  29. Case, S.M.; Swanson, D.B. Extended-matching items: A practical alternative to free-response questions. Teach. Learn. Med. Int. J. 1993, 5, 107–115. [Google Scholar] [CrossRef]
  30. Frey, A.; Leutritz, T.; Backhaus, J.; Hörnlein, A.; König, S. Item format statistics and readability of extended matching questions as an effective tool to assess medical students. Sci. Rep. 2022, 12, 20982. [Google Scholar] [CrossRef]
  31. Naylor, K.; Hislop, J.; Torres, K.; Mani, Z.A.; Goniewicz, K. The Impact of Script Concordance Testing on Clinical Decision-Making in Paramedic Education. Healthcare 2024, 12, 282. [Google Scholar] [CrossRef] [PubMed]
  32. Humbert, A.J.; Besinger, B.; Miech, E.J. Assessing clinical reasoning skills in scenarios of uncertainty: Convergent validity for a Script Concordance Test in an emergency medicine clerkship and residency. Acad. Emerg. Med. 2011, 18, 627–634. [Google Scholar] [CrossRef]
  33. Hamui, M.; Ferreira, J.P.; Torrents, M.; Torres, F.; Ibarra, M.E.; Ossorio, M.F.; Urrutia, L.; Ferrero, F. Script Concordance Test: First nationwide experience in pediatrics. Arch. Argent. Pediatr. 2018, 116, e151–5. [Google Scholar]
  34. Sibert, L.; Darmoni, S.J.; Dahamna, B.; Weber, J.; Charlin, B. Online clinical reasoning assessment with the Script Concordance test: A feasibility study. BMC Med. Inform. Decis. Mak. 2005, 5, 18. [Google Scholar] [CrossRef] [PubMed]
  35. Kania, R.E.; Verillaud, B.; Tran, H.; Gagnon, R.; Kazitani, D.; Huy, P.T.; Herman, P.; Charlin, B. Online script concordance test for clinical reasoning assessment in otorhinolaryngology: The association between performance and clinical experience. Arch. Otolaryngol. Head. Neck Surg. 2011, 137, 751–755. [Google Scholar] [CrossRef]
  36. Hornos, E.H.; Pleguezuelos, E.M.; Brailovsky, C.A.; Harillo, L.D.; Dory, V.; Charlin, B. The practicum script concordance test: An online continuing professional development format to foster reflection on clinical practice. J. Contin. Educ. Health Prof. 2013, 33, 59–66. [Google Scholar] [CrossRef] [PubMed]
  37. Naylor, K.; Chrzanowska-Wąsik, M.; Okońska, P.; Kucmin, T.; Al-Wathinani, A.M.; Goniewicz, K. Adapting to a Pandemic: Web-Based Residency Training and Script Concordance Testing in Emergency Medicine During COVID-19. Disaster Med. Public Health Prep. 2023, 17, e541. [Google Scholar] [CrossRef]
  38. Begun, Y.S.; Stange, L.R.; Davis, C.R. Script concordance testing in genetic counseling training: A pilot study. J. Genet. Couns. 2023, 32, 1121–1130. [Google Scholar] [CrossRef]
  39. Bhardwaj, P.; Black, E.W.; Fantone, J.C., 3rd; Lopez, M.; Kelly, M. Script Concordance Tests for Formative Clinical Reasoning and Problem-Solving Assessment in General Pediatrics. MedEdPORTAL 2022, 18, 11274. [Google Scholar] [CrossRef]
  40. Omega, A.; Wijaya Ramlan, A.A.; Soenarto, R.F.; Heriwardito, A.; Sugiarto, A. Assessing clinical reasoning in airway related cases among anesthesiology fellow residents using Script Concordance Test (SCT). Med. Educ. Online 2022, 27, 2135421. [Google Scholar] [CrossRef]
  41. O’Mahony, D.; Cherubini, A.; Guiteras, A.R.; Denkinger, M.; Beuscart, J.B.; Onder, G.; Gudmundsson, A.; Cruz-Jentoft, A.J.; Knol, W.; Bahat, G.; et al. STOPP/START criteria for potentially inappropriate prescribing in older people: Version 3. Eur. Geriatr. Med. 2023, 14, 625–632. [Google Scholar] [CrossRef]
  42. O’Mahony, D.; O’Sullivan, D.; Byrne, S.; O’Connor, M.N.; Ryan, C.; Gallagher, P. STOPP/START criteria for potentially inappropriate prescribing in older people: Version 2. Age Ageing 2015, 44, 213–218. [Google Scholar] [CrossRef]
  43. Aldekhayel, S.A.; Alselaim, N.A.; Magzoub, M.E.; Al-Qattan, M.M.; Al-Namlah, A.M.; Tamim, H.; Al-Khayal, A.; I Al-Habdan, S.; Zamakhshary, M.F. Constructing a question bank based on script concordance approach as a novel assessment methodology in surgical education. BMC Med. Educ. 2012, 12, 100. [Google Scholar] [CrossRef]
  44. Habes, E.V.; Kolk, J.E.M.; Van Brunschot, M.F.M.; Bouwes, A. Development of script concordance test for assessment of clinical reasoning in nursing: Lessons learned regarding construct validity. Heliyon 2024, 10, e35151. [Google Scholar] [CrossRef] [PubMed]
  45. Iyasere, C.A.; Wing, J.; Martel, J.N.; Healy, M.G.; Park, Y.S.; Finn, K.M. Effect of Increased Interprofessional Familiarity on Team Performance, Communication, and Psychological Safety on Inpatient Medical Teams: A Randomized Clinical Trial. JAMA Intern. Med. 2022, 182, 1190–1198. [Google Scholar] [CrossRef] [PubMed]
  46. Reagans, R.; Argote, L.; Brooks, D. Individual experience and experience working together: Predicting learning rates from knowing who knows what and knowing how to work together. Manag. Sci. 2005, 51, 869–881. [Google Scholar] [CrossRef]
  47. Harrison, D.A.; Mohammed, S.; Mcgrath, J.E.; Florey, A.T.; Vanderstoep, S.W. Time Matters in Team Performance: Effects of Member Familiarity, Entrainment, and Task Discontinuity on Speed and Quality. Pers. Psychol. 2003, 56, 633–669. [Google Scholar] [CrossRef]
Figure 1. Group progression through Tuckman’s stages during SCT development.
Figure 1. Group progression through Tuckman’s stages during SCT development.
Ime 05 00008 g001
Table 1. Tuckman Model Stages and Characteristics.
Table 1. Tuckman Model Stages and Characteristics.
StageCharacteristics
FormingInitial stage where group members get acquainted and establish ground rules.
StormingMembers experience conflicts and competition as they assert their opinions.
NormingGroup begins to establish norms and cohesion, resolving conflicts.
PerformingGroup reaches optimal functioning, focusing and working effectively towards goals.
Table 2. Timing and outputs for each SCT development during the workshop.
Table 2. Timing and outputs for each SCT development during the workshop.
SCT ComponentNo. of ItemsTime (min)
Q1521
Q2528
Q3316
Q4519
Q5520
Q6412
Q7311
Average418
Note: Average is rounded to the nearest whole number.
Table 3. Observed Feasibility Indicators.
Table 3. Observed Feasibility Indicators.
Feasibility IndicatorsEvidence from the StudyAssessment
Time Efficiency 127 min for 7 vignettes (18 min/vignette average) Efficient development rate
Resource Requirements 4 participants, 3 h workshop, basic materials Minimal institutional resource needs
Medical Educators Acceptability Positive focus group responses, sustained engagement. Participants acceptance overserved and reported
Workload Manageability Collaborative approach reduced individual burden, natural role division. Workload for participants appeared manageable.
Table 4. Summary of observed group behaviours mapped to Tuckman’s stages.
Table 4. Summary of observed group behaviours mapped to Tuckman’s stages.
StageGroup ActionDynamics
FormingRapid establishment of roles and task orientation.- Focused immediately on task with minimal introductions.
- Quick agreement on roles and responsibilities.
- Participant 2 emerges as an informal leader.
StormingMinor conflicts and collaborative resolution.- Generally agreeable interactions with occasional disagreements.
- Quick resolution of disagreements through open discussion.
- Building on each other’s ideas.
NormingDeveloped a shared understanding of the task, how to approach it and a collaborative rhythm.- Mutual reinforcement of ideas and approaches.
- Brainstormed and refined SCT questions collaboratively.
PerformingDemonstrated high productivity and adaptability in SCT case development.- Adapted strategies to challenges in real-time.
- Deep engagement with complex clinical scenarios.
- Critical evaluation of their own work throughout the process
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alhossaini, R.M.; Cox, A.R.; Pontefract, S.K. Feasibility of Script Concordance Test Development: A Qualitative Study of Medical Educators’ Experiences. Int. Med. Educ. 2026, 5, 8. https://doi.org/10.3390/ime5010008

AMA Style

Alhossaini RM, Cox AR, Pontefract SK. Feasibility of Script Concordance Test Development: A Qualitative Study of Medical Educators’ Experiences. International Medical Education. 2026; 5(1):8. https://doi.org/10.3390/ime5010008

Chicago/Turabian Style

Alhossaini, Reem M., Anthony Richard Cox, and Sarah Katie Pontefract. 2026. "Feasibility of Script Concordance Test Development: A Qualitative Study of Medical Educators’ Experiences" International Medical Education 5, no. 1: 8. https://doi.org/10.3390/ime5010008

APA Style

Alhossaini, R. M., Cox, A. R., & Pontefract, S. K. (2026). Feasibility of Script Concordance Test Development: A Qualitative Study of Medical Educators’ Experiences. International Medical Education, 5(1), 8. https://doi.org/10.3390/ime5010008

Article Metrics

Back to TopTop