Next Article in Journal
Establishing a Centralized Virtual Visit Support Team: Early Insights
Next Article in Special Issue
Perceptions of Patient Safety Culture among Triage Nurses in the Emergency Department: A Cross-Sectional Study
Previous Article in Journal
Ultrasound Imaging of Subtalar Joint Instability for Chronic Ankle Instability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feedback for Emergency Ambulance Staff: A National Review of Current Practice Informed by Realist Evaluation Methodology

1
School of Psychology, University of Leeds, Leeds LS2 9JT, UK
2
Research and Development Department, Yorkshire Ambulance Service NHS Trust, Wakefield WF2 0XQ, UK
3
Yorkshire Quality and Safety Research Group, Bradford Institute for Health Research, Bradford BD9 6RJ, UK
4
Faculty of Health and Education, Manchester Metropolitan University, Manchester M15 6BH, UK
*
Author to whom correspondence should be addressed.
Current address: Faculty of Health, Medicine and Social Care, Anglia Ruskin University, Chelmsford CM1 1SQ, UK.
Healthcare 2023, 11(16), 2229; https://doi.org/10.3390/healthcare11162229
Submission received: 30 June 2023 / Revised: 27 July 2023 / Accepted: 7 August 2023 / Published: 8 August 2023
(This article belongs to the Special Issue Quality of Pre-hospital Care)

Abstract

:
Research suggests that feedback in Emergency Medical Services (EMS) positively affects quality of care and professional development. However, the mechanisms by which feedback achieves its effects still need to be better understood across healthcare settings. This study aimed to understand how United Kingdom (UK) ambulance services provide feedback for EMS professionals and develop a programme theory of how feedback works within EMS, using a mixed-methods, realist evaluation framework. A national cross-sectional survey was conducted to identify feedback initiatives in UK ambulance services, followed by four in-depth case studies involving qualitative interviews and documentary analysis. We used qualitative content analysis and descriptive statistics to analyse survey responses from 40 prehospital feedback initiatives, alongside retroductive analysis of 17 interviews and six documents from case study sites. Feedback initiatives mainly provided individual patient outcome feedback through “pull” initiatives triggered by staff requests. Challenges related to information governance were identified. Our programme theory of feedback to EMS professionals encompassed context (healthcare professional and organisational characteristics), mechanisms (feedback and implementation characteristics, psychological reasoning) and outcomes (implementation, staff and service outcomes). This study suggests that most UK ambulance services use a range of feedback initiatives and provides 24 empirically based testable hypotheses for future research.

1. Introduction

Feedback on clinical performance is well researched, and systematic reviews suggest it results in small to moderate improvements in patient care by enhancing healthcare professionals’ compliance with desired clinical practice across healthcare settings, including Emergency Medical Services (EMS) [1,2]. EMS professionals work autonomously in an environment characterised by complexity, uncertainty and extreme stressors [3,4]. They make complex decisions, including assessing and treating patients at home to avoid unnecessary hospital attendance and reduce demand [5,6]. For EMS staff, receiving feedback on clinical practice and decision-making is vital, yet often difficult due to practical constraints in obtaining information concerning patients not conveyed to hospital and barriers to information transfer across the boundary marked by hospital admission.
Research from North America and the United Kingdom (UK) suggests that EMS staff view current feedback provision as inadequate and desire more feedback, especially concerning patient outcomes [7,8,9]. Patient outcome feedback involves providing clinicians with information regarding what happened to their patients to support reflection and self-evaluation of clinical performance [10]. Meanwhile, clinical performance feedback predominantly involves auditing clinical performance over time and feeding this back to clinicians in summary form (‘audit and feedback’) [11].
A recent meta-analysis supported a moderate positive effect for various feedback types within EMS, including clinical performance and patient outcome feedback [2]. The review highlighted a need for more theoretical underpinning and robust evaluation of existing feedback in EMS regarding outcomes and understanding the mechanisms by which feedback operates within this context [2].
The broader literature on clinical performance feedback suggests that while feedback overall has small to moderate positive effects, these effects vary greatly depending on the context, initial performance and mode of feedback provision [1]. Despite researchers attempting to optimise the effects of feedback across healthcare settings [12,13,14], key active components and mechanisms by which feedback works still need to be better understood [15]. The most recent theory development within audit and feedback in healthcare is Clinical Performance Feedback Intervention Theory (CP-FIT), based on a meta-synthesis of 73 feedback interventions [16]. Although none of the 73 interventions were within EMS, our provisional logic model developed from qualitative work in one ambulance trust found CP-FIT to have good face validity when exploring how feedback impacts patient care, patient safety and staff wellbeing [9].
Examples of clinical performance and patient outcome feedback within EMS in the UK include paramedics receiving feedback on cardiac arrests [17] and accessing post-box schemes at emergency departments to directly follow up on patients [18] or request facilitator-mediated feedback [19]. Considerable variation exists between these initiatives, which appear to be primarily isolated schemes developed to meet local priorities and needs without systematic evaluation [2]. This may prevent collective learning and increases the risk of individual feedback initiatives failing [20].
This study aimed to understand how UK ambulance services provide feedback to their staff and develop a programme theory of how feedback works within EMS. Programme theories make sense of complex interventions by drawing out the underlying principles of how an intervention causes outcomes within a specific context [21,22,23]. The following questions were addressed:
  • What are the types and design elements of interventions providing feedback for EMS professionals in the UK?
  • What is it about EMS feedback interventions that works, for whom, in what circumstances, how and to what extent?

2. Materials and Methods

2.1. Study Design

This was a mixed-methods study informed by realist evaluation methodology. Realist approaches seek to explain underlying mechanisms of action and encourage using substantive theory from other disciplines, for example, where research into feedback is more advanced [24,25]. The study consisted of two work packages (WPs): a cross-sectional survey of EMS feedback initiatives in the UK (WP1) followed by purposively sampled case studies involving documentary analysis and semi-structured interviews (WP2) (Figure 1).
The study was carried out in accordance with the UK Policy Framework for Health and Social Care Research [26]. It was approved by the University of Leeds ethics committee (PSYC-242 7 July 2021) and the Health Research Authority (IRAS project ID 287110). The RAMESES II reporting standards for realist evaluations [27] and Checklist for Reporting Of Survey Studies [28] guided study reporting.

2.2. Setting and Participants

The study setting was UK emergency ambulance services. Eligible participants for WP1 were leads of EMS feedback initiatives, defined as “providing feedback to emergency ambulance staff regarding their performance within prehospital practice and/or patient outcomes”. We used a sampling framework to purposively select four case studies from the feedback initiatives identified in WP1. The sampling framework was stratified according to contextual and feedback intervention characteristics, including whether feedback was actively sought out by EMS staff (‘pull feedback’) or provided without solicitation (‘push feedback’). Eligible participants for WP2 were initiative leads and emergency ambulance staff who had received feedback through a case study initiative.

2.3. Data Collection

The survey and interview guides (Supplementary Files S1–S3) were developed by an early career paramedic researcher (C.W.) with input from a senior health services researcher with experience in evaluating feedback initiatives (J.B.). The survey for WP1 was hosted on Qualtrics (Qualtrics, Provo, UT) from September 2021 to January 2022 and contained six sections: respondent demographics, feedback initiative description, mechanisms, implementation, consequences and unique characteristics. The questions were open- and closed-format and included qualitative (e.g., project aims, barriers) and quantitative (e.g., start date, numbers of requests) items.
No formal sample size calculation was conducted as this was an exploratory survey aiming to capture all eligible EMS feedback initiatives. We used purposive and snowball sampling to send the survey to programme leads identified through grey literature, networking and clinical practice. The survey was also advertised on social media and via the research departments of participating UK ambulance services. Informed consent was obtained at the beginning of the survey, which automatically ended if users opted not to consent. The anonymity of individual participants was maintained.
To begin WP2, initiative leads were invited to participate in a semi-structured interview and provide existing documents linked to their initiative using the contact details provided in WP1. A further 2–4 interviewees within each case study were sampled using snowball sampling. C.W. conducted the interviews online from November 2021 to August 2022 and transcribed them verbatim. Prior to each interview, C.W. obtained written informed consent in electronic format. In line with realist evaluation methodology, interviews and documents were used to inspire, validate, falsify and modify hypotheses about how the intervention works [29]. During interviews, this was achieved by C.W. describing an aspect of the initial programme theory and allowing participants to explain and clarify based on their perspectives [30].

2.4. Data Analysis

Data analysis was conducted by an early career paramedic-researcher (C.W.) with input from the wider research team of senior health services researchers and behavioural scientists (G.J., R.L., J.B.). Qualitative survey items were analysed using content analysis supported by NVivo (Version 12 Plus, QSR International, Burlington, MA, USA). Quantitative survey items were analysed using descriptive statistics. Descriptive statistics were presented as frequencies (percentages) and median (interquartile ranges (IQR)). Free-text responses were categorised by C.W., with 25% (n = 10) validated by J.B.

2.5. Development of Programme Theory

The initial programme theory was developed by triangulating a logic model developed by the research team using abductive analysis in a prior interview study [9] with the results from WP1 and mapping this onto a Context–Mechanism–Outcome configuration (CMOc) in line with realist methodology [24]. These configurations set out the causal links between the specific mechanisms that trigger intervention outcomes within particular contexts [31,32]. Context here pertains to the background and implementation setting, while mechanisms describe how resources bring about change and outcomes relate to intended and unintended consequences [32,33,34]. We followed the CMOc approach by Dalkin et al. [34], dividing ‘mechanisms’ into ‘reasoning’ and ‘resources’ to capture how an intervention brings about change:
  • Context: What are the barriers and facilitators to implementing feedback in EMS?
  • Mechanism—Resources: What are the main characteristics of the type of feedback being provided?
  • Mechanism—Reasoning: What are the psychological responses by which feedback achieves its effects?
  • Outcome: What are the EMS feedback interventions’ perceived effects and impact?
The programme theory was refined by analysing the interviews and documents in WP2 using a retroductive approach that uses inductive and deductive reasoning—including researcher insights—to theorise hidden mechanisms [35,36,37]. In addition, we drew upon a systematic review of feedback in EMS [2], as well as existing middle-range theories from implementation science (CP-FIT [16], implementation outcomes taxonomy [38]) and behaviour change theory (mechanisms of action [39]).

3. Results

The survey yielded 46 responses, of which we excluded duplicate responses (n = 2) and incomplete questionnaires (n = 4), where participants entered their details but did not answer feedback intervention questions. Therefore, 40 unique EMS feedback initiatives were included in the analysis.

3.1. Characteristics of Survey Participants

Participants completing the survey were mainly based in emergency departments (14 consultants, 2 advanced clinical practitioners) or ambulance services (n = 21), often with managerial or specialist clinical responsibilities. The remaining participants were specialist healthcare professionals based in hospitals (n = 2) or regional networks (n = 1). Responses were from England (n = 27), Scotland (n = 8), Wales (n = 4) and Northern Ireland (n = 1), and from within 11 of the 14 UK ambulance services.

3.2. Types and Design Elements of Interventions Providing Feedback for EMS Professionals

Table 1 summarises the characteristics of the 40 EMS feedback initiatives. Feedback initiatives most frequently provided patient outcome feedback (n = 31, 77.5%) and less frequently clinical performance feedback (n = 4, 10.0%) or a combination of the two (n = 5, 12.5%). Patient outcome feedback included information on either the hospital diagnosis, thereby allowing paramedics to confirm their prehospital diagnosis (n = 36, 90%); patients’ hospital treatment, investigations and care trajectory to inform paramedics’ future clinical reasoning (n = 25, 62.5%); or answers to specific queries (n = 9, 22.5%). Clinical performance feedback included information on paramedics’ decision-making, such as appropriateness of conveyance and destination (n = 7, 17.5%) or compliance with EMS protocols and key performance indicators (n = 5, 12.5%), e.g., on-scene times and adherence to checklists.
Of the initiatives, 70.0% (n = 28) were active when the survey was completed, with the remaining 12 initiatives (30.0%) either permanently terminated or paused temporarily due to the COVID-19 pandemic or other service pressures. Thirty-eight respondents provided a start date for their initiative, with 77.5% (n = 31) set up in the last five years. The earliest start date was 2004; however, this initiative was no longer active. Of the currently active initiatives, the earliest was set up in 2011.
Initiatives had most commonly delivered 10–49 (n = 14, 35.0%) or 50–99 instances of feedback (n = 11, 27.5%) and predominantly provided feedback to frontline EMS staff (n = 36, 90.0%) at an individual level (n = 32, 80.0%). Email or an electronic document was the most frequently used format (n = 35, 87.5%), with provision usually occurring on an ad hoc basis (n = 29, 72.5%). The median lag time was 21 days (IQR 7-30). Feedback was most commonly for individual patient cases (n = 38, 95.0%) using a ‘pull’ model (n = 25, 62.5%), which involved paramedics actively requesting feedback. Initiatives were usually not part of a broader organisational or educational initiative (n = 34, 85.0%), did not include an action plan (n = 28, 70.0%) and were not underpinned by existing theory (n = 24, 60.0%).
The most frequently stated initiative aims were to close the prehospital feedback loop (n = 19, 47.5%) and enhance EMS professionals’ ability to reflect and learn (n = 21, 52.5%). The main barriers surrounding initiative development were information governance (e.g., patient confidentiality, data protection, data sharing agreements, Caldicott guardians, confidentiality advisory group approval; n = 18, 45.0%) and technology (e.g., lack of data linkage, poor data quality, data security limitations; n = 7, 17.5%). Ongoing concerns related to initiatives not being sustainable due to the time and resources required to generate the feedback (n = 16, 40.0%). The main facilitators were collaborative working and support from hospitals and broader networks (n = 14, 35.0%), the enthusiasm of EMS professionals to receive feedback (n = 9, 22.5%) and engagement from hospital staff (n = 8, 20.0%), which one initiative supported by accrediting involvement as part of continuing professional development portfolios.

3.3. Characteristics of Case Study Sites

The case study sites included two patient outcome feedback initiatives, one clinical performance initiative and one encompassing both patient outcome and clinical performance feedback (Table 2). They were a mixture of ‘push’ and ‘pull’ initiatives from across the UK. Seventeen interviews and six documents were analysed. The interviews lasted 29–93 min, with a median duration of 45 min. Most of the 17 participants were male (n = 11, 64.7%), and the median participant length in service was 14 years, ranging from 4 to 38 years. Participants were EMS managers or specialist clinicians (n = 7), regional network clinicians (n = 1), paramedics (n = 8) and emergency medical technicians (n = 1).

3.3.1. Case Study 1

This initiative was set up at a rural hospital where staff turnover was high and ambulance handover delays were long. EMS staff “did not typically feel like they could go and ask [hospital staff] for feedback” (CS1-P1) because they “did not want to bother people” (CS1-P2), and there was no existing formal feedback mechanism. Local frontline staff generated the idea for this feedback initiative and co-designed it with the quality improvement team, including the format of requesting feedback (paper form) and receiving the follow-up information (Email). Co-design resulted in a sense of ownership (“we said, we did” CS1-P1). Participants looked after the physical post-box and understood how the initiative worked, so fewer instructions were required. The initiative “successfully created a process for ambulance staff to obtain feedback locally” (CS1-D1) but was terminated after six months due to not being sustainable because it relied on the initiative lead manually retrieving the request forms, making contact with hospital clinicians and Emailing the feedback to the requester.
Staff receiving the feedback were “very satisfied” or “satisfied” (CS1-D2) with the feedback process and were optimistic about its format and timeliness, including that it fit well with their current way of working: “If you have a patient you are interested in, just fill out these details and then there’s a post-box in the drugs cupboard—literally just stick it in there” (CS1-P3).
Feedback recipients indicated that the feedback positively impacted their reflective or clinical practice as they “normally never get closure” (CS1-D1). Participants described how the feedback had increased their confidence by allowing them to “hear that you’ve been right about something, especially when you’re not 100% sure” (CS1-P3) and provided them with closure as it “puts your mind at rest” (CS1-P4). It also changed their behaviour during subsequent patient encounters, with one participant describing that it “informed my practice further if I saw a patient in a similar situation again” (CS1-P2) and another recalling a specific patient encounter: “I was very glad to have taken her to hospital and will always remember to consider pulmonary embolism in back pain in future” (CS1-D2).
Feedback recipients expressed concerns that some staff may not engage with a feedback initiative due to viewing feedback within the ambulance service as punitive. Others might over-use it if it was rolled out at scale. However, the initiative lead’s impression was that “people were quite conservative with it” (CS1-P1) and talked about users “rationing” their requests (CS1-P2).

3.3.2. Case Study 2

A condition-specific regional network set up this ‘breach-reporting’ initiative to provide feedback when EMS staff deviated from their protocol stipulating direct transport to a specialist unit for a subset of patients. Terminology for this initiative varied, whereby “the hospital call it breach reporting […], but we [in the ambulance service] call it feedback because the primary thing that we want to do with the feedback is not to berate the crew […] but take some learning from it to improve the next time they go to a job” (CS2-P2).
The initiative involved hospital clinicians feeding back information on inappropriate conveyance decisions to their specialist unit to a dedicated senior ambulance clinician. The senior ambulance clinician would then “review the incident as lead clinician and Email it out to the local ambulance team leaders for face-to-face feedback to the crews” (CS2-P2), where their decision-making was discussed to “allow them to learn from their experience and make sure a similar patient in the next situation received the correct care that the pathway states” (CS2-P1).
The senior clinician also aggregated feedback reports and took any “patterns back to the learning forum to share wider” (CS2-P2) and “put some education packages together” (CS2-P1). Engagement from hospital staff was achieved by introducing the new breach-reporting system during the “staff huddle”, ensuring the reporting did not take long to complete and identifying “champions” who sent the breach-reporting Email (CS2-P1). Engagement with senior clinicians and managers from the ambulance service was ensured early in the initiative’s development. However, no frontline staff were involved, justified by team leaders facilitating the feedback face-to-face to avoid “demoralising” frontline staff by sending them an Email (CS2-P1).
Initiative leads noted that since implementation several years ago, the number of breach reports had decreased, potentially indicating that the appropriateness of paramedics’ conveyance decisions had improved. However, initiative leads reported that hospital staff told them they were “too busy at times nowadays to fill these [breach-reporting Emails] in” (CS2-P1), which could also be the cause for decreased breach reporting. A further explanation could be increased informal feedback conversations due to improved relationships between hospital and EMS staff resulting in hospital staff being “quite accepting in terms of asking why maybe the [ambulance] crew have done that [i.e., inappropriately conveyed a patient] rather than telling them they have done wrong” (CS2-P1).
This initiative focused predominantly on learning; no participants mentioned staff wellbeing or closure as considerations.

3.3.3. Case Study 3

This initiative used an electronic dashboard to provide patient outcome feedback to EMS professionals. A local specialist paramedic set it up at two hospital trusts using local information-sharing agreements to allow EMS staff to “audit the quality of their work” by displaying the paramedic’s “provisional [prehospital] diagnosis and then comparing this with the final [hospital] diagnosis” (CS3-D1).
Although this initiative involved information being pushed into the dashboard, access was only granted to EMS professionals who requested this from the initiative lead, who therefore classed it as a ‘pull’ initiative. Feedback recipients described that it “prompted them to reflect”, undertake “further reading to refresh and broaden their knowledge”, and led to “clinical discussions with colleagues” (CS3-D3), thereby improving “decision-making confidence and self-efficacy” (CS3-D2). Receiving feedback was considered a novel experience that assisted with “pattern recognition”, with a paramedic explaining that “if I went to a similar job again, I would think I wonder if this is what it is and I wouldn’t have had that knowledge had I not looked it up” (CS3-P3).
Feedback was considered particularly useful “if the ED diagnosis differed from their prehospital clinical impressions” (CS3-D3) and for patients “that were treatable in the community” where “with the right knowledge and reason, you might leave somebody at home rather than take them to hospital” (CS3-P5).
One participant compared this feedback initiative to working in primary care or hospital trusts where patient information can be accessed in a central location. Another participant thought that the rural setting of this formal feedback initiative may have contributed to the high engagement from EMS staff, as there were limited opportunities for informal feedback provision (i.e., paramedics not returning to the same hospital later in their shift and missing out on asking hospital staff about their patient’s outcome). Initiative leads reported that technological problems meant the initiative had been paused for over a year when the interviews were conducted.
When prompted, participants reported that the feedback “provides an opportunity to clear up mysteries and support wellbeing” (CS3-D2). Participants also noted that receiving confirmation of their prehospital clinical impression improved confidence (“almost like a pat on the back” CS3-P4) and aided job satisfaction by making feedback recipients “feel more valued and more appreciated because somebody was investing time in you” (CS3-P3). However, the main aim of this initiative was to support reflection, learning and professional development.

3.3.4. Case Study 4

This initiative combined patient outcome feedback with clinical performance feedback. A senior ambulance clinician set the initiative up following an audit that indicated poor compliance in performing a specific prehospital skill, with the aim of “sharing outcomes of what happened to patients, so the individual EMS clinicians can learn by closing the learning loop as well as have a positive impact on the wellbeing of staff because it provides a bit of closure for them” (CS4-P1).
The initiative lead noted that gaps identified as part of this audit and feedback initiative led to an “organisational change in how we teach and…guide our staff to manage that clinical scenario” (CS4-P1), including a checklist to support EMS clinicians. The feedback included how compliant recipients were with the checklist, alongside patient outcomes and comments from hospital clinicians regarding how well the prehospital skill was carried out. The relevance of hospital clinicians’ comments varied with the initiative lead noting that “prehospital care is a specialist area that hospital clinicians will not necessarily have any experience of” (PS4-P1) and, therefore, advice may not be in line with paramedics’ scope of practice or EMS guidelines.
The lag time of this initiative was long (1–6 months), with the initiative lead concerned that providing feedback so long after the event may be “a waste of time” (PS4-P1) and all four feedback recipients noting they had already informally followed up on patients. However, the particular skill in question was rarely performed, meaning that all feedback recipients recalled the patient in question and that the written feedback contained “a lot more useful information about specific injuries [that] gives you a more comprehensive understanding of what happened” (CS4-P2). Participants expressed that they felt appreciated and that receiving feedback via this initiative boosted morale because it went above and beyond what they usually received.
Participants reported that they “didn’t really even know there was a feedback initiative going on” (CS4-P3) and that receiving the feedback Email “was a bit of a surprise” (CS4-P2). They described that the “gentle and respectful” (CS4-P2) tone of the written feedback meant they engaged positively despite the initial surprise. All participants discussed the feedback with their crew mates and “reflected on how we handled the situation” (CS4-P2). Several participants mentioned that the feedback “made me more confident for the next time” (PS4-P5) and had changed their practice, for example, increasing the detail of their documentation. In contrast, others stated that they kept their practice the same as they had sought feedback informally. Participants described themselves as having a positive attitude towards feedback and gave examples of when they “provided feedback to ambulance colleagues and students, as well as receiving it” (PS4-P4). They hypothesised that people with a less positive outlook on feedback working in an organisation that “hasn’t got a particularly well-established culture of feedback” (CS4-P4) might “find it challenging” (CS4-P2) to receive written feedback auditing their clinical performance.

3.4. Programme Theory of Feedback within EMS

Drawing upon analysis of the case studies, we developed 24 CMOCs, which explain how EMS feedback should work and what factors might influence its implementation and effectiveness. The CMOCs were divided into those relating to the implementation of EMS feedback interventions (Table 3, n = 11) and those relating to the effectiveness of patient outcome (Table 4, n = 9) and clinical performance feedback (Table 5, n = 4). The overall programme theory encompassed feedback and implementation characteristics (mechanism—resources), healthcare professional and organisational characteristics (context), processing and actioning feedback (mechanism—reasoning) and implementation, staff and service outcomes (outcomes), as visualised in our logic model (Figure 2).

4. Discussion

This study sought to draw upon multiple data sources to provide a comprehensive and theoretically informed review of practice in the under-researched area of EMS feedback. We received survey responses from 40 initiatives providing feedback to EMS professionals across the majority of UK ambulance services (11 out of 14). Most initiatives provided individual patient outcome feedback to individual frontline EMS staff using an Email format triggered by a request for feedback. Drawing upon synthesis from multiple data sources across the review of practice study and prior theory, we developed 24 CMOCs describing the mechanisms by which feedback in EMS operates to underpin our programme theory. The programme theory explains how the characteristics and implementation of an EMS feedback intervention (mechanism-resources) situated within particular healthcare professional and organisational characteristics (context) trigger individuals to process and action the feedback (mechanism-reasoning), resulting in implementation, staff and service outcomes (outcomes).
The survey revealed that EMS feedback initiatives in the UK focused predominantly on patient outcome feedback (n = 31). This contrasts with the published evidence synthesised in our systematic review [2], where most interventions focused on clinical performance feedback. The definition of feedback was the same for both studies (‘an initiative providing feedback to emergency ambulance staff regarding their performance within prehospital practice and/or patient outcomes’). Publication bias may explain the difference between what is done in practice and what gets reported in the peer-reviewed literature. Another explanation could be that EMS professionals more frequently associate ‘feedback’ with patient outcome rather than clinical performance feedback. This is consistent with findings from our interview study [9], where patient outcome feedback was the feedback type most frequently mentioned.
Patient outcome feedback was predominantly provided via pull-feedback initiatives, possibly due to this information being more challenging to obtain routinely within EMS. Although only CS3 used an electronic dashboard to provide patient outcome feedback, considerable opportunity exists to enhance feedback provision to EMS personnel through data linkage and integrated datasets spanning service boundaries [40,41]. Increased electronic data capture and feedback provision may address the lack of sustainability caused by resource constraints in generating feedback within EMS. However, careful consideration should be given to the support offered to EMS staff receiving feedback through automated electronic initiatives.
Around half of the surveyed EMS feedback initiatives reported challenges with information governance, echoing previous qualitative work [9,42]. Data governance issues related to patient outcomes being shared with clinicians that were no longer involved in a patient’s care [42], with EMS feedback initiatives taking various approaches to overcome this issue. The framing of feedback as an essential part of the learning cycle and agreeing on common terminology is vital to support the development of EMS feedback.
All initiatives in our survey involved feedback from a hospital trust or regional network to EMS staff. Within the UK, the work of paramedics routinely involves referrals to community services or primary care physicians [43,44], but we have yet to identify any feedback initiatives spanning these boundaries. A UK-wide study on EMS non-conveyance [45] found high variability in conveyance rates, with a systematic review suggesting that feedback interventions may improve patient safety in this area [46]. Our previous qualitative work suggests that EMS professionals strongly desire feedback on non-conveyed patients [9], indicating a clear need for feedback intervention development in this area, particularly in light of current policies encouraging non-conveyance [44].
Our analysis provided insight into the theoretical basis of existing EMS feedback interventions and the need to strengthen theory in this field. Participants’ understanding of ‘underlying theory’ encompassed various factors, such as expert opinion, replicating other initiatives, clinical governance, education, job satisfaction and continued professional development. In the social sciences, a theory is defined as “an ordered set of assertions about a generic behaviour or structure assumed to hold throughout a significantly broad range of specific instances” [47] (p. 9). There are different levels of theory, i.e., programme theories (as developed in this study), mid-range theories (such as CP-FIT, which are restricted to a subset of a social phenomenon within a specific context) and grand theories (which seek to offer a comprehensive meta-narrative applicable to any context) [22,48]. Our participants’ examples of theories deviated from the traditional definition or levels of theory, suggesting a possible misunderstanding or lack of awareness of existing feedback theory. The absence of theoretical foundations in feedback initiatives has been critiqued in other healthcare settings, as it impedes progress in feedback intervention science [20,49].
We have previously used CP-FIT—a mid-range theory of clinical performance feedback [16]—to explain participants’ involvement with feedback at an abstract level [9]. However, the CP-FIT definition of mechanisms as “underlying explanations of how and why an intervention works” [16] (p. 2) did not seem to fully address mechanisms as interpreted within our chosen realist methodology, i.e., “a combination of resources offered by the social programme under study and stakeholders’ reasoning in response” [34] (p. 3). Adding an established list of mechanisms of action from behaviour change theory [39] to our data analysis allowed us to consider in-depth psychological processes by which outcomes were achieved. Utilising realist methodology allowed us to explore how these mechanisms linked to contexts and outcomes. This enabled the current study to go further than our systematic review of the published literature by adding depth of understanding [2].

4.1. Implications and Future Research

Our study demonstrates that feedback initiatives are common within EMS but that there is considerable divergence between research evidence into effective audit and feedback gleaned largely from other healthcare settings and implementation within EMS. Therefore, an important question is whether this implementation gap is due to unique challenges in the EMS setting for this type of intervention or other factors that can be addressed through future implementation research. Our findings suggest that the traditionally poor feedback culture within ambulance services and the practical difficulty of sharing data across hospital-EMS boundaries may be reasons for this divergence.
Our study demonstrates that existing theory can be used to evaluate existing feedback initiatives within EMS. Our programme theory will inform improvements and changes to ongoing EMS feedback initiatives. To support this, we are developing a best-practice guide for EMS feedback in collaboration with stakeholders based on our study findings.
Future research should explore EMS culture around the provision of feedback and seek to understand the prevalence, predictors and effects of feedback using quantitative methods. Our context–mechanism–outcome matrices (Table 3, Table 4 and Table 5) serve as a source of empirically based testable hypotheses for future research on EMS feedback. Our hypotheses expand others developed specifically within audit and feedback [12,14,16] to include ‘pull’ initiatives requiring feedback-seeking behaviour and alternative feedback types and outcomes, such as closure from patient outcome feedback. By not just limiting our review of practice to clinical performance feedback, we were able to develop a more holistic programme theory of feedback within EMS, allowing prospective initiative leads to consider a variety of feedback types and effects [50].
Although respondents anecdotally reported the effectiveness of local programmes, research is needed to provide robust quasi-experimental evidence to support feedback models in EMS. Our logic model expanded CP-FIT’s effectiveness outcomes [16] to include Proctor’s implementation outcomes [38] and could be used to design hybrid effectiveness-implementation trials of EMS feedback initiatives. Hybrid designs combine questions concerning intervention effectiveness with questions about how best to implement it in one study using a range of research designs [51,52].

4.2. Strengths and Limitations

Our review of current practice focused on EMS systems within the UK. Although a good representation of the 14 emergency ambulance services was achieved, it remains for future research to establish the degree to which these findings generalise to EMS in other health systems. The limits of existing international research in EMS feedback make generalisation difficult. However, our work draws upon established theory within the international implementation science audit and feedback domain, which should lay the groundwork for future comparative analysis. There is no prior literature regarding the number of EMS feedback initiatives in the UK to allow us to estimate the population size. However, we received many responses using a relatively small sampling framework, i.e., only eliciting responses from initiative leads.
Whilst our survey coverage of UK practice spanned 11 of 14 EMS, it did not include all EMS feedback initiatives currently or previously active. It was impossible to calculate a response rate for our survey, as we could not determine how many people received the survey due to the advertisement method. We received no responses from initiatives within London, South Central and the Isle of Wight, though the lead researcher’s professional network suggests the presence of feedback initiatives in some of those geographical areas.
The study protocol asked participating ambulance trusts to identify feedback initiatives based on or by identifying existing data-sharing agreements. However, we avoided asking initiative leads what these were. Some ambulance trusts expressed concern that not all initiatives would have robust information governance approvals. Initiatives without these approvals were, therefore, potentially underrepresented in our sample.
Our case study approach facilitated insightful analysis of specific practice initiatives that would not have been possible with other study designs. We collected less data for Case study 2 than other sites, which may have introduced selection bias. This was due to difficulty identifying participants who had received feedback despite extensive efforts to liaise directly with frontline ambulance staff, initiative leads and the ambulance service’s research department. We were unable to determine whether this was due to the feedback being so embedded in routine practice that recipients were unaware of it being linked to a formal feedback initiative, a lack of willingness to participate in research or disparity between actual practice and initiative leads’ perception of staff engagement with the initiative.

5. Conclusions

This study suggests that initiatives providing feedback for EMS staff are common in practice, with most UK ambulance trusts currently having examples of feedback initiatives within their footprint. EMS feedback initiatives are motivated by various factors but are challenging to implement and sustain effectively. The published literature differs from current practice in the focus and scope of EMS feedback initiatives. There is considerable opportunity to strengthen the methodological and theoretical basis for innovations in practice in this area. This includes providing guidance on the design of interventions and robust evaluation of different feedback approaches, especially those which might be unique in meeting the challenges of the EMS setting.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/healthcare11162229/s1, File S1: Survey measure; File S2: Interview topic guide for feedback initiative leads; File S3: Interview topic guide for feedback recipients.

Author Contributions

C.W. conceived the study and developed the survey instrument and topic guide under the supervision of J.B.; G.J. and J.B. obtained research funding; C.W. developed the study protocol, obtained relevant ethics and governance approvals, collected the data, analysed the data and drafted the manuscript under supervision and guidance from G.J., R.L. and J.B.; J.B. was the primary supervisor for this project; C.W. takes full responsibility for the paper as a whole. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Institute for Health Research (NIHR) Yorkshire and Humber Patient Safety Translational Research Centre (NIHR Yorkshire and Humber PSTRC). The views expressed are those of the authors and not necessarily those of the NIHR or the Department of Health and Social Care.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the University of Leeds ethics committee (PSYC-242 7 July 2021) and the Health Research Authority (IRAS project ID 287110).

Informed Consent Statement

Before the survey and interviews, respondents were provided with standardised research information regarding the study’s purpose, and informed consent was obtained.

Data Availability Statement

The datasets generated and analysed during the current study are not publicly available as sharing the raw data would violate the agreement to which participants consented; however, the manuscript includes an extensive list of quotes, and the datasets are available from the corresponding author on reasonable request.

Acknowledgments

The authors would like to thank the study participants for completing the survey and participating in interviews and the research departments of participating ambulance trusts for their support with advertisement and recruitment. Thank you to Julia Williams, University of Hertfordshire/South East Coast Ambulance Service NHS Trust, and A. Niroshan Siriwardena, University of Lincoln, for peer-reviewing the study protocol. Thank you to Matthew Morys-Edge, Research and Implementation Assistant, Bradford Institute for Health Research, for assisting with transcription.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the study design; in the collection, analyses or interpretation of data; in the writing of the manuscript or in the decision to publish the results.

Abbreviations

CMOcContext–Mechanism–Outcome Configuration
CP-FITClinical Performance Feedback Intervention Theory
EMSEmergency Medical Services
IQRInterquartile Range
UKUnited Kingdom
WPWork Package

References

  1. Ivers, N.M.; Jamtvedt, G.; Flottorp, S.; Young, J.M.; Odgaard-Jensen, J.; French, S.D.; O’Brien, M.A.; Johansen, M.; Grimshaw, J.; Oxman, A.D. Audit and feedback: Effects on professional practice and healthcare outcomes. Cochrane Database Syst. Rev. 2012, 6, CD000259. [Google Scholar] [CrossRef] [PubMed]
  2. Wilson, C.; Janes, G.; Lawton, R.; Benn, J. Types and effects of feedback for emergency ambulance staff: A systematic mixed studies review and meta-analysis. BMJ Qual. Saf. 2023. [Google Scholar] [CrossRef] [PubMed]
  3. Fisher, J.D.; Freeman, K.; Clarke, A.; Spurgeon, P.; Smyth, M.; Perkins, G.D.; Sujan, M.A.; Cooke, M.W. Health services and delivery research. In Patient Safety in Ambulance Services: A Scoping Review; NIHR Journals Library: Southampton, UK, 2015. [Google Scholar]
  4. Lawn, S.; Roberts, L.; Willis, E.; Couzner, L.; Mohammadi, L.; Goble, E. The effects of emergency medical service work on the psychological, physical, and social well-being of ambulance personnel: A systematic review of qualitative research. BMC Psychiatry 2020, 20, 348. [Google Scholar] [CrossRef] [PubMed]
  5. Paulin, J.; Kurola, J.; Koivisto, M.; Iirola, T. Ems non-conveyance: A safe practice to decrease ed crowding or a threat to patient safety? BMC Emerg. Med. 2021, 21, 115. [Google Scholar] [CrossRef]
  6. Blodgett, J.M.; Robertson, D.J.; Pennington, E.; Ratcliffe, D.; Rockwood, K. Alternatives to direct emergency department conveyance of ambulance patients: A scoping review of the evidence. Scand. J. Trauma Resusc. Emerg. Med. 2021, 29, 4. [Google Scholar] [CrossRef]
  7. Morrison, L.; Cassidy, L.; Welsford, M.; Chan, T.M. Clinical performance feedback to paramedics: What they receive and what they need. AEM Educ. Train. 2017, 1, 87–97. [Google Scholar] [CrossRef] [Green Version]
  8. Cash, R.E.; Crowe, R.P.; Rodriguez, S.A.; Panchal, A.R. Disparities in feedback provision to emergency medical services professionals. Prehospital Emerg. Care 2017, 21, 773–781. [Google Scholar] [CrossRef]
  9. Wilson, C.; Howell, A.-M.; Janes, G.; Benn, J. The role of feedback in emergency ambulance services: A qualitative interview study. BMC Health Serv. Res. 2022, 22, 296. [Google Scholar] [CrossRef]
  10. Cifra, C.L.; Sittig, D.F.; Singh, H. Bridging the feedback gap: A sociotechnical approach to informing clinicians of patients’ subsequent clinical course and outcomes. BMJ Qual. Saf. 2021, 30, 591–597. [Google Scholar] [CrossRef]
  11. Jamtvedt, G.; Young, J.M.; Kristoffersen, D.T.; O’Brien, M.A.; Oxman, A.D. Does telling people what they have been doing change what they do? A systematic review of the effects of audit and feedback. Qual. Saf. Health Care 2006, 15, 433–436. [Google Scholar] [CrossRef]
  12. Brehaut, J.C.; Colquhoun, H.L.; Eva, K.W.; Carroll, K.; Sales, A.; Michie, S.; Ivers, N.; Grimshaw, J.M. Practice feedback interventions: 15 suggestions for optimizing effectiveness. Ann. Intern. Med. 2016, 164, 435–441. [Google Scholar] [CrossRef]
  13. Colquhoun, H.L.; Michie, S.; Sales, A.; Ivers, N.; Grimshaw, J.M.; Carroll, K.; Chalifoux, M.; Eva, K.; Brehaut, J. Reporting and design elements of audit and feedback interventions: A secondary review. BMJ Qual. Saf. 2017, 26, 54–60. [Google Scholar] [CrossRef]
  14. Colquhoun, H.L.; Carroll, K.; Eva, K.W.; Grimshaw, J.M.; Ivers, N.; Michie, S.; Sales, A.; Brehaut, J.C. Advancing the literature on designing audit and feedback interventions: Identifying theory-informed hypotheses. Implement. Sci. 2017, 12, 117. [Google Scholar] [CrossRef]
  15. Rapin, J.; Pellet, J.; Mabire, C.; Gendron, S.; Dubois, C.-A. How does feedback shared with interprofessional health care teams shape nursing performance improvement systems? A rapid realist review protocol. Syst. Rev. 2019, 8, 182. [Google Scholar] [CrossRef] [Green Version]
  16. Brown, B.; Gude, W.T.; Blakeman, T.; van der Veer, S.N.; Ivers, N.; Francis, J.J.; Lorencatto, F.; Presseau, J.; Peek, N.; Daker-White, G. Clinical performance feedback intervention theory (cp-fit): A new theory for designing, implementing, and evaluating feedback in health care based on a systematic review and meta-synthesis of qualitative research. Implement. Sci. 2019, 14, 40. [Google Scholar] [CrossRef] [Green Version]
  17. Lyon, R.M.; Clarke, S.; Milligan, D.; Clegg, G.R. Resuscitation feedback and targeted education improves quality of pre-hospital resuscitation in scotland. Resuscitation 2012, 83, 70–75. [Google Scholar] [CrossRef]
  18. Sommers, N.; Evans, L.; Collingwood, T.; Dykes, L. PP29 ‘Breaking down Barriers: Experiences from the first year of an ED to EMS feedback ‘Postbox. Emerg. Med. J. 2019, 36, e13. [Google Scholar] [CrossRef]
  19. Snowsill, M.; Cracolici, G.; Wieder, T.; Allen, G. Facilitated hospital to pre-hospital feedback for professional development (phem feedback): A service evaluation using a self-reported questionnaire to understand the experiences of participating pre-hospital clinicians in the first year of operation. Br. Paramed. J. 2023, 8, 42–52. [Google Scholar] [CrossRef]
  20. Foy, R.; Skrypak, M.; Alderson, S.; Ivers, N.M.; McInerney, B.; Stoddart, J.; Ingham, J.; Keenan, D. Revitalising audit and feedback to improve patient care. BMJ 2020, 368, m213. [Google Scholar] [CrossRef] [Green Version]
  21. Davidoff, F. Understanding contexts: How explanatory theories can help. Implement. Sci. 2019, 14, 23. [Google Scholar] [CrossRef]
  22. Davidoff, F.; Dixon-Woods, M.; Leviton, L.; Michie, S. Demystifying theory and its use in improvement. BMJ Qual. Saf. 2015, 24, 228–238. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Funnell, S.; Rogers, P. Purposeful Program Theory: Effective Use of Theories of Change and Logic Models; Wiley: Hoboken, NJ, USA, 2011. [Google Scholar]
  24. Pawson, R.; Tilley, N. Realistic Evaluation; Sage: Los Angeles, CA, USA, 1997. [Google Scholar]
  25. Pawson, R.; Manzano-Santaella, A. A realist diagnostic workshop. Evaluation 2012, 18, 176–191. [Google Scholar] [CrossRef] [Green Version]
  26. NHS Health Research Authority. UK Policy Framework for Health and Social Care Research; NHS Health Research Authority: London, UK, 2017. [Google Scholar]
  27. Wong, G.; Westhorp, G.; Manzano, A.; Greenhalgh, J.; Jagosh, J.; Greenhalgh, T. Rameses ii reporting standards for realist evaluations. BMC Med. 2016, 14, 96. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Sharma, A.; Duc, N.T.M.; Thang, T.L.L.; Nam, N.H.; Ng, S.J.; Abbas, K.S.; Huy, N.T.; Marušić, A.; Paul, C.L.; Kwok, J.; et al. A consensus-based checklist for reporting of survey studies (cross). J. Gen. Intern. Med. 2021, 36, 3179–3187. [Google Scholar] [CrossRef]
  29. Manzano, A. The craft of interviewing in realist evaluation. Evaluation 2016, 22, 342–360. [Google Scholar] [CrossRef] [Green Version]
  30. Mukumbang, F.C.; Marchal, B.; Van Belle, S.; van Wyk, B. Using the realist interview approach to maintain theoretical awareness in realist studies. Qual. Res. 2019, 20, 485–515. [Google Scholar] [CrossRef] [Green Version]
  31. De Weger, E.; Van Vooren, N.J.E.; Wong, G.; Dalkin, S.; Marchal, B.; Drewes, H.W.; Baan, C.A. What’s in a realist configuration? Deciding which causal configurations to use, how, and why. Int. J. Qual. Methods 2020, 19, 1609406920938577. [Google Scholar] [CrossRef]
  32. De Brún, A.; McAuliffe, E. Identifying the context, mechanisms and outcomes underlying collective leadership in teams: Building a realist programme theory. BMC Health Serv. Res. 2020, 20, 261. [Google Scholar] [CrossRef]
  33. Greenhalgh, J.; Manzano, A. Understanding ‘context’ in realist evaluation and synthesis. Int. J. Soc. Res. Methodol. 2022, 25, 583–595. [Google Scholar] [CrossRef]
  34. Dalkin, S.; Greenhalgh, J.; Jones, D.; Cunningham, B.; Lhussier, M. What’s in a mechanism? Development of a key concept in realist evaluation. Implement. Sci. 2015, 10, 49. [Google Scholar] [CrossRef] [Green Version]
  35. Gilmore, B.; McAuliffe, E.; Power, J.; Vallières, F. Data analysis and synthesis within a realist evaluation: Toward more transparent methodological approaches. Int. J. Qual. Methods 2019, 18, 1609406919859754. [Google Scholar] [CrossRef] [Green Version]
  36. Jagosh, J. Retroductive theorizing in pawson and tilley’s applied scientific realism. J. Crit. Realism 2020, 19, 121–130. [Google Scholar] [CrossRef]
  37. Mukumbang, F.C.; Kabongo, E.M.; Eastwood, J.G. Examining the application of retroductive theorizing in realist-informed studies. Int. J. Qual. Methods 2021, 20, 16094069211053516. [Google Scholar] [CrossRef]
  38. Proctor, E.; Silmere, H.; Raghavan, R.; Hovmand, P.; Aarons, G.; Bunger, A.; Griffey, R.; Hensley, M. Outcomes for implementation research: Conceptual distinctions, measurement challenges, and research agenda. Adm. Policy Ment. Health Ment. Health Serv. Res. 2011, 38, 65–76. [Google Scholar] [CrossRef] [Green Version]
  39. Carey, R.N.; Connell, L.E.; Johnston, M.; Rothman, A.J.; de Bruin, M.; Kelly, M.P.; Michie, S. Behavior change techniques and their mechanisms of action: A synthesis of links described in published intervention literature. Ann. Behav. Med. A Publ. Soc. Behav. Med. 2019, 53, 693–707. [Google Scholar] [CrossRef] [Green Version]
  40. Porter, A.; Badshah, A.; Black, S.; Fitzpatrick, D.; Harris-Mayes, R.; Islam, S.; Jones, M.; Kingston, M.; LaFlamme-Williams, Y.; Mason, S.; et al. Electronic health records in ambulances: The era multiple-methods study. Health Serv. Deliv. Res. 2020, 8, 1–140. [Google Scholar] [CrossRef]
  41. Cuk, S.; Wimmer, H.; Powell, L.M. Problems associated with patient care reports and transferring data between ambulance and hospitals from the perspective of emergency medical technicians. Issues Inf. Syst. 2017, 18, 16–26. [Google Scholar]
  42. Eaton-Williams, P.; Mold, F.; Magnusson, C. Exploring paramedic perceptions of feedback using a phenomenological approach. Br. Paramed. J. 2020, 5, 7–14. [Google Scholar] [CrossRef]
  43. O’Hara, R.; Johnson, M.; Siriwardena, A.N.; Weyman, A.; Turner, J.; Shaw, D.; Mortimer, P.; Newman, C.; Hirst, E.; Storey, M.; et al. A qualitative study of systemic influences on paramedic decision making: Care transitions and patient safety. J. Health Serv. Res. Policy 2015, 20, 45–53. [Google Scholar] [CrossRef] [Green Version]
  44. NHS Improvement. Operational Productivity and Performance in English NHS Ambulance Trusts: Unwarranted Variations; NHS Improvement: London, UK, 2018. [Google Scholar]
  45. O’Cathain, A.; Knowles, E.; Bishop-Edwards, L.; Coster, J.; Crum, A.; Jacques, R.; James, C.; Lawson, R.; Marsh, C.; O’Hara, R.; et al. Understanding Variation in Ambulance Service Non-Conveyance Rates: A Mixed Methods Study; NIHR Journals Library: Southampton, UK, 2018. [Google Scholar]
  46. Ebben, R.H.A.; Vloet, L.C.M.; Speijers, R.F.; Tönjes, N.W.; Loef, J.; Pelgrim, T.; Hoogeveen, M.; Berben, S.A.A. A patient-safety and professional perspective on non-conveyance in ambulance care: A systematic review. Scand. J. Trauma Resusc. Emerg. Med. 2017, 25, 71. [Google Scholar] [CrossRef] [Green Version]
  47. Weick, K.E. Theory construction as disciplined imagination. Acad. Manag. Rev. 1989, 14, 516–531. [Google Scholar] [CrossRef]
  48. Kislov, R.; Pope, C.; Martin, G.P.; Wilson, P.M. Harnessing the power of theorising in implementation science. Implement. Sci. 2019, 14, 103. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Ivers, N.M.; Sales, A.; Colquhoun, H.; Michie, S.; Foy, R.; Francis, J.J.; Grimshaw, J.M. No more ‘business as usual’ with audit and feedback interventions: Towards an agenda for a reinvigorated intervention. Implement. Sci. 2014, 9, 14. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Denisi, A.; Sockbeson, C.E.S. Feedback sought vs. feedback given: A tale of two literatures. Manag. Res. J. Iberoam. Acad. Manag. 2018, 16, 320–333. [Google Scholar] [CrossRef]
  51. Curran, G.M.; Landes, S.J.; McBain, S.A.; Pyne, J.M.; Smith, J.D.; Fernandez, M.E.; Chambers, D.A.; Mittman, B.S. Reflections on 10 years of effectiveness-implementation hybrid studies. Front. Health Serv. 2022, 2, 1053496. [Google Scholar] [CrossRef]
  52. Curran, G.M.; Bauer, M.; Mittman, B.; Pyne, J.M.; Stetler, C. Effectiveness-implementation hybrid designs: Combining elements of clinical effectiveness and implementation research to enhance public health impact. Med. Care 2012, 50, 217–226. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Study flow chart [2,9].
Figure 1. Study flow chart [2,9].
Healthcare 11 02229 g001
Figure 2. Logic model depicting our programme theory of EMS feedback.
Figure 2. Logic model depicting our programme theory of EMS feedback.
Healthcare 11 02229 g002
Table 1. Characteristics of feedback initiatives in EMS (n = 40).
Table 1. Characteristics of feedback initiatives in EMS (n = 40).
CharacteristicsN%
Feedback typePatient outcome feedback3177.5
Clinical performance feedback410.0
Clinical performance + patient outcome feedback512.5
Currently activeYes2870.0
No1230.0
Year started
Median 2019
(IQR 2017–2020)
2004–201325.0
2014–20181332.5
2019–20222357.5
No response25.0
Instances of feedback<1012.5
10–491435.0
50–991127.5
100–199410.0
200–29925.0
300–39925.0
400–49912.5
>50025.0
No response37.5
Feedback recipient
(categories are not mutually exclusive)
Frontline EMS staff3690.0
Helicopter EMS staff1435.0
EMS managers922.5
Emergency operations centre staff25.0
EMS organisation37.5
Combination1640.0
FormatEmail or electronic document3587.5
Face-to-face37.5
Written letter12.5
No response12.5
FrequencyDaily25.0
Weekly37.5
Bi-weekly12.5
Monthly37.5
Quarterly12.5
Ad hoc2972.5
No response12.5
Lag time
Median 21 days
(IQR 7–30 days)
Within 24 h37.5
Within 7 days820.0
1–2 weeks615.0
2–4 weeks1435.0
4–12 weeks615.0
>12 weeks25.0
No response12.5
Recipient levelIndividual3280.0
Individual + team37.5
Individual + organisation12.5
Individual + team + organisation410.0
Patient casesIndividual3895.0
Aggregate25.0
Push or pullPush820.0
Pull2562.5
Combination615.0
No response12.5
Stand-aloneStand-alone3485.0
Part of an organisational or education initiative615.0
Action plan includedYes1230.0
No2870.0
Underlying theoryYes1537.5
No2460.0
No response12.5
Aim
(categories are not mutually exclusive)
Aid reflection, education + learning2152.5
Provide feedback (“close the loop”)1947.5
Improve staff wellbeing + job satisfaction922.5
Improve patient care615.0
Improve relationships between EMS + hospital staff615.0
Increase protocol adherence410.0
Satisfy EMS staff’s curiosity37.5
Clinical governance25.0
Barriers
(categories are not mutually exclusive)
Information governance1845.0
Time and resources required to generate the feedback1640.0
Technology717.5
No barriers410.0
EMS managers410.0
Lack of awareness from EMS staff that the initiative existed37.5
Identifying hospital collaborators and securing buy-in from hospital staff25.0
Concerns from EMS staff that errors identified in clinical practice would be used against them25.0
Facilitators
(categories are not mutually exclusive)
Collaboration1435.0
Enthusiasm from EMS staff922.5
Engagement from hospital staff820.0
Initiative lead with time + enthusiasm25.0
Alignment with organisational priorities25.0
No existing initiative providing feedback12.5
Information governance12.5
Conducting a pilot of the initiative first12.5
Inclusion of specific expertise within the team (e.g., data analyst)12.5
Table 2. Characteristics of case study sites.
Table 2. Characteristics of case study sites.
CharacteristicsCase Study 1Case Study 2Case Study 3Case Study 4
Data sourcesInitiative lead interview (n = 1), Feedback recipient interviews (n = 3), documents (n = 2)Initiative lead interviews (n = 2), Feedback recipient interview (n = 1), documents (n = 1)Initiative lead interviews (n = 2), Feedback recipient interviews (n = 3), documents (n = 3)Initiative lead interview (n = 1), Feedback recipient interviews (n = 4)
ContextRuralUrbanRuralRural, urban + suburban
Catchment area1 hospitalRegional2 hospitalsNational
Initiative leadEMS managerRegional network clinician + EMS specialist clinicianEMS specialist clinician + EMS managerEMS specialist clinician
Patient groupAll patients1 conditionAll patients1 condition
Feedback typePatient outcomeClinical performancePatient outcomeClinical performance + patient outcome feedback
Push or pullPullPushPullPush
Year started2018201520142020
Instances of feedback50–99Unknown>50010–49
Feedback recipientFrontline EMS staffFrontline EMS staffFrontline EMS staffFrontline + helicopter EMS staff
FormatEmailFace-to-faceElectronic dashboardEmail
FrequencyWhenever requestedWhen breach occurredWeeklyAfter every event
Lag-time2–4 weeksA few weeksUp to 7 days1–6 months
Table 3. Context–mechanism–outcome matrix of the implementation of EMS feedback interventions.
Table 3. Context–mechanism–outcome matrix of the implementation of EMS feedback interventions.
Mechanism (Resources)ContextOutcomesMechanism (Reasoning)Data Sources
Ifan initiative is co-designedwithfeedback recipients that have a positive feedback attitude and leadership supportthenacceptability is likely to be higher
becausethere is a sense of ownership, understanding and credibilityCS1-P1, CS1-P3
Ifa senior clinician provides push feedback on protocol adherencewithinan organisation that has leadership supportthenit is likely to be accepted
becausethere is a sense of credibilityCS2-P1, CS2-P2, CS4-P2, CS4-P3
Ifan initiative provides patient outcome feedback via a ‘pull’ initiativetohealthcare professionals that display feedback-seeking behaviourthenadoption and appropriateness are likely to be higher
becausethere is a sense of compatibilityCS1-P2, CS1-P3, CS1-P4, CS3-P2, CS3-P3
Ifa feedback initiative provides patient outcome feedbackwithinan organisation where informal follow-up opportunities are limitedthenadoption is likely to be higher
becausethere is a relative advantageCS1-P1, CS1-P1, CS3-P2, CS3-P3
Ifthe initiative lead guides feedback providers on paramedics’ scope of practicewhenfeedback providers do not have any experience or knowledge of EMS clinical practicethenfeedback is more likely to be appropriate
becauseit is actionableCS1-P1, CS4-P1
Iffeedback is meant to be generated with limited trainingwithinorganisations that can appoint champions to generate feedbackthenit is more likely to be feasible
becausethere is a resource matchCS1-P1, CS2-P1
Ifthe number of processed feedback requests is limited by having specific eligibility criteriawithinan organisation that has limited resources ring-fenced for this initiativethensustainability would be better
becausethere would be a better resource matchCS1-P1, CS4-P1
Ifpatient outcome feedback does not identify which patient it relates towithinan organisation where staff attend multiple patients per shift and feedback is delayedthenacceptability and adoption are likely to be poor
becauseit would be complex for recipients to identify whom the feedback related toCS1-P4
CS3-P5, CS4-P1, CS4-P2
Iffeedback is pushed out without training, support or raising awarenesswithinan organisation that traditionally has poor feedback culturethenacceptability is likely to be poor
becausethere is poor compatibilityCS2-P3, CS4-P2, CS4-P3, CS4-P4
Iffeedback is provided with a long lag-timetoa staff member that can informally follow up on patientsthenadoption is likely to be poor
becausethere is no relative advantageCS4-P1, CS4-P2, CS4-P3, CS4-P4
Ifan ambulance trust leads a feedback initiativewithoutclearly articulating how the feedback is going to be used by the trustthenadoption may be low
becauseit could be perceived as being punitiveCS1-P2, CS1-P4, CS4-P2, CS4-P4
Ifa feedback initiative has a high resource cost for one individualwithinan organisation where there are competing prioritiesthenit is not likely to be sustainable
becausethere is no resource matchCS1-P1, CS1-D2, CS4-P1
Note. indicates a positive outcome and indicates a negative outcome.
Table 4. Context–mechanism–outcome matrix of interventions providing patient outcome feedback for EMS staff.
Table 4. Context–mechanism–outcome matrix of interventions providing patient outcome feedback for EMS staff.
Mechanism (Resources)ContextOutcomesMechanism (Reasoning)Data Sources
Ifpatient outcome feedback allows comparisons between prehospital working impression and hospital diagnosisinan organisation that monitors this on an aggregate levelthenit is likely to improve patient safety
becauseKnowledge is improved by organisations arranging additional training; the environment is changed by organisations increasing alternative pathwaysCS3-D1, CS3-D2, CS3-D3, CS3-P3, CS4-P1
Ifpatient outcome feedback is providedtoambulance staff interacting with patients who are anxious about what might happen at the hospitalthenit is likely to improve service quality
becausestaff have a belief about consequences and can reassure patientsCS1-P2, CS3-P3
Ifpatient outcome feedback is provided regarding patients with a difference in prehospital/hospital diagnosis, patients frequently calling the ambulance service or patients with common conditionstohealthcare professionals that have a positive feedback attitude and within an organisation that supports autonomous decisions by staffthenit is likely to lead to individual behaviour change and better patient care
becauseit improves beliefs about consequencesCS1-P1, CS1-D1, CS3-P3, CS3-P5, CS4-P2, CS4-P5
Ifadditional learning materials are provided alongside patient outcome feedbacktohealthcare professionals that have a positive feedback attitudethenit is likely to lead to individual behaviour change and better patient care
becauseit improves knowledgeCS4-P2, CS1-P1, CS3-P5
Ifan initiative provides patient outcome feedback via a ‘pull’ initiativetohealthcare professionals that display feedback-seeking behaviourthenit may lead to behaviour change and improved clinical performance/patient safety
becauseit improves knowledge and decision-making processesCS1-P1, CS1-P2, CS1-P3, CS3-P2
Ifpatient outcome feedback is provided on patients whom staff have a particular interest intostaff who do not routinely find out what happens to their patientsthenit is likely to improve staff mental health
becauseit allows an element of closureCS1-P1, CS1-P2, CS1-P4, CS1-D1, CS3-D2
Ifa senior clinician sets up an initiative providing patient outcome feedbackwithinan organisation that has traditionally viewed patient outcome information as an ‘add on’thenit is likely to lead to increased job satisfaction
becauseit makes staff feel appreciatedCS3-P3, CS3-P4, CS4-P2, CS4-P3, CS4-P4, CS4-P5
Ifpatient outcome feedback is provided only by emergency department staffbutpatients are admitted to hospital wardsthenstaff behaviour change may be limited
becausethere is only a partial understanding of the consequencesCS1-P1, CS1-P3, CS3-P5
Note. indicates a positive outcome and indicates a negative outcome.
Table 5. Context–mechanism–outcome matrix of interventions providing clinical performance feedback for EMS staff.
Table 5. Context–mechanism–outcome matrix of interventions providing clinical performance feedback for EMS staff.
Mechanism (Resources)ContextOutcomesMechanism (Reasoning)Data Sources
Ifpathway adherence feedback is providedwithinan organisation that has a centralised model of care (i.e., bypassing local hospitals to attend specialist treatment centres)thenit will lead to behaviour change and improved patient safety
becausestaff will have had the pathway reinforcedCS2-P1, CS2-P2, CS4-P2, CS4-P5
Ifpush-feedback on protocol adherence is providedtoa healthcare professional that has a positive feedback attitudethenit is likely to lead to behaviour change
becauseit improves knowledgeCS2-P1, CS2-P2,
CS3-P2, CS3-P3, CS4-P2, CS4-P4
Iffeedback on protocol adherence is providedwithinan organisation that prioritises the protocol topicthenit is likely to improve clinical performance
becauseit improves knowledge and skillsCS2-P1, CS2-P2,
CS4-P1
Ifpush-feedback is providedtoa healthcare professional that has a low sense of confidencethenit may negatively affect staff mental health
becauseit reduces beliefs about capabilitiesCS1-P3, CS2-P1,
CS4-P2, CS4-P4
Note. indicates a positive outcome and indicates a negative outcome.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wilson, C.; Janes, G.; Lawton, R.; Benn, J. Feedback for Emergency Ambulance Staff: A National Review of Current Practice Informed by Realist Evaluation Methodology. Healthcare 2023, 11, 2229. https://doi.org/10.3390/healthcare11162229

AMA Style

Wilson C, Janes G, Lawton R, Benn J. Feedback for Emergency Ambulance Staff: A National Review of Current Practice Informed by Realist Evaluation Methodology. Healthcare. 2023; 11(16):2229. https://doi.org/10.3390/healthcare11162229

Chicago/Turabian Style

Wilson, Caitlin, Gillian Janes, Rebecca Lawton, and Jonathan Benn. 2023. "Feedback for Emergency Ambulance Staff: A National Review of Current Practice Informed by Realist Evaluation Methodology" Healthcare 11, no. 16: 2229. https://doi.org/10.3390/healthcare11162229

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop