Combining performance and outcome indicators can be used in a standardized way: a pilot study of two multidisciplinary, full-scale major aircraft exercises
© Rådestad et al.; licensee BioMed Central Ltd. 2012
Received: 19 March 2012
Accepted: 19 August 2012
Published: 28 August 2012
Disaster medicine is a fairly young scientific discipline and there is a need for the development of new methods for evaluation and research. This includes full-scale disaster exercisers. A standardized concept on how to evaluate these exercises, could lead to easier identification of pitfalls caused by system-errors in the organization. The aim of this study was to demonstrate the feasibility of using a combination of performance and outcome indicators so that results can be compared in standardized full-scale exercises.
Two multidisciplinary, full-scale exercises were studied in 2008 and 2010. The panorama had the same setup. Sets of performance indicators combined with indicators for unfavorable patient outcome were recorded in predesigned templates. Evaluators, all trained in a standardized way at a national disaster medicine centre, scored the results on predetermined locations; at the scene, at hospital and at the regional command and control.
All data regarding the performance indicators of the participants during the exercises were obtained as well as all data regarding indicators for patient outcome. Both exercises could therefore be compared regarding performance (processes) as well as outcome indicators. The data from the performance indicators during the exercises showed higher scores for the prehospital command in the second exercise 15 points and 3 points respectively. Results from the outcome indicators, patient survival and patient complications, demonstrated a higher number of preventable deaths and a lower number of preventable complications in the exercise 2010. In the exercise 2008 the number of preventable deaths was lower and the number of preventable complications was higher.
Standardized multidisciplinary, full-scale exercises in different settings can be conducted and evaluated with performance indicators combined with outcome indicators enabling results from exercises to be compared. If exercises are performed in a standardized way, results may serve as a basis for lessons learned. Future use of the same concept using the combination of performance indicators and patient outcome indicators may demonstrate new and important evidence that could lead to new and better knowledge that also may be applied during real incidents.
KeywordsAirplane crash Disaster preparedness Disaster management Disaster response Educational model Field exercise Major incident Simulation
The overall aim for medical responses to disasters and major incidents (MIs) is to achieve the best possible outcome for the most number of victims. However, in order to evaluate the outcome of a response you also need to assess the processes involved. To be able to do this there is a need for standardization and for sets of goals (bench-marks) reflecting what is to be considered as good and less good performance. These processes and goals must be developed in a way that allows them to be systematically studied, analyzed and with the possibility to compare results and experiences[1–3]. Several researchers have indicated the need for validated assessment methods performed in a scientifically genuine manner in order to measure the effectiveness of disaster medicine response[4–8].
In most healthcare systems the use of quality indicators are mandatory for determining standards and measuring quality. The challenge is to address indicators even in the field of disaster medicine. The use of measurable quality indicators to determine the level of performance against predetermined goals and objectives has been addressed in previous studies[9–11]. In these studies the quality indicators represent standards (times and content) of what is desirable performance of the management during MIs and disasters and were results from process- and concept modelling and/or opinion of an expert panel.
A few educational models for simulation of disaster management and MIs response are accepted and used in Sweden[12, 13]. However, there are a limited number of full-scale exercises (FSE) and systematic and validated evaluation tools are still in the process of being developed. Therefore the competence and effectiveness of the emergency medical services (EMS) and hospital performance seldom are assessed in a way that allows results to be compared. The question arises whether a standardized evaluation methodology would be applicable in a FSE to assess if and how the management is related to the patient outcome. The aim of this study was to demonstrate the feasibility of using a combination of performance and outcome indicators so that results can be compared in standardized full-scale exercises.
Description of the exercises
A quantitative evaluation method was applied in two multidisciplinary, FSE (major aircraft accident) in two regions in Sweden. The two exercises studied were conducted in October 2008 and in April 2010 at two different airports. The time interval, 18 months, between the exercises is due to the willingness of the two different airports and county councils to participate in this study. According to national regulations, each international airport is mandated to perform an exercise every other year, and there were, to our knowledge, no other airport exercises during this time interval.
The scenario was: an aircraft carrying passengers and crew that crashed during a landing attempt resulting in 99 respectively 100 victims (99 for the reason that one person, aimed to be figurant, felt sick on a late notice). In both exercises the participants, EMS and hospital personnel, were all familiar with the management structure (doctrine) at MI and disaster and expected to respond according to plans and procedures as stated in the national regulations issued by the Swedish National Board of Health and Welfare.
The victim’s conditions were expressed in physiological parameters. The basis for the victim bank has been developed on a national consensus among traumatologists. ETS is a simulation tool for education and training in disaster medicine which can be used both for table top exercises and field exercises as well as templates for evaluating results of performance and the treatment of the victims. A similar system is the Mass Casualty Simulation system (MACSIM) but with further developed injury card.
Exercise design and mock victim distribution
Instead of actually transporting the victims to the hospital by ambulances, ambulances were in both settings used only at the scene of the accident, and for a short transportation to the point where victims data were reported to the receiving hospitals. This data, regarding measures performed on each victim, was at this point collected by trained nurses and reported by telephone to the hospital ED. Also, the expected time of arrival (according to real transportation time) was reported. This meant, that a mock-victim in 2008 and magnet victim symbol in 2010, were presented with exactly the same data as the mock-victim that had participated on-scene, could be presented at the ED. By using this method, fewer ambulances were needed at the scene of the accident although the realism was the same. In a real MI probably 25–30 ambulances would rapidly respond to the requirement through the dispatch centre. For obvious reason access to ambulances is limited in FSE. However, for testing the concept merely two hospitals participated, one in each exercise, and received 17 victims, respectively. Both hospitals were emergency hospitals, and had a capacity of 302 and 309 beds, respectively. Available intensive care unit beds were 13 and 12 respectively, and operating rooms were nine and ten respectively. The time period in hospital, for each victim, included the time spent at the ED, x-ray, surgery (OR) and the estimated time for post operative- and intensive care. Performance- and patient outcome indicators were included in the evaluation of both exercises.
Sets of performance indicators and standards used for evaluation in full-scale disaster exercises
Prehospital command and control
Standard (time frame in min)
Putting on tabard*
First report to dispatch
Content on first report
Formulate guidelines for response
Establish contract with strategic level of command
Liaison with fire and police
Second report from scene
Content of second report
Verifying first report. Indicating first patient transport
Establish level of medical ambition
First patient evacuated
Information to media on scene
Regional command and control
Standard (time frame in min)
Declaring major incident
Deciding level of preparedness
Decision on additional resources to scene
Deciding on receiving hospitals
Establishing contact with incident officers at scene
Deciding on guidelines for referring hospitals
Brief information to media
Formulate general guidelines in accordance with guidelines from scene
Make sure there is information for definitive referral guidelines
Evaluated if capacity of own organisation is sufficient
Notify guidelines on referring hospitals
Hospital command and control
Standard (time frame in min)
Decide on level of preparedness
Formulate guidelines for hospital response
Give information about resources to strategic level
Ensuring that there is a medical officer in emergency operation
Estimate need of ICU beds
First information to staff
Estimate endurance of staff
Evaluate and report estimated shortage of own capacity
Evaluate influence on the daily hospital activities
Information plan for patients with postponed appointments and operations
Regional and Hospital staff procedure skills
Standard (time frame in min)
Assigning functions to all the staff members directly upon arrival
Directly on arrival
Placement in room according to function in staff
Designated telephone numbers
Introduction of newly arrived staff member
Utilization of available equipmenta
Maximum 8 min for “staff briefing”
Content of “staff briefing”b
Telephone discipline during “staff briefing”
Drawing and content of “staff schedule”
Summary after session, orally
Summary after session, written
The protocols were used for evaluation of performance at two different levels of medical management, regional level (in international literature often called strategic or gold level) and local level i.e. on-scene and in hospital (Figure1).
The performance indicators covered early decision- making (management skills), as well as staff performance. The level was scored by giving 0, 1, or 2 points where 0 point was given if the performance was not timely or adequate, 1 point if the result was somewhat correct and 2 points for a timely and by content, correct performance. Each template contained 11 indicators, thereby allowing for 22 points being maximum achievable level of result (Table1). Eleven out of 22 points where considered satisfactory.
The outcome indicator, patient outcome, was evaluated in terms of risk for preventable death or complication. Each victim was assigned to specific measures in order to be expected to have a favorable outcome, all according to templates that are included in the ETS. If these treatments were not performed or performed too late, according to the ETS stipulated timeframe, the patient will have a preventable complication or death. In ETS every specific treatment or transfer are performed according to a “real-time” approach.
All logistical and clinical data were registered at the scene of the accident and in hospital, regarding the care performed in real time, by participants who had been trained and instructed in ETS template.
All data regarding the performance indicators of the participants during the exercises were obtained as well as all data regarding indicators for patient outcome. Both exercises could therefore be compared regarding performance (process) as well as outcome (result) indicators.
Results based on templates of performance indicators, expressed in points
Command & control*
Staff procedure* skills
Command & Control*
Staff procedure* skills
Prehospital (Local level)
Hospital (Local level)
Patient outcome expressed as preventable complications and preventable death in two, full-scale disaster exercises
This study demonstrates the possibility of conducting standardized exercises with built-in evaluation methodology that can produce comparable results. The discussion in this paper focuses more on the methodology than the results from the exercises per se. A prerequisite for results from evaluations to be comparable is that the exercises are evaluated the same way, making it possible to obtain valid data to be recorded. In the present study we have demonstrated the possibility to apply both performance and outcome indicators in multidisciplinary, FSE. Indicators should be based on well-developed standards established and accepted by the organization to ensure comparability and reproducibility. The selected indicators were derived from a national concept and process modeling conducted by the National Board of Health and Welfare in Sweden. Idvall et al. describes this as ‘the key to good quality indicators’ and the only certain way of knowing what is good or less good by comparing performance against the standard. Furthermore, Idvall et al. clarifies that the most important indicator is the patient outcome, and it is a product influenced by all activities in an organization and should be measured together with other indicators. Standards (bench-marks) for evaluation methods that examine and describe the relation between the performance in disaster management and the patient outcome are yet to be described.
The evaluation model exposed several challenges faced in the initial decision-making procedures that were repeatedly observed and had major impact on patient outcome. For example, insufficient reporting and inconsistent coordination among responders that limited the ability to evacuate severely injured victims and resulted in comparatively high numbers of preventable death and complications. Despite good access to available ambulances in both exercises failure of rapid evacuation indicates that decisions made in this context were less than optimal which could possibly have been reflected in the patient outcome. By using the proposed model each agency will have increased possibilities to specifically identify what part of the command and control that needs to be improved. The fact that the difference between the prehospital command and control, as illustrated in Table2, did not show in patient outcome can however, not be explained and this is something that may have to be addressed using different research methods. At this stage, after only a few exercises, it is still too early to assess any relationship between performance and outcome. There is also a need to study if timely decisions or actions are more important than the content of the performance. In the present study the two prehospital organizations had different training concepts with regard to command and control, even though they had to follow the same regulations in performing.
Efficient disaster management requires information sharing, follow up and coordinated decision-making among involved agencies. Jufferman and Bierens studied five national disaster responses and noted that many shortages are repeated despite changes in protocols, legislation and organizations. For this reason there is a need for interagency training to improve operational decisions. Circumstances at the scene such as time -period, geography, weather conditions, reliable communication systems and security are other essential factors that may affect the response and therefore have to be considered in the assessment. This problem was recently described after a real aircraft incident where there were difficulties to evacuate the victims and this delayed the distribution of victims to hospitals.
The use of standardized education and training models, with built-in evaluation possibilities, allows comparisons and could lead to better knowledge and improve disaster medicine evidence in management system. In this study the first exercise was reproduced and the same type of evaluation, obtaining the same types of comparable results, was used in the second exercise. To be able to draw valid conclusions, the number of exercises performed must be higher. But this is also depending on number of indicators and methodology chosen. However, in our opinion, we must already start now to collect data as described in this study in order for the lessons learned in the processes not to be delayed further. The rational for this is the development of a robust design of the evaluation model used in the present study.
In the planning phase consideration was, in this study, taken both to each hospitals request of participation as well as their concerns regarding financing of staff which had impact upon the exercise design (Figure1). In our view, the difference in methodology regarding live figurants vs. virtual victims used at hospital, shows the models robustness in several settings. It is our belief that before engaging several hospitals in this type of structured exercises it should be prioritized to evaluate the first steps of the process. When this model becomes more accepted, additional units that can be trained and evaluated should be introduced.
Significance of this study for the future
To achieve greater acceptance methods with valid indicators need to be promoted by the health care providers. By using the same indicators in the evaluation of real incidents, in relation to available patient data, it would be possible to not only compare the performance indicators but also to validate the outcome indicators chosen in the exercises. The comparison with real incidents will give clues if we are educating and training the staff in the best way, and comparing results will provide information about possible improvements of the pedagogic educational simulation system used today. Knowledge that will émerge from this will no doubt, improve the possibility to conduct measurements in order to improve the disaster response before a real incident occurs. It will also reduce the time consumption for planning and after-action report writing.
Indicators are now gradually introduced in different regions in Sweden, so far, for evaluating the effect of training in order to measure quality in performance related to the management doctrine, reflected by regional regulations, supported by the Swedish National Board of Health and Welfare. Also in an international study results from using the same sets of indicators has recently been published.
In this study we have not considered structure indicators relating to staff, time and financial costs. Results from future studies with a systematic approach to the structure indicators will provide important information on how train and exercises should be conducted in the best way.
Competency and quality between different levels of management and organisations in this paper can depend on variation of personnel’s professional qualifications, education, experience and training in disaster medicine and might be a limitation. This applies also to the evaluators. Different evaluators may give different scoring for different processes. There is also a need to study if timely decisions or actions are more important than the content of the performance.The used simulation system may have limitations for evaluating patient outcome, but until studies from real incidents (where performance indicators as well as outcome indicators) are available we believe that this method will provide valid results.
The basis to use 11 points as acceptable level for measuring performance is based on results from training and examination sessions in command and control during several years of experience[9–11]. However, it could well be that any of the indicators could be more important than the others. Correlations studies may give answer to this. It may also well be that the indicators in the future must be weighted.
The figurant in the first exercise, that fell sick on late notice did not simulate a critical injured patient and was not intended to be sent to the participating hospital. Therefore we do not believe that this has any effect on the results.
Cost effectiveness, in regards to this subject, was not addressed in this study. This is due to the lack of systematic documentation of total costs for all involved staff including time for planning as well as the actual exercise.
Standardized FSE in different settings can be conducted and evaluated with performance indicators combined with outcome indicators enabling results from exercises to be compared. If exercises are performed in a standardized way, results may serve as a basis for lessons learned. Future use of the same concept using the combination of performance indicators and patient outcome indicators may demonstrate new and important evidence that could lead to new and better knowledge that also may be applied during real incidents.
This study was supported by grants from the Stockholm County Council. The authors thank the Air Navigation Services of Sweden and the County Administrative Board of Stockholm and the County Administrative Board of Jämtland for the opportunity to participate in the exercises that resulted in this study, as well as the support from different organizations that contributed to this project.
- Sundnes KO, Birnbaum ML: Health disaster management guidelines for evaluation and research in the Utstein style. Prehosp Disaster Med. 2003, 17 (Suppl 3): 113-127.Google Scholar
- Lennquist S: Promotion disaster medicine to a scientific discipline – A slow and painful but necessary process. Editorial. International Journal of Disaster Medicine. 2003, 2: 95-96.View ArticleGoogle Scholar
- Auf Der Heide E: The importance of evidence-based disaster planning. Ann Emerg Med. 2006, 47: 34-49. 10.1016/j.annemergmed.2005.05.009.View ArticlePubMedGoogle Scholar
- Kaji AH, Lewis RJ: Assessment of the reliability of the Johns Hopkins/Agency for Healthcare Research and Quality hospital disaster drill evaluation tool. Ann Emerg Med. 2008, 52: 204-210. 10.1016/j.annemergmed.2007.07.025.View ArticlePubMedGoogle Scholar
- Jenckes M, Catlett C, Hsu E: Development of evaluation modules for use in hospital drills. Am J Disaster Med. 2007, 2: 87-95.PubMedGoogle Scholar
- Juffermans J, Bierens JJ: Recurrent medical response problems during five recent disasters in the Netherlands. Prehosp Disaster Med. 2010, 25 (2): 137-138.View ArticleGoogle Scholar
- Savoia E, Biddinger PD, Burstein J, Stoto MA: Inter-agency communication and operations capabilities during a hospital functional exercise: reliability and validity of a measurement tool. Prehosp Disaster Med. 2010, 25 (1): 52-58.PubMedGoogle Scholar
- Legemaate G, Burkle F, Bierens J: The evaluation of research methods during disaster exercises: Applicability for improving disaster health management. Prehosp Disaster Med. 2012, 27 (1): 18-26. 10.1017/S1049023X11006789.View ArticlePubMedGoogle Scholar
- Rüter A, Örtenwall P, Vikström T: Performance indicators for major incident medical management – a possible tool for quality control?. International Journal of Disaster Medicine. 2004, 2: 52-55. 10.1080/15031430410023355.View ArticleGoogle Scholar
- Rüter A, Lundmark T, Ödmansson E, Vikström T: The development of a national doctrine for management of major incidents and disasters. Scand J Trauma Resusc Emerg Med. 2006, 14: 189-194.Google Scholar
- Rüter A, Örtenwall P, Vikström T: Staff procedure skills in management groups during exercises in disaster medicine. Prehosp Disaster Med. 2007, 22 (4): 318-321.PubMedGoogle Scholar
- Emergo Train System®. 2012, Available at websitehttp://www.emergotrain.com Accessed 29 May,
- Mass Casualty Simulation system (MACSIM) ®. 2012, Available at websitehttp://www.macsim.se. Accessed 29 May,
- Swedish transport agency. Regulation TFSF 2010:114. 2012, Available at websitehttp://www.transportstyrelsen.se Accessed 16 February,
- The Swedish National Board of Health and Welfare: Regulation SOSFS 2005. 2012, Available at websitehttp://www.socialstyrelsen.se/sosfs/2005-13 Accessed 24 August, 13,Google Scholar
- Idvall E, Rooke L, Hamrin E: Quality indicators in clinical nursing: a review of the literature. J Adv Nurs. 1997, 25: 6-17. 10.1046/j.1365-2648.1997.1997025006.x.View ArticlePubMedGoogle Scholar
- Nilsson H, Rüter A: Management of resources at major incidents and disasters in relation to patient outcome: a pilot study of an educational model. Eur J Emerg Med. 2008, 15: 162-165. 10.1097/MEJ.0b013e3282f4d14b.View ArticlePubMedGoogle Scholar
- Postma ILE, Winkelhagen J, Bloemers FW, Heetveld MJ, Bijlsma TS, Gosling JC: February 2009 Airplane Crash at Amsterdam Schiphol Airport: An Overview of Injuries and Patient Distribution. Prehosp Disaster Med. 2011, 26 (4): 299-304. 10.1017/S1049023X11006467.View ArticlePubMedGoogle Scholar
- France JM, Nichols D, Dong S: Increasing emergency medicine residents confidence in disaster management: use of an Emergency Department simulator and an expedited curriculum. Prehosp Disaster Med. 2012, 27 (1): 31-35. 10.1017/S1049023X11006807.View ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.