Skip to main content

Database quality assessment in research in paramedicine: a scoping review

Abstract

Background

Research in paramedicine faces challenges in developing research capacity, including access to high-quality data. A variety of unique factors in the paramedic work environment influence data quality. In other fields of healthcare, data quality assessment (DQA) frameworks provide common methods of quality assessment as well as standards of transparent reporting. No similar DQA frameworks exist for paramedicine, and practices related to DQA are sporadically reported. This scoping review aims to describe the range, extent, and nature of DQA practices within research in paramedicine.

Methods

This review followed a registered and published protocol. In consultation with a professional librarian, a search strategy was developed and applied to MEDLINE (National Library of Medicine), EMBASE (Elsevier), Scopus (Elsevier), and CINAHL (EBSCO) to identify studies published from 2011 through 2021 that assess paramedic data quality as a stated goal. Studies that reported quantitative results of DQA using data that relate primarily to the paramedic practice environment were included. Protocols, commentaries, and similar study types were excluded. Title/abstract screening was conducted by two reviewers; full-text screening was conducted by two, with a third participating to resolve disagreements. Data were extracted using a piloted data-charting form.

Results

Searching yielded 10,105 unique articles. After title and abstract screening, 199 remained for full-text review; 97 were included in the analysis. Included studies varied widely in many characteristics. Majorities were conducted in the United States (51%), assessed data containing between 100 and 9,999 records (61%), or assessed one of three topic areas: data, trauma, or out-of-hospital cardiac arrest (61%). All data-quality domains assessed could be grouped under 5 summary domains: completeness, linkage, accuracy, reliability, and representativeness.

Conclusions

There are few common standards in terms of variables, domains, methods, or quality thresholds for DQA in paramedic research. Terminology used to describe quality domains varied among included studies and frequently overlapped. The included studies showed no evidence of assessing some domains and emerging topics seen in other areas of healthcare. Research in paramedicine would benefit from a standardized framework for DQA that allows for local variation while establishing common methods, terminology, and reporting standards.

Background

ParamedicineFootnote 1 is increasingly recognized as a distinct healthcare profession with a unique body of knowledge [1,2,3]. Numerous studies and position papers have cited the need for research to develop quality benchmarks, investigate interventions, and evaluate outcomes within paramedic practice [2, 4,5,6,7,8]. While research in paramedicine continues to grow and evolve, the field faces key barriers to its ongoing development [6, 9, 10]. Among these, access to high-quality records of paramedic clinical practice (hereafter, paramedic data) has been identified as a critical barrier to linking to patient outcomes and researching paramedic care [2, 7, 8, 11].

The paramedic practice environment poses unique challenges to data collection [5, 10, 12,13,14]. Paramedic work is dynamic and complex, and takes place in uncontrolled and unpredictable environments, often subject to time and other pressures. Data collection is frequently delayed or shared among practitioners also providing care, resulting in potential data loss or inaccuracy [14, 15]. Records of paramedic care, historically paper-based, are transitioning to electronic platforms, but face continuing challenges to implementation in many jurisdictions [16, 17]. Paramedic services (as well as other emergency response agencies) typically organize documentation based on the incident, not the patient. Incident-based record keeping then requires linkage to subsequent files to assess outcomes for individual patients [18]. Data linkage using paramedic records varies in terms of success, not least in relation to the quality of initial data, and the linkage process can be susceptible to various forms of bias [11, 19].

Electronic health records in all contexts have benefits and limitations, but all require consistent ways of describing, assessing, and integrating information about data quality [20,21,22] . These needs apply equally to paramedic data. Despite challenges to data collection and analysis, research capacity in paramedicine will depend on consistent and valid methods of data collection, as well as a common language of quality assessment and standards of transparent reporting. Other healthcare professions have addressed these goals by developing conceptual tools for assessing data quality [23, 24]. Usually termed data quality assessment (DQA) frameworks, these tools provide both templates for data evaluation and guidance for future data collection. They establish baseline methodological standards, which in turn support the methodological quality of future research and the validity of results.

DQA frameworks cover a wide range of settings and purposes. Typically, they are organized by domains – distinct aspects of data that together make up a total picture of data quality in any particular field. The number of domains included in any framework can vary widely, and the terms used to describe similar concepts frequently overlap. Although as many as 49 different domains have been described in one practice area, [25]. frameworks typically include between one and eight domains, with key concepts such as completeness, accuracy, and timeliness appearing most frequently [25]. These and similar examples of domains from other healthcare disciplines have not been adopted in paramedicine. Although some position statements on data capture and reporting have been published, [26, 27]. no comprehensive framework dedicated to the paramedic work environment has been developed, and the adaptability of existing ones to the unique circumstances of paramedicine has not been determined.

As paramedic research continues to evolve, studies that rely on records of paramedic clinical practice will require a common language and standard of data assessment to support methodological rigor. In the absence of a paramedic-specific DQA framework, the landscape of data-quality practices remains uncharted. No prior reviews have collected information on this topic, and reporting of DQA practices within paramedicine remains sporadic. Currently, we lack a comprehensive view of what data are assessed, methods for doing so, and the ensuing results. Recognizing a need to understand the extent to which paramedicine researchers have embedded information about data quality into their research products, this manuscript describes the results of a scoping review that was conducted to describe the range, extent, and nature of DQA practices reported in paramedicine research.

Methods

A protocol of the methods has been previously registered with the Open Science Framework (https://doi.org/10.17605/OSF.IO/Z287T) and published [28]. Reporting follows the guidelines of the Preferred Reporting Items in Systematic Reviews and Meta-Analyses extension for scoping reviews (PRISMA-ScR) [29].

Aim

This scoping review asks, what are the range, extent and nature of DQA practices in paramedic research? It aims to document these characteristics to support ongoing development of methodological standards in research in paramedicine.

Search strategy

With the support of a professional librarian and in accordance with established methods, a search was constructed to reflect the population, context, and concept of the research question [30, 31]. Paramedic research studies that assessed data quality as a major goal and reported quantitative DQA results from the paramedic practice environment were included. This environment included urban, rural, remote, and military settings, but excluded special circumstances (disaster and mass-casualty situations). Studies were excluded if they were protocols, commentaries, case studies, interviews, simulations, or used experimental data-processing techniques. Studies that were not primarily concerned with paramedic data, or studies that evaluated databases that incidentally included paramedic information, were also excluded. No restrictions were placed on language. After iterative refinement of search terms and pilot testing of date ranges, the search was limited to 2011–2021 to balance comprehensiveness with recency. The search was applied to the following databases: MEDLINE (National Library of Medicine), Embase (Elsevier), Scopus (Elsevier) and CINAHL (EBSCO). The searches as applied are available in “Additional file 1” and reflect the specific terminology, logical combinations, and formatting of each database. Generic keywords and subject headings are listed for illustration: “emergency medical services”, “emergency medical technicians”, “ambulance”, “paramedic”, “paramed*”, “prehospital”, “first respond*”, “emergency services”, “quality improvement”, “quality assurance”, “health care”, “information storage”, “information retrieval”, “data collection”, “medical records”, “electronic health records”, “health records, personal”, “medical record linkage”, “medical records systems, computerized”, “patient regist*”, “data quality”, “electronic medical record”, “record linkage”, “paramedic record”.

Screening

Search results were imported into a data-management program (Covidence, Veritas Health Innovation, Melbourne, Australia). After duplicate citations were removed, all authors participated in title and abstract screening of 250 records to practice and discuss the application of inclusion criteria. All remaining records were independently screened by at least two reviewers, and any record selected by any reviewer was retained for full-text screening. Full-text records were assessed independently by two reviewers (NM, RP); differences were resolved with discussion, including the third reviewer (NL).

Data extraction

Data were extracted using a custom-designed data-extraction form (“Additional file 2”). This form included 13 fields grouped according to the range, extent, and nature of DQA practices. Range was defined by geographic location, year of publication, study purpose, and topic (whether a clinical area, population, or specific circumstance). Extent was documented by the level, breadth, and number of records assessed. Within extent, level refers to the organizational area of the primary data and includes five categories: local (municipal or small area); regional (such as a regional health authority); sub-national jurisdiction (state/province/county); national; and international. Breadth contains two components: the number of services included and the number of linkages between paramedic data and other types of databases. The nature of the DQA was summarized by the specific variables or fields assessed, the methods of assessment, results, the domain of data quality being assessed, and the presence of any quality threshold. In accordance with guidance on scoping reviews, critical appraisal was not performed [29]. Data were extracted iteratively, and key information was summarized for reporting, either quantitatively or with representative examples.

Protocol amendments

These methods correspond to the registered study protocol with following exceptions. Each change was based on the consensus of reviewers during data extraction. (1) The duration of data assessed was replaced by number of records assessed. (2) A field to record any quality threshold or summary rating of data quality was added. (3) The study protocol called for data-quality domains to be recorded both as identified by the study, and according to an existing framework used by the Canadian Institute for Health Information (CIHI) [32]. Since the included studies used a wide variety of descriptions to identify assessment domains, terms covering similar concepts were grouped under the domain name that was most applicable or appeared most frequently (with all alternative terms listed). As well, categorization of assessment domains under the CIHI framework yielded only two categories (Accuracy & Reliability and Comparability & Coherence). As these results added little interpretive value, they have not been reported.

Results

Database searching identified 10,105 unique articles (Fig. 1). After title and abstract screening, 199 remained for full-text review. Of these, four were in languages other than English (one each of German, Spanish, Russian, and Persian [Farsi]); these were professionally translated for further assessment. Among all articles selected for full-text assessment, 102 were excluded for reasons cited. Additional duplicates (n = 18) identified at this stage included abstracts for which full articles using the same data and substantially similar results were also present. Ninety-seven articles were included in the analysis.

Fig. 1
figure 1

PRISMA-ScR flow diagram of study selection

Study characteristics

Table 1 lists the main characteristics of included studies, as well as selected extracted data. (“Additional file 3” lists full citations of all included studies.)

Table 1 Characteristics of included studies [inserted at end of document]

Range of included articles

Among the 97 included articles, 39 (40%) were published from 2019 to 2021, with the remainder spread relatively evening across the preceding years. Forty-nine studies (51%) were conducted in the United States (US); Australia (n = 10), the United Kingdom (n = 8), and Canada (n = 6) were the next most frequent locations. Figure 2 lists all countries, as well as the breakdown of US States, where applicable. Abstracts (as well as one letter) accounted for 27 (28%) included items; the remainder (n = 70, 72%) were full articles. Included articles studied diverse topics spanning clinical areas, populations, and specific situations. Studies were coded to allow for multiple subject areas; Fig. 3 illustrates the number of studies per topic out of all mentioned (n = 111). Topics related to data linkage or the data management without reference to a clinical area (labelled, “Data”) were the most frequent area of study (n = 27, 24%). The next most common topic was trauma (n = 21, 19%), followed by out-of-hospital cardiac arrest (OHCA) (n = 20, 18%). These three areas made up the majority (68/111, 61%) of all areas studied.

Fig. 2
figure 2

Geographic location of data quality assessment studies in research in paramedicine (n = 97), listing the number of studies by country (main panel), and by State (or national / military) among studies from the United States (inset)

Fig. 3
figure 3

Topic (clinical area / population / situation) of data quality assessment studies in research in paramedicine, listing the number of areas (total = 111) among all studies (n = 97)

Extent of included studies

Figure 4 displays the extent of included studies according to the identified sub-categories. The level at which studies assessed data was spread relatively evenly among local (n = 28, 29%), regional (n = 25, 26%), and state/province/county (n = 28, 29%) (Fig. 4A). The majority of studies (n = 51, 53%) assessed data belonging to one paramedic or prehospital agency (Fig. 4B). In terms of linkage, 39 (40%) studies did not link paramedic or prehospital data to any other sources, whereas forty-four (45%) linked to a single type of database (whether hospital, emergency department, or other related source), and 14 (14%) linked to multiple databases of different kinds (Fig. 4C). The majority of studies reviewed between 100 and 9,999 records (n = 59, 61%), with only 6 (6%) reviewing fewer than 100 and 4 (4%) reviewing more than 1 million (Fig. 4D). Considering combinations of the level of data assessed (Fig. 4BA), the number of services (Fig. 4B), and number of linkages (Fig. 4C), the three largest exclusive groups of characteristics involved: a local, single service linked to a single type of database (13/97); state-level data, represented by 10 or more services, linked to a single type of database (13/97); and a local, single service with no linkage (12/97).

Fig. 4
figure 4

The extent of data quality assessment studies in research in paramedicine, measured by A the level of data assessed, B the number of services included, C the number of types of linkages to other databases, and D, the number of records assessed. Each chart includes all studies (n = 97)

Nature of included studies

Table 2 summarizes the domain names and explanations derived from how the studies described their assessment. It also includes any quality measures applied by included studies, grouped by domain. As listed in Table 1, some studies assessed multiple areas, yielding 126 instances of an assessed domain.

Table 2 Summary of data quality assessment domains in studies on research in paramedicine

As incidental findings, one study adapted a DQA framework from public-health surveillance and applied some domains to its prehospital data [33]. Similarly, two studies applied a reporting guideline specific to the methodology of database linkage [34, 35]. No other DQA reporting guidelines were noted.

The DQA domains of the included studies are summarized below, with examples of representative and unique studies.

Completeness

The included studies used a variety of terms that can be summarized as assessments of completeness (Table 2). Based on the practices described, completeness measured how often a variable was present when expected or required. It was usually expressed as a proportion or percent of all potential entries. Depending on the purpose of the study or the nature of the results, this was often represented as its complement, missingness. This domain appeared most frequently, and was present in 57 studies, accounting for 45% of all domains documented (n = 126).

Among included studies, completeness frequently measured the variables deemed most important to each study’s purpose. For example, Abir et. al. found only five of 18 key variables were present in over 90% of cases [36]. Other large studies provided similar ranges, [37] although some report wide discrepancies among individual services in aggregated data [38, 39] . Certain categories, such as mechanism of injury, frequently showed relatively low values [40]; emergency department (ED) disposition, where reported, was negligibly complete in paramedic databases (cited in one study at less than 5% [41]). Additional contrasts in the completeness of basic variables can be seen between different settings, such as helicopter emergency medical services (EMS) agencies and the military, where completion rates were consistently high and low, respectively [42,43,44,45].

Linkage

Thirty-four studies (representing 27% of all domains) assessed how well paramedic or prehospital data could be linked to other sources of information. Included studies detailed a range of techniques for linkage, broadly divided between deterministic and probabilistic approaches, occasionally supplemented by manual review for confirmation or optimization [46, 47].

Overall rates of linkage varied among the included studies. In one case, an optimized iterative deterministic approach yielded 97% success in linking records of EMS patients transported to an ED, with no false positives [19]. Other studies found similar results with a variety of optimization strategies [35, 48,49,50]. Contrasting results appeared in several studies linking trauma patients to hospital outcomes, ranging between 15 and 88%, and 49–60% specifically for ground transport [51, 52]. Others examining OHCA (34% [53]) and stroke (26% [54]) marked the lowest reported rates within those clinical areas.

Accuracy

Among a range of terms used by the included studies to describe similar concepts, accuracy summarizes practices that evaluated the extent to which a variable recorded what it was designed to measure. When it was assessed, accuracy was measured against a reference thought to be valid or true, sometimes referred to as a gold standard. It was expressed in terms of proportions, percents, and diagnostic test statistics (sensitivity, specificity, positive predictive value, negative predictive value). Evaluations of accuracy were present in 14 studies, accounting for 11% of all domains assessed.

Several topics featured multiple studies assessing accuracy, including OHCA, [55,56,57] vital-sign documentation, [58, 59] [37, 60] and patient history. [61, 62] Within OHCA, three studies evaluated the accuracy of documented events and timepoints in the paramedic record in comparison to video or audio recordings or data from a defibrillator/monitor – in each case, a source thought to represent a gold standard. All showed discrepancies between written and recorded data, including, for example, detection of return of spontaneous circulation and re-arrest, [56] the rate and depth of chest compressions, [57] and total CPR time and total adrenaline dose [55].

Reliability

In addition to assessing accuracy, some included studies also measured the extent to which measurements and documentation were consistent or how much variety would appear over repeated measures. This was most commonly described as reliability, although agreement, consistency, and other terms were used for the same domain (Table 2). In contrast to measures of accuracy and validity, reliability assessed agreement between two values without assuming that one represented a reference standard. In place of statistics that measure proximity to a value, reliability was expressed in terms of correlation, kappa, intraclass correlation coefficient, difference, differences in proportions, and unique measures derived by individual studies [40, 63]. Ten studies presented quantitative data falling under these headings, representing 8% of domains evaluated.

Whereas several studies evaluated the accuracy of prehospital documentation of patient medical history in comparison to hospital records, some analyses assessed the same information in terms of agreement. For example, Coventry et. al. found that paramedic and hospital documentation showed high agreement in recording the presence of chest pain among patients with myocardial infarctions (adjusted kappa, k = 0.87).[62].

When applied specifically to the spread or clustering of measurements, reliability is commonly termed precision. (This was also referred to as granularity in the case of time stamps [40].) In assessing documented event times in OCHA in comparison to audio recordings, Frisch et al. found wide variability in reported times – imprecision that they argue should be accounted for in future analyses [64]. Precision has also been assessed in terms of how many different ways variables are recorded, both within and across datasets. Staff et al. examined whether vital signs in trauma calls were recorded as exact numbers, categories, or inferred from free-text [65] . Common variables recorded differently both within and across datasets were cited in other instances, including vital signs, [66] chief-complaint coding among different services, [38], and even ostensibly standardized variables in OHCA reporting [67].

Representativeness

Studies that examined the extent to which data corresponded to reference populations or to the degree to which data could be applied outside of the study group assessed representativeness (or generalizability, bias, concordance, or external validity). Among included articles, representativeness was assessed most often by comparisons of proportions, although correlation, agreement, and unique statistics were also used [68] Eleven studies included assessments of representativeness, accounting for 9% of domains.

Studies in paramedic research used a variety of approaches to defining a reference group. Mann et. al. assessed the generalizability of the 2012 National Emergency Medical Services Information System (NEMSIS, a national database of EMS information in the United States) by comparing patient ages as documented in NEMSIS to the ages of all ED arrivals documented in other sources (the results showed high correlation, r > 0.9).[41] Lerner et al. (2021) evaluated a pediatric-specific database with the complete cohort of all pediatric records in NEMSIS and found meaningful differences in patient race and chief complaints between the two groups [69].

Other linkage studies assessed their results for bias by examining differences between linked and unlinked cohorts. Within particular clinical areas, such as stroke and OHCA, indications of bias between linked and unlinked groups were seen within topic-related factors, such as age, event location, bystander CPR, or return of spontaneous circulation [57, 68, 70, 71] . Another study linking paramedic and hospital records tracked the degree to which an optimized strategy for case matching mitigated bias found in a standard approach [19].

Quality thresholds

Also included as an attribute of the nature of studies on research in paramedicine, the concept of quality thresholds appeared sporadically among the included studies. Despite these mentions, there are no established guides, thresholds, or systems for defining what constitutes quality data or determining what is high versus low quality. Many studies discussed the relevance of their results, finding them to be feasible or applicable (or not) in individual cases. Few studies reported applying any quality threshold; those that did are described below.

The domain of completeness offered clear and simple options for testing. In one study, completeness of less than 90% (or greater than 10% missingness) was judged to be low quality [36]. Others used similar thresholds [45, 72,73,74]. Within studies examining linkage of paramedic data with other sources, papers sometimes applied a pre-specified probability cut-off that determined a match or non-match, with those at or near the threshold value being selected for manual review. This was often listed as a probability at or straddling 0.9, [39, 51, 75] although 0.5 was also used, [49] as were levels that varied within the study according to patient block [53]. Other studies used ratings of match quality depending on the number or type of variables that established the link [70, 76, 77].

Within the domain of representativeness, few studies worked with a standard beyond reporting different proportions among their study groups. In contrast, Lerner et al. (2021) described applying a threshold of plus or minus 5% as indicating a meaningful difference between their sample and reference populations [69]. Oostema et al. used an absolute standardized difference, defined as the average difference of each variable as a percent of its standard deviation, with values greater than 0.1 indicating a significant difference [68].

Discussion

The studies identified in this scoping review make up a sample of DQA practices in research in paramedicine. This collection varies widely across many factors, including country of origin, topic assessed, and purpose. In many cases, the DQA component appeared to be ad hoc, reflecting the unique methodological requirements of individual studies and often presented as an accompanying abstract or article to an investigation with some other aim. Where evident, accumulated expertise developed over the course of multiple studies appeared within related research groups, rather than across researchers within the profession [38, 39, 51, 69, 75, 78]. The variety in purpose was also related to the extent of included studies. Many featured a single service examining its own data or linking to a single hospital or ED. In contrast, there were several examples of regional, state, or national-level data being integrated with multiple external databases with high levels of linkage success, either for specific research purposes or routine outcome evaluation [19, 34, 51, 70, 79]. These examples demonstrate progress in overcoming oft-noted barriers to data linkage and outcome evaluation [2, 11].

While the results of individual studies were too variable to draw specific conclusions about paramedic data quality, some generalizations about the nature of DQA practices emerged. Many authors emphasized the central priority of data completeness in paramedic research. Although a relatively simple concept, completeness was seen as a foundation supporting other domains – not only as a baseline indicator of data quality, but also as an essential precursor to linkage with other databases and outcome evaluation. Apart from this consensus, there were few (if any) common standards in terms of variables, domains, methods, or quality thresholds for DQA in paramedic research. A DQA framework was mentioned by only one included study (which was only partially applicable to prehospital data) [33]. Relatedly, although a reporting guideline exists for data-linkage methodology, it was referenced by only two papers out of 34 reporting linkage results [34, 35]. As in existing frameworks, the terminology and application of some DQA practices among the included studies featured variable or inconsistent meanings. This variety highlights the need for clear and consistent terminology to support transparency and comparability in DQA practices.

These characteristics of DQA practices point to both the relative youth of research in paramedicine and continuing barriers to research and data collection in the field in general [9, 10]. These barriers are discussed at length by several articles, and key findings reiterate the difficulty of collecting high-quality information (especially accurate demographic details) in the clinical environment [36]. Incomplete or unreliable data limit the effectiveness of deterministic linkage, [52] and inconsistent reporting of common data fields complicates studies using aggregated data. Problems with varied reporting were observed among a range of topic areas, including defining trauma calls, [52] coding chief complaints, [38] reporting OHCA variables, [67] and even the ages defining pediatric patients, which ranged among included studies from 0–4 to 0–21 [69, 80,81,82]. These inconsistencies overlapped with observed difficulties in both coding and extracting information from free-text data [38, 83]. Data linkage is complex, labour-intensive, and expensive, presenting challenges to single services aiming to assess outcomes [70]. Finally, the need to establish data-sharing agreements between organizations that collaborate in patient care constitutes another barrier to outcome assessment [52].

Although challenges to data quality were widely described, fewer studies remarked on strategies for assurance or improvement. Among those that did, Mann et al. referenced a system of over 300 logic rules that assess data quality prior to acceptance in NEMSIS [41]. (While logic rules are commonly applied, one paper observed the unintended consequence of a “bare minimum effect” when forcing documentation [36].) Several studies showed improved documentation after focused and dedicated internal training [83,84,85]. Others noted improved outcomes with the introduction of electronic forms or databases [86,87,88]. Methodological refinements in case ascertainment, handling missing data, and linkage strategies were shown to maximize data quality [39, 51, 75].

Beyond the barriers and strategies for improvement for data quality in general, the included studies speak to DQA practices both by what they describe and by what they do not. Existing DQA frameworks feature domains and sub-domains that did not appear among the reviewed studies, including broad categories such as accessibility, clarity, and timeliness [32, 89] These domains (as well as synonyms and related concepts such as punctuality, relevance, interpretability, comparability) largely reflect the needs of researchers in gaining access to databases, the timing of data updates and their availability, and supporting documentation [14, 32, 89] (Occasional studies have assessed the timeliness of the availability of the paramedic record for clinical use, but not for research purposes [90, 91]) The absence of these domains might be seen also to reflect the relative youth of paramedic databases and remaining barriers to incorporating them into administrative repositories.

Considering DQA along a spectrum of progress highlights current issues and how they might be incorporated into the next iterations of guidelines for paramedic data. As an example, recent research has foregrounded comprehensive reporting of sex and gender and the inadequacy of binary options to encompass multi-dimensional concepts [92]. Sex and gender reporting has been evaluated in other electronic health datasets, [93] and the implications of its limitations on record linkage were considered in one included study [94]. In a similar approach, the COVID-19 pandemic has accelerated efforts to examine outcomes through the lens of data equity, [95] and current guidance on race-based data collection emphasizes a range of system features that might be considered preconditions for the responsible collection and use of this information [96]. Finally, knowledge of patient and public perspectives related to individual data items translates to awareness of public involvement and engagement in data management as a precursor to maintaining social license for healthcare research [97, 98]. While concepts such as data ownership, stewardship, and patient and public involvement do not address quality in the same way as ensuring birthdates are collected accurately, they undoubtedly have a role in how data is collected, accessed, and used – and therefore a role in ensuring the most basic definition of data quality, that it is fit for use [32].

Limitations

While comprehensive, the search strategy employed in this review was necessarily exploratory. It was iteratively refined to ensure capture of known key papers, but the possibility of missed articles cannot be excluded, and the resulting sample could be biased in unknown ways. Extreme heterogeneity among included studies presents difficulty in summarizing results. Alternative ways of categorizing and interpreting the data are possible, and the approach taken here potentially reflects biases among the reviewers. Although small, the review team included members with clinical, administrative, and methodological expertise in order to guard against this possibility. In keeping with the nature of scoping reviews, these results should be taken as a preliminary description of the field of study, with analyses and conclusions interpreted cautiously.

Conclusions

This scoping review of DQA practices in paramedic research summarizes diverse approaches applied largely as needed in individual studies or research programs. Although there are many opportunities and options for improving the quality of data collected at the source, the results of this review point to additional considerations for practice leaders. Databases of health information collected by paramedics would benefit from a standardized framework for DQA that allows for local variation while establishing common methods, terminology, and reporting standards. As paramedic research continues to grow, there is an opportunity to integrate progressive concepts of availability, stewardship, and ownership into emerging constructs.

Availability of data and materials

All source material is available in the public domain. Translations of included articles not in English are available upon reasonable request.

Notes

  1. * Other common descriptions of paramedicine and paramedics reflect distinct aspects of the profession, whether in terms of the provider (emergency medical technicians, emergency responder, allied health provider) or the setting (prehospital care, out-of-hospital care, remote and retrieval medicine). No single description encompasses the international range of past and emerging practice models. Acknowledging this limitation, this review will use “paramedic” and “paramedicine” as generally inclusive of varied roles and settings.

Abbreviations

DQA:

Data quality assessment

PRISMA-ScR:

Preferred Reporting Items in Systematic Reviews and Meta-analyses extension for scoping reviews

CIHI:

Canadian Institute for Health Information

OHCA:

Out-of-hospital cardiac arrest

ED:

Emergency department

EMS:

Emergency medical services

NEMSIS:

National Emergency Medical Services Information System

References

  1. Williams B, Beovich B, Olaussen A. The definition of paramedicine: an international Delphi study. J Multidiscip Healthc. 2021;14:3561–70.

    Article  PubMed  PubMed Central  Google Scholar 

  2. Tavares W, Allana A, Beaune L, Weiss D, Blanchard I. Principles to guide the future of paramedicine in Canada. Prehosp Emerg Care. 2022;26(5):728–38.

    Article  CAS  PubMed  Google Scholar 

  3. Reed B, Cowin L, O’Meara P, Wilson I. Professionalism and professionalisation in the discipline of paramedicine. Australas J Paramed. 2019;16:1.

    Article  Google Scholar 

  4. Jensen JL, Bigham BL, Blanchard IE, Dainty KN, Socha D, Carter A, et al. The Canadian national EMS research agenda: a mixed methods consensus study. CJEM. 2013;15(2):73–82.

    Article  PubMed  Google Scholar 

  5. Carter H, Thompson J. Defining the paramedic process. Aust J Prim Health. 2015;21(1):22–6.

    Article  PubMed  Google Scholar 

  6. Maurin Söderholm H, Andersson H, Andersson Hagiwara M, Backlund P, Bergman J, Lundberg L, et al. Research challenges in prehospital care: the need for a simulation-based prehospital research laboratory. Adv Simul. 2019;4:3.

    Article  Google Scholar 

  7. Vloet LCM, Hesselink G, Berben SAA, Hoogeveen M, Rood PJT, Ebben RHA. The updated national research agenda 2021–2026 for prehospital emergency medical services in the Netherlands: a Delphi study. Scand J Trauma Resusc Emerg Med. 2021;29(1):162.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Newgard CD, Fu R, Malveau S, Rea T, Griffiths DE, Bulger E, et al. Out-of-hospital research in the era of electronic health records. Prehosp Emerg Care. 2018;22(5):539–50.

    Article  PubMed  Google Scholar 

  9. Olaussen A, Beovich B, Williams B. Top 100 cited paramedicine papers: a bibliometric study. Emerg Med Australas. 2021;33(6):975–82.

    Article  PubMed  Google Scholar 

  10. Carter AJE, Jensen JL, Petrie DA, Greene J, Travers A, Goldstein JP, et al. State of the evidence for emergency medical services (EMS) care: the evolution and current methodology of the prehospital evidence-based practice (PEP) program. Healthc Policy. 2018;14(1):57–70.

    PubMed  PubMed Central  Google Scholar 

  11. Cone DC, Irvine KA, Middleton PM. The methodology of the Australian prehospital outcomes study of longitudinal epidemiology (APOStLE) project. Prehosp Emerg Care. 2012;16(4):505–12.

    Article  PubMed  Google Scholar 

  12. Denecke K, Meier L, Bauer JG, Bender M, Lueg C. Information capturing in pre-hospital emergency medical settings (EMS). Stud Health Technol Inform. 2020;270:613–7.

    PubMed  Google Scholar 

  13. Reichard AA, Marsh SM, Moore PH. Fatal and nonfatal injuries among emergency medical technicians and paramedics. Prehosp Emerg Care. 2011;15(4):511–7.

    Article  PubMed  Google Scholar 

  14. Mashoufi M, Ayatollahi H, Khorasani-Zavareh D. A review of data quality assessment in emergency medical services. Open Med Inform J. 2018;12:19–32.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Denecke K, Meier L, Bauer JG, Bender M, Lueg C. Information capturing in pre-hospital emergency medical settings (EMS). Stud Health Technol Inform. 2020;16(270):613–7.

    Google Scholar 

  16. Landman AB, Lee CH, Sasson C, Van Gelder CM, Curry LA. Prehospital electronic patient care report systems: early experiences from emergency medical services agency leaders. PLoS ONE. 2012;7(3):e32692.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  17. Porter A, Badshah A, Black S, Fitzpatrick D, Harris-Mayes R, Islam S, et al. Health Services and Delivery Research. Electronic health records in ambulances: the ERA multiple-methods study. Southampton (UK): NIHR Journals Library.

  18. Cox S, Martin R, Somaia P, Smith K. The development of a data-matching algorithm to define the “case patient.” Aust Health Rev. 2013;37(1):54–9.

    Article  PubMed  Google Scholar 

  19. Blanchard IE, Williamson TS, Ronksley P, Hagel B, Niven D, Dean S, Shah MN, Lang ES, Doig CJ. Linkage of emergency medical services and hospital data: a necessary precursor to improve understanding of outcomes of prehospital care. Prehosp Emerg Care. 2022;26(6):801–10.

    Article  CAS  PubMed  Google Scholar 

  20. Hersh WR, Weiner MG, Embi PJ, Logan JR, Payne PR, Bernstam EV, et al. Caveats for the use of operational electronic health record data in comparative effectiveness research. Med Care. 2013;51(8 Suppl 3):S30–7.

    Article  PubMed  PubMed Central  Google Scholar 

  21. Kahn MG, Callahan TJ, Barnard J, Bauck AE, Brown J, Davidson BN, et al. A harmonized data quality assessment terminology and framework for the secondary use of electronic health record data. EGEMS. 2016;4(1):1244.

    Article  PubMed  PubMed Central  Google Scholar 

  22. Verheij RA, Curcin V, Delaney BC, McGilchrist MM. Possible sources of bias in primary care electronic health record data use and reuse. J Med Internet Res. 2018;20(5):e185.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Chan KS, Fowles JB, Weiner JP. Review: electronic health records and the reliability and validity of quality measures: a review of the literature. Med Care Res Rev. 2010;67(5):503–27.

    Article  PubMed  Google Scholar 

  24. Weiskopf NG, Weng C. Methods and dimensions of electronic health record data quality assessment: enabling reuse for clinical research. J Am Med Inform Assoc. 2013;20(1):144–51.

    Article  PubMed  PubMed Central  Google Scholar 

  25. Chen H, Hailey D, Wang N, Yu P. A review of data quality assessment methods for public health information systems. Int J Environ Res Public Health. 2014;11(5):5170–207.

    Article  PubMed  PubMed Central  Google Scholar 

  26. Gunderson MR, Florin A, Price M, Reed J. NEMSMA position statement and white paper: process and outcomes data sharing between EMS and receiving hospitals. Prehosp Emerg Care. 2021;25(2):307–13.

    Article  PubMed  Google Scholar 

  27. Canadian standards association (CSA Group). Functional requirements and core data set for a Canadian paramedic information system (CSA Z1635:22). Toronto, Ontario: CSA Group; 2022.

  28. McDonald N, Kriellaars D, Doupe M, Giesbrecht G, Pryce RT. Database quality assessment in research in paramedicine: a scoping review protocol. BMJ Open. 2022;12(7):e063372.

    Article  PubMed  PubMed Central  Google Scholar 

  29. Tricco AC, Lillie E, Zarin W, O’Brien KK, Colquhoun H, Levac D, et al. PRISMA extension for scoping reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. 2018;169(7):467–73.

    Article  PubMed  Google Scholar 

  30. Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2005;8(1):19–32.

    Article  Google Scholar 

  31. Peters MD, Godfrey C, McInerney P, Munn Z, Tricco AC, Khalil H. Scoping reviews. Joanna Briggs Inst Rev Man. 2017;2015:1–24.

    Google Scholar 

  32. CIHI. CIHI's information quality framework. Ottawa: Canadian Institute for Health Information; 2017.

  33. Goldstick J, Ballesteros A, Flannagan C, Roche J, Schmidt C, Cunningham RM. Michigan system for opioid overdose surveillance. Injury Prev. 2021;27(5):500–5.

    Article  Google Scholar 

  34. Carroll T, Muecke S, Simpson J, Irvine K, Jenkins A. Quantification of NSW ambulance record linkages with multiple external datasets. Prehosp Emerg Care. 2015;19(4):504–15.

    Article  PubMed  Google Scholar 

  35. Chikani V, Blust R, Vossbrink A, Wightman P, Bissell S, Graw J, et al. Improving the continuum of care by bridging the gap between prehospital and hospital discharge data through stepwise deterministic linkage. Prehosp Emerg Care. 2020;24(1):1–7.

    Article  PubMed  Google Scholar 

  36. Abir M, Taymour RK, Goldstick JE, Malsberger R, Forman J, Hammond S, et al. Data missingness in the Michigan NEMSIS (MI-EMSIS) dataset: a mixed-methods study. Int J Emerg Med. 2021;14(1):22.

    Article  PubMed  PubMed Central  Google Scholar 

  37. Alstrup K, Petersen JAK, Knudsen L, Barfod C, Moller TP, Rognas L. The Danish helicopter emergency medical service database: high quality data with great potential. Scand J Trauma Resuscitation Emerg Med. 2019;27(1):38.

    Article  Google Scholar 

  38. Lerner EB, Dayan PS, Brown K, Fuchs S, Leonard J, Borgialli D, et al. Characteristics of the pediatric patients treated by the pediatric emergency care applied research network’s affiliated EMS agencies. Prehosp Emerg Care. 2014;18(1):52–9.

    Article  PubMed  Google Scholar 

  39. Newgard C, Malveau S, Staudenmayer K, Wang NE, Hsia RY, Mann NC, et al. Evaluating the use of existing data sources, probabilistic linkage, and multiple imputation to build population-based injury databases across phases of trauma care. Acad Emerg Med. 2012;19(4):469–80.

    Article  PubMed  PubMed Central  Google Scholar 

  40. Andrews R, Wynn MT, Ter Hofstede AHM, Vallmuur K, Bosley E, Rashford S, et al. Leveraging data quality to better prepare for process mining: an approach illustrated through analysing road trauma pre-hospital retrieval and transport processes in Queensland. Int J Environ Res Public Health. 2019;16(7):1138.

    Article  PubMed  PubMed Central  Google Scholar 

  41. Mann NC, Kane L, Dai M, Jacobson K. Description of the 2012 NEMSIS public-release research dataset. Prehosp Emerg Care. 2015;19(2):232–40.

    Article  PubMed  Google Scholar 

  42. Tonsager K, Rehn M, Ringdal KG, Lossius HM, Virkkunen I, Osteras O, et al. Collecting core data in physician-staffed pre-hospital helicopter emergency medical services using a consensus-based template: international multicentre feasibility study in Finland and Norway. BMC Health Serv Res. 2019;19(1):151.

    Article  PubMed  PubMed Central  Google Scholar 

  43. Saviluoto A, Björkman J, Olkinuora A, Virkkunen I, Kirves H, Setälä P, et al. The first seven years of nationally organized helicopter emergency medical services in Finland—the data from quality registry. Scand J Trauma Resusc Emerg Med. 2020;28(1):46.

    Article  PubMed  PubMed Central  Google Scholar 

  44. Therien SP, Nesbitt ME, Duran-Stanton AM, Gerhardt RT. Prehospital medical documentation in the Joint theater trauma registry: a retrospective study. J Trauma. 2011;71(1 Suppl):S103–8.

    PubMed  Google Scholar 

  45. Robinson JB, Smith MP, Gross KR, Sauer SW, Geracci JJ, Day CD, et al. Battlefield documentation of tactical combat casualty care in Afghanistan. US Army Med Dept J. 2016;2–16:87–94.

    Google Scholar 

  46. Redfield C, Tlimat A, Halpern Y, Schoenfeld DW, Ullman E, Sontag DA, et al. Derivation and validation of a machine learning record linkage algorithm between emergency medical services and the emergency department. J Am Med Inform Assoc. 2020;27(1):147–53.

    Article  PubMed  Google Scholar 

  47. Rajagopal S, Booth SJ, Brown TP, Ji C, Hawkes C, Siriwardena AN, et al. Data quality and 30-day survival for out-of-hospital cardiac arrest in the UK out-of-hospital cardiac arrest registry: a data linkage study. BMJ Open. 2017;7(11):e017784.

    Article  PubMed  PubMed Central  Google Scholar 

  48. Fix J, Ising AI, Proescholdbell SK, Falls DM, Wolff CS, Fernandez AR, et al. Linking emergency medical services and emergency department data to improve overdose surveillance in North Carolina. Public Health Rep. 2021;136(1_suppl):54–61.

    Article  Google Scholar 

  49. Redfield C, Schoenfeld DW, Ullman E, Tlimat A, Nathanson LA, Halpern Y, et al. Derivation and validation of a machine learning record linkage algorithm between emergency medical services and the emergency department. J Am Med Inform Assoc. 2020;27(1):147–53.

    Article  PubMed  Google Scholar 

  50. Seymour CW, Kahn JM, Martin-Gill C, Callaway CW, Angus DC, Yealy DM. Creating an infrastructure for comparative effectiveness research in emergency medical services. Acad Emerg Med. 2014;21(5):599–607.

    Article  PubMed  PubMed Central  Google Scholar 

  51. Newgard CD, Malveau S, Zive D, Lupton J, Lin A. Building A longitudinal cohort from 9-1-1 to 1-year using existing data sources, probabilistic linkage, and multiple imputation: a validation study. Acad Emerg Med. 2018;25(11):1268–83.

    Article  PubMed  Google Scholar 

  52. Engels PT, Coates A, MacDonald RD, Ahghari M, Welsford M, Dodd T, et al. Toward an all-inclusive trauma system in central south Ontario: development of the trauma-system performance improvement and knowledge exchange (T-SPIKE) project. Can J Surg. 2021;64(2):E162–72.

    Article  PubMed  PubMed Central  Google Scholar 

  53. Mumma BE, Diercks DB, Danielsen B, Holmes JF. Probabilistic linkage of prehospital and outcomes data in out-of-hospital cardiac arrest. Prehosp Emerg Care. 2015;19(3):358–64.

    Article  PubMed  Google Scholar 

  54. Ibrahim G, Nickles AV, Wall SR, O’Brien SL, Scorcia-Wilson T, Wahl R, et al. Assessing the accuracy of a linkage between the Michigan emergency medical services information system and the Michigan coverdell acute stroke registry. Stroke. 2019;50(Supplement 1):316.

    Google Scholar 

  55. Dewolf P, Rutten B, Wauters L, Van den Bempt S, Uten T, Van Kerkhoven J, et al. Impact of video-recording on patient outcome and data collection in out-of-hospital cardiac arrests. Resuscitation. 2021;165:1–7.

    Article  PubMed  Google Scholar 

  56. Sundermann ML, Salcido DD, Koller AC, Menegazzi JJ. Inaccuracy of patient care reports for identification of critical resuscitation events during out-of-hospital cardiac arrest. Am J Emerg Med. 2015;33(1):95–9.

    Article  PubMed  Google Scholar 

  57. Jaureguibeitia X, Aramendi E, Irusta U, Alonso E, Aufderheide TP, Schmicker RH, et al. Methodology and framework for the analysis of cardiopulmonary resuscitation quality in large and heterogeneous cardiac arrest datasets. Resuscitation. 2021;168:44–51.

    Article  PubMed  Google Scholar 

  58. Perez O, Barnhart BJ, Hu C, Spaite DW, Helfenbein E, Babaeizadeh S, et al. Prehospital blood pressure measurement in major traumatic brain injury: concordance between EMS provider documentation and non-invasive monitor data tracking. Circulation. 2017;136(Supplement 1):14669.

    Google Scholar 

  59. Perez O, Barnhart BJ, Spaite DW, Gaither JB, Denninghoff KR, Keim SM, et al. Accuracy of ems hypoxia documentation compared to continuous non-invasive monitor data in major traumatic brain injury. J Emerg Med. 2017;53(3):443.

    Article  Google Scholar 

  60. Poulsen NR, Kløjgård TA, Lübcke K, Lindskou TA, Søvsø MB, Christensen EF. Completeness in the recording of vital signs in ambulances increases over time. Dan Med J. 2020;67(2):A07190421.

    PubMed  Google Scholar 

  61. Foster A, Florea V, Fahrenbruch C, Blackwood J, Rea TD. Availability and accuracy of EMS information about chronic health and medications in cardiac arrest. West J Emerg Med. 2017;18(5):864–9.

    Article  PubMed  PubMed Central  Google Scholar 

  62. Coventry LL, Bremner AP, Williams TA, Jacobs IG, Finn J. Symptoms of myocardial infarction: concordance between paramedic and hospital records. Prehosp Emerg Care. 2014;18(3):393–401.

    Article  PubMed  Google Scholar 

  63. Reisner AT, Chen L, Reifman J. The association between vital signs and major hemorrhagic injury is significantly improved after controlling for sources of measurement variability. J Crit Care. 2012;27(5):533e1-e10.

    Article  Google Scholar 

  64. Frisch A, Reynolds JC, Condle J, Gruen D, Callaway CW. Documentation discrepancies of time-dependent critical events in out of hospital cardiac arrest. Resuscitation. 2014;85(8):1111–4.

    Article  PubMed  Google Scholar 

  65. Staff T, Søvik S. A retrospective quality assessment of pre-hospital emergency medical documentation in motor vehicle accidents in south–eastern Norway. Scand J Trauma Resusc Emerg Med. 2011;19:20.

    Article  PubMed  PubMed Central  Google Scholar 

  66. Tonsager K, Kruger AJ, Ringdal KG, Rehn M. Data quality of glasgow coma scale and systolic blood pressure in scientific studies involving physician-staffed emergency medical services: systematic review. Acta Anaesthesiol Scand. 2020;64(7):888–909.

    Article  PubMed  Google Scholar 

  67. Nishiyama C, Brown SP, May S, Iwami T, Koster RW, Beesems SG, et al. Apples to apples or apples to oranges? International variation in reporting of process and outcome of care for out-of-hospital cardiac arrest. Resuscitation. 2014;85(11):1599–609.

    Article  PubMed  PubMed Central  Google Scholar 

  68. Oostema JA, Nickles A, Reeves MJ. A comparison of probabilistic and deterministic match strategies for linking prehospital and in-hospital stroke registry data. J Stroke Cerebrovasc Dis. 2020;29(10):105151.

    Article  PubMed  Google Scholar 

  69. Lerner EB, Browne LR, Studnek J, Mann NC, Dai M, Hoffman C, et al. Novel use of the national emergency medical services information system to create a pediatric emergency care applied research network-specific emergency medical services patient registry. Acad Emerg Med. 2021;28(SUPPL 1):S113.

    Google Scholar 

  70. Ji C, Lall R, Scomparin C, Horton J, Smyth MA, Quinn T, et al. Feasibility of data linkage in the PARAMEDIC trial: a cluster randomised trial of mechanical chest compression in out-of-hospital cardiac arrest. BMJ Open. 2018;8(7):e021519.

    Article  PubMed  PubMed Central  Google Scholar 

  71. Stromsoe A, Svensson L, Axelsson AB, Goransson K, Todorova L, Herlitz J. Validity of reported data in the Swedish cardiac arrest register in selected parts in Sweden. Resuscitation. 2013;84(7):952–6.

    Article  CAS  PubMed  Google Scholar 

  72. Garcia Minguito L, Casas Sanchez JdD, Rodriguez Albarran MS. [A proposed scale to analyze the quality of injury reports in cases of gender violence] Propuesta de baremo (de escala) para analizar la calidad de los partes de lesiones en casos de violencia de genero. 2012;26(3):256–60.

  73. Bradley NL, Garraway N, Bell N, Lakha N, Hameed SM. Data capture and communication during transfers to definitive care in an inclusive trauma system. Injury. 2017;48(5):1069–73.

    Article  PubMed  Google Scholar 

  74. Randell D. Documentation mnemonic and rubric substantially improved documentation. Educator Update. 2020 (Winter 2019-20):13–16.

  75. Newgard CD, Zive D, Weathers C, Jui J, Daya M. Electronic versus manual data processing: evaluating the use of electronic health records in out-of-hospital clinical research. Acad Emerg Med. 2012;19(2):217–27.

    Article  PubMed  PubMed Central  Google Scholar 

  76. Kearney AS, George N, Karim N, Aluisio AR, Levine AC, Kabeja LM, et al. Development of a trauma and emergency database in Kigali. Rwanda Afr J Emerg Med. 2016;6(4):185–90.

    Article  PubMed  Google Scholar 

  77. Swor R, Qu L, Putman K, Sawyer KN, Domeier R, Fowler J, et al. Challenges of using probabilistic linkage methodology to characterize post-cardiac arrest care in Michigan. Prehosp Emerg Care. 2018;22(2):208–13.

    Article  PubMed  Google Scholar 

  78. Newgard CD, Zive D, Malveau S, Leopold R, Worrall W, Sahni R. Developing a statewide emergency medical services database linked to hospital outcomes: a feasibility study. Prehosp Emerg Care. 2011;15(3):303–19.

    Article  PubMed  Google Scholar 

  79. MacDougall L, Smolina K, Otterstatter M, Zhao B, Chong M, Godfrey D, et al. Development and characteristics of the provincial overdose cohort in British Columbia, Canada. PLoS ONE. 2019;14(1):e0210129.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  80. Fein M, Quinn J, Watt K, Nichols T, Kimble R, Cuttle L. Prehospital paediatric burn care: New priorities in paramedic reporting. Emerg Med Australas. 2014;26(6):609–15.

    Article  PubMed  Google Scholar 

  81. Babcock L, Lloyd J, Semenova O, Meinzen-Derr J, Depinet H. Prehospital capture of variables commonly used in ED sepsis screening tools. Pediatrics. 2019;144(2_MeetingAbstract):412.

  82. Deasy C, Hall D, Bray JE, Smith K, Bernard SA, Cameron P. Paediatric out-of-hospital cardiac arrests in Melbourne, Australia: improved reporting by adding coronial data to a cardiac arrest registry. Emerg Med J. 2013;30(9):740–4.

    Article  PubMed  Google Scholar 

  83. Oud FRW, Kooij FO, Burns BJ. Long-term effectiveness of the airway registry at sydney helicopter emergency medical service. Air Med J. 2019;38(3):161–4.

    Article  PubMed  Google Scholar 

  84. Hern HG, Alter H, Barger J, Teves M, Hamilton K, Mueller L. A focused educational intervention increases paramedic documentation of patient pain complaints. Acad Emerg Med. 2012;19(SUPPL. 1):S202–3.

    Google Scholar 

  85. Timóteo MD, Dantas RA, Costa IC, Silva TT, Santos KV, Oliveira ED, Dantas DV. Implementation of improvement cycle in health records of mobile emergency prehospital care. Revista Brasileira de Enfermagem. 2020;17:73.

    Google Scholar 

  86. Katzer R, Barton DJ, Adelman S, Clark S, Seaman EL, Hudson KB. Impact of implementing an EMR on physical exam documentation by ambulance personnel. Appl Clin Inform. 2012;3(3):301–8.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  87. Lippert F, Folke F, Christensen HC, Blomberg SN. Transition of medical records from paper to electronic records—implications for out-of-hospital cardiac arrest registration. Resuscitation. 2019;142(Supplement 1):e78.

    Google Scholar 

  88. Ko PC-I, Chiang W-C, Chiu AW-H, Lin C-H, Lin H-T, Chen Q-M, et al. Abstract 208: Innovative web-based e-registry enhances survival after out-of-hospital cardiac arrest. Circulation. 2012;126(suppl_21):A208–A.

  89. Smith M, Lix LM, Azimaee M, Enns JE, Orr J, Hong S, et al. Assessing the quality of administrative data for research: a framework from the Manitoba centre for health policy. J Am Med Inform Assoc. 2018;25(3):224–9.

    Article  PubMed  Google Scholar 

  90. O’Connor K, Golding M. Assessment of the availability and utility of the paramedic record in the emergency department. Emerg Med Australas. 2021;33(3):485–90.

    Article  Google Scholar 

  91. Watts T. Abstract P118: process improvement for stroke/EMS run sheets available in medical record. Stroke. 2021;52(Suppl1):118.

    Google Scholar 

  92. Lau F, Antonio M, Davison K, Queen R, Devor A. A rapid review of gender, sex, and sexual orientation documentation in electronic health records. J Am Med Inform Assoc. 2020;27(11):1774–83.

    Article  PubMed  PubMed Central  Google Scholar 

  93. Thompson HM, Kronk CA, Feasley K, Pachwicewicz P, Karnik NS. Implementation of gender identity and assigned sex at birth data collection in electronic health records: where are we now? Int J Environ Res Public Health. 2021;18(12):1–12.

    Article  Google Scholar 

  94. Rahilly-Tierney C, Altincatal A, Agan A, Albert S, Ergas R, Larochelle L, et al. Linking ambulance trip and emergency department surveillance data on opioid-related overdose Massachusetts. Public Health Rep. 2021;136(suppl):47s–53s.

    Article  PubMed  PubMed Central  Google Scholar 

  95. Morey BN, Chang RC, Thomas KB, Tulua A, Penaia C, Tran VD, Pierson N, Greer JC, Bydalek M, Ponce N. No equity without data equity: data reporting gaps for native Hawaiians and Pacific Islanders as structural racism. J Health Politics, Policy law. 2022;47(2):159–200.

    Article  Google Scholar 

  96. Canadian Institute for Health Information. Guidance on the use of standards for race-based and indigenous identity data collection and health reporting in Canada. Ottawa: ON Canadian Institute for Health Information; 2022.

    Google Scholar 

  97. Paprica PA, Sutherland E, Smith A, Brudno M, Cartagena RG, Crichlow M, et al. Essential requirements for establishing and operating data trusts: practical guidance co-developed by representatives from fifteen Canadian organizations and initiatives. Int J Popul Data Sci. 2020;5(1):1353.

    PubMed  PubMed Central  Google Scholar 

  98. Aitken M, Tully MP, Porteous C, Denegri S, Cunningham-Burley S, Banner N, et al. Consensus statement on public involvement and engagement with data intensive health research. Int J Popul Data Sci. 2019;4(1):586.

    PubMed  PubMed Central  Google Scholar 

Download references

Acknowledgements

The authors would like to thank librarian Hal Loewen for support with the search strategy.

Funding

The work was supported by the Pamela Hardisty Graduate Fellowship from the University of Manitoba, held by the lead author.

Author information

Authors and Affiliations

Authors

Contributions

NM, RP, and MD conceived the idea for the review. NM developed the methods under the supervision of RP, DK, GG, and MD. NM, RP, and NL conducted the search, screening, and data extraction. NM drafted the manuscript with substantial input from all co-authors.

Corresponding author

Correspondence to Neil McDonald.

Ethics declarations

Ethics approval and consent to participate

Not Applicable.

Consent for publication

Not Applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1

Documentation of searches.

Additional file 2

Data extraction form.

Additional file 3

Citations to all articles included in the review.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

McDonald, N., Little, N., Kriellaars, D. et al. Database quality assessment in research in paramedicine: a scoping review. Scand J Trauma Resusc Emerg Med 31, 78 (2023). https://doi.org/10.1186/s13049-023-01145-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13049-023-01145-2

Keywords