Article Text

PDF

Disentangling quality and safety indicator data: a longitudinal, comparative study of hand hygiene compliance and accreditation outcomes in 96 Australian hospitals
  1. Virginia Mumford1,
  2. David Greenfield1,
  3. Anne Hogden1,
  4. Deborah Debono1,
  5. Elena Gospodarevskaya2,
  6. Kevin Forde3,
  7. Johanna Westbrook2,
  8. Jeffrey Braithwaite1
  1. 1Centre for Clinical Governance Research, Australian Institute of Health Innovation, UNSW, Sydney, New South Wales, Australia
  2. 2Centre for Health Systems and Safety Research, Australian Institute of Health Innovation, UNSW, Sydney, New South Wales, Australia
  3. 3School of Public Health and Community Medicine, UNSW, Sydney, New South Wales, Australia
  1. Correspondence to Dr Virginia Mumford; v.mumford{at}unsw.edu.au

Abstract

Objectives The study aims are twofold. First, to investigate the suitability of hand hygiene as an indicator of accreditation outcomes and, second, to test the hypothesis that hospitals with better accreditation outcomes achieve higher hand hygiene compliance rates.

Design A retrospective, longitudinal, multisite comparative survey.

Setting Acute public hospitals in New South Wales, Australia.

Participants 96 acute hospitals with accreditation survey results from two surveys during 2009–2012 and submitted data for more than four hand hygiene audits between 2010 and 2013.

Outcomes Our primary outcome comprised observational hand hygiene compliance data from eight audits during 2010–2013. The explanatory variables in our multilevel regression model included: accreditation outcomes and scores for the infection control standard; timing of the surveys; and hospital size and activity.

Results Average hand hygiene compliance rates increased from 67.7% to 80.3% during the study period (2010–2013), with 46.7% of hospitals achieving target compliance rates of 70% in audit 1, versus 92.3% in audit 8. Average hand hygiene rates at small hospitals were 7.8 percentage points (pp) higher than those at the largest hospitals (p<0.05). The association between hand hygiene rates, accreditation outcomes and infection control scores is less clear.

Conclusions Our results indicate that accreditation outcomes and hand hygiene audit data are measuring different parts of the quality and safety spectrum. Understanding what is being measured when selecting indicators to assess the impact of accreditation is critical as focusing on accreditation results would discount successful hand hygiene implementation by smaller hospitals. Conversely, relying on hand hygiene results would discount the infection control related research and leadership investment by larger hospitals. Our hypothesis appears to be confounded by an accreditation programme that makes it more difficult for smaller hospitals to achieve high infection control scores.

Statistics from Altmetric.com

Strengths and limitations of this study

  • The main strengths of this study relate to the use of a comprehensive data set involving the number of acute hospitals (96) participating in the accreditation process and the length of follow-up over eight hand hygiene audit points and two accreditation cycles.

  • This study also addresses an important research question in terms of identifying and assessing the components of hospital accreditation and quantifying their inter-related benefits.

  • The results have important implications for health policymakers internationally in terms of designing health service accreditation programmes that accurately monitor a wide range of hospital sizes and types.

  • The main limitation was the lack of a control group as all hospitals in the survey were accredited. This meant that it was not possible to assess direct or reverse causal relationships or to prevent an omitted variable bias.

  • Other limitations include potential measurement error resulting from the use of self-reported hand hygiene data; however, the data collection methods adhered to the WHO's best practice guidelines.

Introduction

Hospital accreditation programmes are designed to set clinical and organisational standards, assess compliance with those standards and strengthen quality improvement efforts. Accreditation is widely practised with national level accreditation agencies active in at least 27 countries.1 The problems associated with measuring accreditation benefits are well documented.2 ,3 A clear understanding of the inputs, in terms of costs and resource use, and outcomes, in terms of improved patient safety and quality, is essential in ensuring that accreditation programmes are achieving their aims of improving patient safety and healthcare quality.4 ,5 Measuring the effects of accreditation on clinical practice and quality of care is important as we need to determine whether the cost burden of data collection and audit processes is outweighed by the expected improvements in quality and safety outcomes.5

In this study, we analyse the relationship between hand hygiene compliance rates and accreditation outcomes in order to test the suitability of hand hygiene as an indicator of accreditation outcomes. Our hypothesis is that hospitals with better accreditation outcomes and infection control scores than others reflect organisational processes that support a positive culture towards improving quality and safety,6 and therefore they would achieve higher hand hygiene compliance rates.

Hand hygiene assessment is an integral component of the infection control standards used to evaluate whether Australian hospitals are compliant with accreditation standards.7–9 Hand hygiene compliance rates have been validated as a potential process indicator for accreditation outcomes. Moreover, healthcare-associated infections are recognised as a leading cause of increased morbidity and healthcare costs.10 A US study estimated that there were 1.7 million healthcare-associated infections in 2002, comprising 4.5% of admissions, and resulting in nearly 99 000 deaths.11 A recent meta-analysis estimated the cost of the five most common healthcare-associated infections at US$9.8 billion per annum.12 In Australia, the most recent figures available indicate that healthcare-associated infections resulted in an extra two million bed days in 2005, with estimated additional costs of $A21 million from postdischarge surgical infections.13

There is increasing evidence that improving hand hygiene reduces healthcare-associated infections and the spread of antimicrobial resistance.14–18 However, it is difficult to demonstrate a causal link between hand hygiene and healthcare-associated infections due to a multiplicity of interventions and scarcity of randomised trials.19–21 Nevertheless, the WHO has identified good hand hygiene as a major factor in reducing healthcare-associated infections based on epidemiological evidence.18

Hand hygiene policies in Australian hospitals follow international best practices. They are based on WHO's recommendations with a multimodal approach incorporating: access to cleaning agents at the point of care; training and education; monitoring and feedback; reminders in the workplace; and development of an institutional safety climate.22 Auditors trained by Hand Hygiene Australia monitor hand hygiene activity by direct observation of hospital staff and compare hand hygiene activity against the total number of potential ‘moments’ for hand hygiene.23 The national target for hand hygiene compliance is 70% and audit results are publicly reported three times a year. The Australian Commission on Safety and Quality in Health Care (ACSQHC) has recommended that hand hygiene programmes need to be repeatedly monitored using both process indicators (compliance rates) and outcome indicators (infection rates).24

Methods

Study design, setting and context

The study comprised a retrospective, longitudinal, multisite comparative survey of hand hygiene compliance rates and accreditation outcomes in acute public hospitals in New South Wales (NSW), Australia. With a population of 7.2 million, NSW comprises 30.5% of the 736 public hospitals and 32.0% of the population in Australia.25 We employed retrospective data matching techniques over the study period, 2009–2013, to analyse the relationship between hand hygiene compliance data and accreditation outcomes.

Data matching and analysis

Hand hygiene compliance data

Hand hygiene policies include five ‘moments’ when hand hygiene should be performed: before touching a patient; before a procedure; after a procedure or body fluid exposure risk; after touching a patient and after touching a patient's surroundings.23 ,26 Audits are carried out three times a year by healthcare workers who have been accredited by Hand Hygiene Australia. Surveys are conducted using a standardised observation assessment tool which measures hand hygiene activity versus the total number of observed possible ‘moments’. Auditors are trained in selecting the wards or units for the audit, and the minimum number of required ‘moments’ for each audit is determined by hospital size and activity. We obtained hand hygiene compliance rates data from late 2010 to early 2013 from the NSW Clinical Excellence Commission, the quality and safety body responsible for implementing the hand hygiene initiative in NSW and collecting hand hygiene audit data.27

Accreditation programme and infection control standards

Data on accreditation outcomes from 2009 to 2013 were provided by the Australian Council on Healthcare Standards (ACHS). The ACHS Evaluation and Quality Improvement Program (EQuIP) was introduced in 1997 and comprises a 4-year cycle with external surveys in years 2 and 4.28 During these surveys, hospitals are assessed by an external team of surveyors against ACHS developed standards. The EQuIP has undergone several revisions, none of which materially affected this study. Our study period included EQuIP4, which was introduced in 2007 and used for the surveys in 2009 and 2010, and EQuIP5, which was introduced in 2011. Accreditation standards were changed significantly following the introduction of national mandatory standards in 2013,9 but the infection control criteria were the same for both versions of the EQuIP standards assessed during the study period (see online supplementary file 1). Surveyors scored facilities on a five-point Likert scale for each standard or criterion assessed in the survey. Scores were designated as Outstanding Achievement (OA), Extensive Achievement (EA), Moderate Achievement (MA), Some Achievement and Little Achievement (LA). Hospitals needed to achieve OA, EA or MA, scores in each mandatory standard or criterion in order to meet accreditation requirements.

Infection control related criteria were part of a broader standard regarding the safe provision of care and services.7 ,8 To meet accreditation requirements, hospitals needed to ensure that the infection control policy: met all regulatory requirements and industry guidelines; had executive support; incorporated ongoing education activities; and included indicators to show both compliance with the policy and effective outcome measurements. Additional activities which counted towards achieving higher (EA and OA) scores included: benchmarking of performance indicators; use of feedback to inform and improve; contributing to infection control research and recognised leadership in infection control systems.7 ,8 Accreditation is often granted to a cluster of facilities within a local health district and therefore reflects conditions at all the facilities within that survey group. These clusters are subject to boundary changes, as seen in the NSW 2011 health system reorganisation, which took place during the study period.29 We therefore identified the different hospitals within each cluster in order to match the accreditation scores with the hand hygiene data from individual hospitals.

Study variables

To analyse the matched data, we used hand hygiene compliance rates as our outcome of interest, expressed as a continuous outcome variable. Data were available at eight different time points from the end of 2010 to early 2013 (see figure 1). We characterised the accreditation scores as either full or partial accreditation. Partial accreditation was defined as either accreditation being granted for a reduced time, or resulting in a recommendation for action. No hospitals in the study were refused accreditation during the study period. We included infection control scores in the model by whether hospitals achieved a higher score in one, none (our reference case) or both surveys. To test for a possible timing effect, we included a variable to identify whether surveys were either carried out in the 2009 and 2011 accreditation cycle (cycle=0), or in the 2010 and 2012 cycle (cycle=1).

Figure 1

Timeline of accreditation surveys and hand hygiene audits.

Acute hospitals were grouped according to the Australian Institute of Health and Welfare activity matrix, based on annual numbers of acute episodes of care adjusted for patient complexity, and geographic location (see online supplementary file 2).30 We used these groups to create a categorical variable with principal referral and specialist hospitals as the reference case (principal=0), with large hospitals scored as 1, medium hospitals as 2 and small hospitals as 3.

Analysis

The nature of the data, with irregular audit dates and clustering within hospitals, indicated that a multilevel model would be most appropriate and would allow adjustment for hospital level variance.31 ,32 After matching the accreditation and hand hygiene data, our sample comprised 96 hospitals each with two accreditation surveys. For some hospitals, we were not able to match the accreditation outcomes because of changes in the accreditation clusters. Missing data were higher during the first two audits while the programme was progressively implemented. We did not impute values for missing data as the pattern of missing results indicated that these were not missing at random. For example, data from some of the smaller hospitals were not included as they did not meet the minimum publication requirements during the study period (50 moments of hand hygiene), which we determined was likely to be related to hospital size.

We tested our main model using hand hygiene data from audits 1 through 8 as the outcome variable, and accreditation outcomes, infection control scores, accreditation cycle and peer groups as our explanatory variables. We tested the model fit versus ordinary least squares by calculating the intraclass correlation to assess the within-hospital effect. We also tested whether to use a random coefficient or random intercept model. Analysis was conducted using Stata statistical software (V.12SE),33 applying a two-sided significance level of 5%. We also fitted a restrictive model using data from audits 3–8 to determine whether the different peer group mix in the first two surveys was affecting the results.

Results

Hand hygiene and accreditation data analysis

We assessed hand hygiene data on 118 hospitals from eight different audit points during 2010–2013. Overall, hand hygiene rates showed an improvement during the study period, with 28 out of 60 hospitals (46.7%) achieving 70% compliance rates in the first audit in 2010 vs 108 out of 117 hospitals (92.3%) in the final audit, in early 2013. Average hand hygiene compliance rates increased from 67.7% in audit 1, to 80.3% in audit 8, and remained above the 70% national target rate from audit 2 onwards. The average audit compliance rates by audit and hospital peer group are shown in figure 2.

Figure 2

Mean hand hygiene compliance rates by audit and hospital peer group, 2010–2013.

During the study period, 2009–2013, 61 hospitals underwent an accreditation survey in cycle 0, and 44 in cycle 1. The accreditation outcomes showed that 59% of organisations were granted full accreditation in the first survey in each cycle (during 2009–2010) versus 77% in the second survey (2011–2012). The number of hospitals receiving higher infection control scores also increased over time, with 13% receiving a high score (EA) in the first survey of each accreditation cycle versus 18% in the second surveys during 2011 and 2012. No hospitals received an OA score for infection control during the study period. We further examined whether there was a difference between meeting, or not meeting, the target compliance rates by comparing the partial data from audit 1 when the programme was being rolled out, and the final audit in our study, audit 8 (see figure 3). Large hospitals showed the biggest increase, with 30% meeting the target in audit 1, rising to 100% by audit 8.

Figure 3

Percentage of hospitals meeting hand hygiene targets by audit and hospital peer group.

We also noted that principal and large hospitals comprised 51.7% of hospitals in audit 1 versus 39.3% in audit 8, suggesting they were early adopters in the programme. We tested whether this would influence the results using the restricted model (comprising data from audits 3–8). The infection control outcomes also improved over time with 3% of organisations receiving high scores in both surveys in cycle 0 versus 17% in cycle 1; however, we note the absolute small numbers involved (n=8). A size effect was noted with 6.5% (n=2) of smaller hospitals receiving a higher infection control score in one or both surveys versus 35.7% (n=10) of principal hospitals and 46.2% (n=6) of large hospitals (see table 1).

Table 1

Summary characteristics of accreditation and infection control scores, breakdown by peer group and timing of surveys

Testing the model

The intraclass correlation coefficient indicated that 38% of the variance was due to within-hospital effects, indicating sufficient variance for using a random intercept model.32 We ran a likelihood ratio test using a null model (without the random intercept). This gave a χ2 test result of 190 (p=0.001), which confirmed our approach versus using ordinary least squares. We also tested the model using infection control scores as a random coefficient, but the results (χ2=0.81, p=0.67) indicated that the random intercept model was more appropriate. We noted a ceiling effect in our data with 100% of large hospitals and 92% of all hospitals meeting the target compliance rates by audit 8, with less incentive to reach higher levels (see figure 3). Although hospitals were incrementally enrolled in the hand hygiene programme, the lack of hand hygiene compliance data in the unenrolled hospitals meant we were not able to use a stepped wedge design to provide controls, or evaluate a before and after effect and our results are also subject to omitted variable bias. A fixed effects panel data model might be a more traditional approach to reduce sample variation, but we determined that the random intercept model would be more appropriate due to the policy requirement for all public hospitals to submit hand hygiene data, and the presence of time-invariant variables.34

Multilevel model

After matching the hand hygiene data with hospitals that underwent two accreditation surveys, our main model included data from 661 hand hygiene audits from 96 hospitals, an average of 6.9 audits per hospital. The results (table 2) show that achieving full accreditation for both surveys was not significantly associated with higher hand hygiene rates versus those hospitals achieving full accreditation in only one survey. The association between hand hygiene rates and infection control scores is less clear. Hospitals achieving high infection control scores (EA) in one survey (n=14) showed 4.2 percentage point (pp) lower hand hygiene rates than hospitals which just met the accreditation standard (MA score), and this result was significant (p=0.033). Hospitals achieving high infection control scores in both surveys (n=8) showed higher rates (2.1 pp), but this was not significant (p=0.40). Small-sized and medium-sized hospitals experienced significantly higher hand hygiene compliance rates (7.8 pp for small hospitals and 3.5 pp for medium hospitals) compared with principal hospitals. The restricted model, using data from audit 3 onwards, also showed a negative relationship between higher infection control scores and hand hygiene audit results. However, the effect was smaller (2.9 pp) and the results were no longer significant (p=0.14). The size effect in the restricted model was consistent with the main model, with small and medium hospitals showing significantly higher hand hygiene rates than principal referral hospitals. These results do not lend support to our study hypothesis that high infection control scores are associated with higher hand hygiene rates.

Table 2

Multilevel model to show effect of accreditation outcomes on hand hygiene audit rates

Discussion

Our analysis of 118 NSW hospitals showed that hand hygiene rates increased from 67.7% in audit 1 in late 2010 to 80.3% by audit 8 in early 2013. This is comparable with rates of 62.2% from a sample of NSW hospitals in February 2007, which were observed following the introduction of the Clean Hand Saves Lives campaign during 2006–2007,15 and continues the improvement shown nationally in Australia with average hand hygiene rates estimated at 68.3% in 2011.35 It is challenging to compare these results internationally. A US study estimated average hand hygiene rates of 56.6% in 40 hospitals using data collected for 1 year before and after the introduction of national hand hygiene guidelines in 2002.36 However, it must be noted that this programme was different to that followed by Hand Hygiene Australia and so the results would not be directly comparable.

Smaller hospitals in our study had higher hand hygiene compliance rates but the relationship between accreditation outcomes and hand hygiene data was less clear. This hospital size effect on hand hygiene compliance has been confirmed by other research investigating the link between hand hygiene rates and healthcare-associated Staphylococcus aureus bacteraemia.37 ,38 We consider that the results from small and medium hospitals in our study can be explained by looking at the organisational infrastructure necessary to meet the hand hygiene and accreditation requirements. Both are dependent on a widespread organisational response in terms of education, monitoring, infrastructure and management involvement.6 Achieving higher infection control scores requires additional benchmarking, feedback and research capabilities.7 ,8 The organisational size effect suggests that small-sized and medium-sized hospitals can effectively embrace multimodal quality improvement strategies as seen by the higher compliance rates. However, the requirements for achieving high infection control scores within an accreditation survey may be measuring different aspects of quality that are not reflected in the hand hygiene compliance rates. The results indicate that smaller hospitals are able to focus on the practical implementation of a national hand hygiene policy. Having the resources to meet the requirements for higher infection control scores, in terms of conducting research or being recognised leaders in infection control, may not be practical for these smaller organisations. Although some smaller hospitals will be accredited as part of a larger cluster of hospitals, which includes principal and large hospitals, this is not always the case. Our hypothesis that higher accreditation scores would be reflected in hand hygiene rates appears to be confounded by an accreditation programme that makes it more difficult for smaller hospitals to achieve high infection control scores.

This study uses one indicator for evaluating accreditation, whereas multiple measures may be more effective.39 For example, outcome indicators are widely used in the US hospital system,40–42 and include hospital-acquired Staphylcoccus aureus bacteraemia (SAB) rates and surgical site infection rates. These incorporate a broader mix of the antimicrobial, hand hygiene and specialist cleaning practice modules of the infection control standard. Using outcome indicators would complete Donabedian's triad of performance measures to include structural (accreditation results), process (hand hygiene compliance rates) and outcome (S. aureus infection rates) measures.43 ,44

Measurement issues in our study include a possible observer effect since the hospital staff might be aware that the audit was taking place. However, although this may increase the compliance rate, the results would still be valid as the standards include requirements for staff education and installation of appropriate infrastructure. Increased rates during an audit imply that the correct infrastructure is in place, in terms of availability of functioning hand washing stations, and that the staff are aware of the hand hygiene policies. In addition, any observer effect is likely to be mitigated as all the hospitals used the same methods for collecting data and thus are equally subjected to this bias. Other methods to measure hand hygiene activity include measuring consumption rates of hand hygiene products, such as alcohol rubs,45 and electronic systems for monitoring compliance,46 but the WHO guidelines suggest that direct observation is still the gold standard.40 We also note that although accreditation surveys are assessed on a five-point scale, only two scores (MA and EA) were used in the infection control standard during the four surveys in our study. The lack of granularity in the accreditation scores makes it difficult to differentiate accreditation performance. Intersurveyor reliability is recognised as a limitation of audit systems that are based on subjective assessments.47 To reduce idiosyncratic scoring, ACHS surveyors need to provide evidence to ACHS for their scoring methods, and in the decision to award higher scores, but some variation between surveyors may remain.

Limitations of the model include reverse causality in that higher compliance rates could lead to higher infection control scores at the next survey where hand hygiene audit rates are used as evidence of implementation during an accreditation survey. This would most likely be the case going forward as ACSQHC includes hand hygiene audit results as part of the evidence of implementation of standard 3 under the new national standards.9 ,48

Implications

Different indicators will give different perspectives on how an organisation approaches and implements relevant policy. However, the costs of measurement in healthcare should be balanced against using a range of indicators to capture a broader mix of infection control policies, and across the different standards assessed during accreditation surveys. Indicator selection should include both process indicators, recognised as a method of measuring organisational changes,49 and outcome indicators. Public reporting of indicator data further increases the requirement to accurately identify and measure the parts of the patient safety and quality spectrum that are being addressed. In this study, a focus on the accreditation results would underestimate the successful implementation of the hand hygiene policy by smaller hospitals. Conversely, just using hand hygiene results would underestimate the research and leadership investment in infection control by larger hospitals. Disentangling these two outcomes within the same safety and quality initiative is a pre-requisite to understanding how they can be effectively assessed and monitored. For example, consideration could be given to changing criteria for awarding higher scores for infection control such that achieving higher scores was evidence based and could be feasible for all hospitals. Although we focus on Australian hospitals in our study, international accreditation programmes will also need to ensure that indicators accurately capture outcomes and reflect performance across a range of hospital sizes and types.

Conclusion

Identifying indicators to measure the effectiveness of accreditation is challenging due to the complexity of implementing a wide range of accreditation-related processes across multiple hospital activities. Our results do not support our study hypothesis that high infection control scores are associated with higher hand hygiene rates. Instead, this study suggests that accreditation outcomes and hand hygiene audit data measure different parts of the quality and safety spectrum. Developing a framework to identify suitable indicators is an important contribution to understanding the impact of hospital accreditation internationally. Policymakers need to appreciate the assumptions behind the choice of indicators and understand exactly what is being measured to ensure that key performance indicators encourage quality improvements in the delivery of hospital services.

Acknowledgments

This study forms part of the Accreditation Collaborative for the Conduct of Research, Evaluation and Designated Investigations through Teamwork (ACCREDIT) project. The ACCREDIT collaboration involves researchers in the Centre for Clinical Governance Research and Centre for Health Systems and Safety Research in the Australian Institute of Health Innovation at the University of New South Wales, Australia. The ACCREDIT research team benefits from a high-profile international advisory group containing researchers in health safety and quality from the UK, Spain and Sweden. The collaboration includes two leading health safety and quality bodies (ACSQHC and the NSW Clinical Excellence Commission) plus three of the major Australian health services accreditation agencies: ACHS, Australian General Practice Accreditation Limited, and the Australian Aged Care Quality Agency. The authors would like to thank ACHS and the NSW Clinical Excellence Commission for providing data for the study. They would also like to acknowledge the feedback received on the first draft of the paper from the PhD group at the Centre for Health Economics Research and Evaluation at the University of Technology Sydney, in particular: Professor Marion Haas; Dr Rebecca Reeve and Ms Bonny Parkinson.

References

View Abstract
  • Supplementary Data

    This web only file has been produced by the BMJ Publishing Group from an electronic file supplied by the author(s) and has not been edited for content.

    Files in this Data Supplement:

Footnotes

  • Contributors JB, DG, JW and VM contributed to the concept and design of the study. VM, JB and DG contributed to the acquisition of data. EG contributed to the model specification and model analysis, and to the drafting of the results sections. KF and AH contributed to the data collection and analysis, and to the drafting of the methods section. VM wrote the first draft of the manuscript. JB and DG provided a critical revision of the first draft. JW contributed to the drafting of the discussion section. All authors reviewed the final manuscript and approved the final manuscript.

  • Funding This research is supported by the National Health and Medical Research Council Program Grant 568612, and under the Australian Research Council's Linkage Projects scheme (project LP100200586). Industry partners (ACHS, Australian General Practice Accreditation Limited, and the Australian Aged Care Quality Agency) contributed to the project design and may provide assistance with its implementation. They may also contribute to data analysis, interpretation of findings and presentation of results. Final decision-making responsibility for all research activities, including the decision to submit research for publication, resides with UNSW.

  • Competing interests None.

  • Ethics approval The University of New South Wales Human Research Ethics Committee has approved this project (approval number HREC 10274).

  • Provenance and peer review Not commissioned; externally peer reviewed.

  • Data sharing statement No additional data are available.

Request permissions

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.