Impact of Physician Scorecards on Emergency Department Resource Use, Quality, and Efficiency Shabnam Jain, MD, MPHa,b, Gary Frank, MDb, Kelly McCormick, MSPHc, Baohua Wu, MSc, Brent A. Johnson, PhDc abstract a Department of Pediatrics and Emergency Medicine, Emory University School of Medicine, Atlanta, Georgia; bChildren’s Healthcare of Atlanta, Atlanta, Georgia; and cDepartment of Biostatistics, Rollins School of Public Health, Emory University, Atlanta, Georgia Dr Jain conceptualized the study, designed the scorecards, interpreted the data, and drafted the initial manuscript; Dr Frank assisted with scorecard implementation and data review and revised the manuscript; Ms McCormick assisted with data analysis and revised the manuscript; Mr Wu worked on data analysis, generated the summary tables and graphs, and revised the manuscript; Dr Johnson analyzed the data and revised the manuscript; and all authors approved the final manuscript as submitted. Variability in practice patterns and resource use in the emergency department (ED) can affect costs without affecting outcomes. ED quality measures have not included resource use in relation to ED outcomes and efficiency. Our objectives were to develop a tool for comprehensive physician feedback on practice patterns relative to peers and to study its impact on resource use, quality, and efficiency. BACKGROUND AND OBJECTIVES: METHODS: We evaluated condition-specific resource use (laboratory tests; imaging; antibiotics, intravenous fluids, and ondansetron; admission) by physicians at 2 tertiary pediatric EDs for 4 common conditions (fever, head injury, respiratory illness, gastroenteritis). Resources used, ED length of stay (efficiency measure), and 72-hour return to ED (return rate [RR]) (balancing measure) were reported on scorecards with boxplots showing physicians their practice relative to peers. Quarterly scorecards were distributed for baseline (preintervention, July 2009–August 2010) and postintervention (September 2010–December 2011). Preintervention, postintervention, and trend analyses were performed. RESULTS: In 51 450 patient visits (24 834 preintervention, 26 616 postintervention) DOI: 10.1542/peds.2014-2363 seen by 96 physicians, we observed reduced postintervention use of abdominal and pelvic and head computed tomography scans, chest radiographs, intravenous antibiotics, and ondansetron (P , .01 for all). Hospital admissions decreased from 7.4% to 6.7% (P = .002), length of stay from 112 to 108 minutes (P , .001), and RR from 2.2% to 2.0%. Trends for use of laboratory tests and intravenous antibiotics showed significant reduction (P , .001 and P , .05, respectively); admission trends increased, and trends for use of computed tomography scans and plain abdominal radiographs showed no change. Accepted for publication Mar 6, 2015 CONCLUSIONS: Physician www.pediatrics.org/cgi/doi/10.1542/peds.2014-2363 Address correspondence to Shabnam Jain, MD, MPH, Emory University, Children’s Healthcare of Atlanta, Pediatric Emergency Medicine, 1645 Tullie Circle NE, Atlanta, GA 30329. E-mail: [email protected] PEDIATRICS (ISSN Numbers: Print, 0031-4005; Online, 1098-4275). Copyright © 2015 by the American Academy of Pediatrics FINANCIAL DISCLOSURE: The authors have indicated they have no financial relationships relevant to this article to disclose. FUNDING: Partially funded by the Children’s Healthcare of Atlanta Research Fund. POTENTIAL CONFLICT OF INTEREST: The authors have indicated they have no potential conflicts of interest to disclose. QUALITY REPORT feedback on practice patterns relative to peers results in reduction in resource use for several common ED conditions without adversely affecting ED efficiency or quality of care. There has been considerable focus on overall resource use and cost of care in our nation’s emergency departments (EDs), particularly for nonemergent conditions that often make up a large share of the case mix in the typical ED. Additionally, wide variation in practice and resource use that cannot be explained by patient-related factors has been demonstrated in both adults and children, including emergency medicine patients.1–12 Excessive use of resources in health care has not been found to improve quality or outcomes, but it does affect cost.11,13–16 The recent Choosing Wisely initiative encourages every specialty to consider reducing use of tests and procedures that are often unnecessary and sometimes can be harmful.17 Traditional ED quality measures have included measures of timeliness such PEDIATRICS Volume 136, number 3, September 2015 as length of stay, time to antibiotics, boarder time, safety measures such as errors, hand-washing, and measures of patient-centeredness such as leftwithout-being-seen rates and patient satisfaction. Unexpected return to ED soon after an initial ED visit is a measure of effectiveness of care that is monitored in most EDs. Recent studies have suggested the importance of measuring efficient use of resources for common pediatric ED conditions.18,19 Efforts to streamline resource use by standardizing practice through evidence-based guidelines have been ongoing, but significant degrees of variation in practice remain.4,5,19 Current ED quality measures do not include measures of efficient use of resources or how they correlate with ED outcomes. In recently proposed indicators of quality in pediatric EDs, none of the 62 evaluated unnecessary tests.20 A lack of system incentives such as quality measures has been noted as one of the reasons for overdiagnosis that can result in a cascade of effects that do not always improve outcomes.21 Audit and feedback of physician practice has been used as a tool to improve resource use and care efficiency; when physicians are provided with data on how they practice, they may be more likely to consider practice changes.22–24 Although insurers have been profiling providers’ resource use patterns, such data are often not available to individual clinicians, usually are not acuity-adjusted, and do not include a combination of metrics that balance resource use and outcome measures. Balanced scorecards have been used in health care to provide a comprehensive picture of performance at the organizational level but have not been studied at the individual provider level.25–28 The impact of providing physicians with comprehensive and balanced feedback on their practice patterns relative to peers in the same setting is not known. In this quality improvement initiative, our objectives PEDIATRICS Volume 136, number 3, September 2015 were to develop a tool for comprehensive feedback to ED physicians on their practice patterns relative to peers and to evaluate the impact of such a physician feedback tool on ED resource use and associated ED quality measures, ED length of stay (efficiency measure), and return to ED (balancing measure). METHODS Setting and Scorecard Development The study was conducted at a large pediatric health care system’s 2 tertiary-level EDs with a combined annual census of .150 000 pediatric visits. Both sites are staffed by pediatric emergency medicine physicians and urgent care pediatricians and have electronic medical records (EMRs). We have previously documented large variation in practice in the ED even after adjusting for patient (eg, acuity, age) and temporal factors.4 We used this persistent variation in practice to highlight individual physician performance relative to the range of practice of their peers. We did this by creating a comprehensive and balanced scorecard showing physicians their resource use and ED quality metrics, relative to their peers, for 4 common ED conditions. These conditions were chosen because they are some of the most common presentations to pediatric EDs. Institutional guidelines for the management of fever have been in place since 2004 and for the 2 most common respiratory illnesses (asthma and bronchiolitis) since 2006. Guidelines for the management of mild head injury were implemented in mid2011, and there are no institutional guidelines for the management of children with gastroenteritis-like illness. Although guidelines can influence practice patterns, such change takes time, especially if not associated with measurement of compliance with guideline recommendations. During our study period, there were no new efforts to reinforce compliance to guidelines. Scorecard Inclusion Criteria Four common ED conditions were included in the scorecard. To avoid bias based on final diagnosis, we based inclusion criteria on patient presenting complaints. Professional coders classified patient chief complaints into “admitting” diagnoses using International Classification of Diseases, Ninth Revision, Clinical Modification codes. The 4 conditions and the corresponding chief complaint International Classification of Diseases, Ninth Revision, Clinical Modification codes were Fever unspecified (780.60) (age .2 months only because infants ,2 months old receive routine screening tests and are often admitted); Head Injury unspecified (959.01) (age .3 months because institutional guidelines apply only to infants .3 months old); Gastroenteritis-like symptoms: Vomiting alone (787.03), Persistent Vomiting (536.2), Diarrhea (787.91), Dehydration (276.51); and Respiratory illness: Other Dyspnea and Respiratory Abnormality (786.09), Cough (786.2), Wheezing (786.07), Unspecified Asthma (493.90), with exacerbation (493.92), with status asthmaticus (493.91). Acuity Adjustment and Peer Comparisons Studies on ED provider feedback have underscored the importance of adjusting for acuity, diagnosis, and patient outcomes, and they recommend making comparisons of resource use rates and outcomes for individual physicians against peerbased norms.24 In our study, we achieved acuity adjustment by including only patients in Emergency Severity Index category 3 (midacuity).29,30 The potential for practice variation is highest for these midacuity patients, and the 4 conditions included in the scorecard represent nearly 40% of all midacuity patients seen in our ED. Peer group performance was shown on the scorecard as a standard boxplot with 25th, 50th, and 75th percentiles, with e671 dashes for 1.5 interquartile ranges. A bold diamond indicated an individual physician’s performance against the peer group’s performance shown on the boxplot (Fig 1). Quality Measures Included in Scorecard Three measures of ED quality of care were included in the scorecard: resource use, ED length of stay, and 72-hour return rate (Table 1). Resource use, a process measure, was monitored specific to each condition. Additionally, because EDs account for .50% of hospital admissions, and arguably this is one of the most expensive resource use decisions made by an ED physician and is considered discretionary, admission rate to the hospital was measured for each of the 4 conditions.31 ED length of stay (LOS; measured as time from physician assuming care of the patient until exit from the ED) was included as a timeliness measure. Rate of 72-hour return to the ED for the same condition was included as a balancing quality measure. The scorecard also included the total number of patients seen by each physician during the reporting period. Exclusion Criteria Patients who left without being seen were excluded. Performance data of physicians who saw ,10 patients during a reporting period in any condition were not included in the peer group (boxplot) calculation for that condition, to prevent proportions with low denominators from unduly influencing percentiles. Furthermore, physicians who saw ,25 patients during either the preintervention or postintervention phase were excluded from analysis because these physicians were not regular ED physicians, or they left the practice at some point, and therefore did not meaningfully contribute to changes in practice patterns. Physician Attribution Physical attribution was based on the attending physician assigned to the patient during the ED visit. When there was transfer of care during the ED visit, resource use decisions were assigned to the first physician, whereas the disposition decision and 72-hour return were assigned to the second (dispositioning) physician. Intervention The intervention (feedback via scorecards) began in September 1, 2010. Feedback was provided through 3 related interventions. The first was quarterly distribution of scorecards electronically. The first scorecard was for the reporting period July 2009 to June 2010, and FIGURE 1 A boxplot showing an individual physician’s performance (shown as a blue diamond) against the peer group’s performance. e672 subsequent scorecards were distributed quarterly, with each reporting quarter representing a previous 12-month rolling average. The second intervention was presentation of scorecards for educational purposes at practice meetings. Third, physician directors of both EDs regularly reviewed with physicians their performance on a 1-on-1 basis, particularly with physicians who had outlier practice patterns. The 14-month period from July 2009 to August 2010 (just before the first scorecard was distributed) was the preintervention phase, and the 16-month period from September 2010 to December 2011 was the postintervention phase. Data Source Data were obtained from EMRs and administrative data that are stored in an institutional data warehouse. The EMRs have electronic signature capture that allows for accurate physician attribution; computerized physician order entry ensured that resource use was accurately assigned to the ordering physician. Data were aggregated at the level of individual providers, who were identified by blinded codes not accessible to investigators. Statistical Analysis To assess the effect of this scorecard intervention, we conceived 2 complementary statistical analyses: compare the aggregate group-level difference in resource use before and after the intervention, and compare physician-level serial trends before and after the intervention to evaluate longer-term effects of the intervention. The former analysis allows comparison of aggregate group-level averages for 11 resources and outcomes of interest before and after the intervention. For this analysis, we used x2 tests (for categorical data), 2-sample t tests (for normally distributed continuous data), and Wilcoxon rank-sum tests (for skewed continuous data, eg, JAIN et al TABLE 1 Quality Measures Included in Scorecard Fever Resources ED LOSa 72-h return rateb Gastroenteritis-like Illness Laboratory tests CBC BMP or CMP Blood culture CRP Imaging Chest radiograph Abdominal radiograph Head Injury Respiratory Illness Imaging Head CT scan Imaging Chest radiograph Laboratory tests CBC BMP or CMP Imaging Abdominal radiograph Abdominal CT scan IV fluids IV ondansetron Hospital admission: reported for all 4 conditions Reported for all 4 conditions Reported for all 4 conditions BMP, basic metabolic profile; CBC, complete blood cell count; CMP, comprehensive metabolic profile; CRP, C-reactive protein. a ED LOS measured as time from physician assuming care of patient until exit from the ED. b 72-h return rate: Return to the ED for same condition within 72 h of an initial visit. LOS). The latter analysis is a secondary analysis of findings to assess the trends both before and after intervention and to determine whether there are statistical differences in these 2 trends. For this analysis, we used piecewise (generalized) linear mixed models to model the scorecard outcomes as a function of each resource or outcome on the scorecard while accounting for physician-level clustering and effect of time. Each month, each physician was a repeated measure, and each physician started at his or her own baseline. A Poisson distribution was used to model responses that are counts. The effect of the intervention was measured by the change in the slope of the trend lines before and after the intervention, showing change in monthly trends over time (as opposed to change in practice at a single point in time), reflecting sustained effect. Results were noted in terms of marginal effects (eg, the change in the probability that a patient receives a test before and after intervention). We performed all statistical analyses on SAS (SAS Institute, Inc, Cary, NC) statistical software.32 The institutional review board exempted this study from review because only aggregate data were used, with no physician or patient PEDIATRICS Volume 136, number 3, September 2015 identifiers and no patient-level intervention. RESULTS During the study period from July 2009 to December 2011, a total of 336 294 patients were seen in the EDs, of whom 128 691 were in Emergency Severity Index 3 acuity. Of these, 51 450 patients met inclusion criteria (preintervention, 24 834; postintervention, 26 616). Ninety-six physicians saw these patients, with a mean of 536 patients per physician. Figure 2 shows a typical scorecard of a physician during the study period for all 4 conditions. Table 2 shows the overall preintervention–postintervention results for the various measures reported on scorecards. Resource categories shown reflect only those that were included in the quality improvement initiative and reported to providers via the scorecard. Overall, statistically significant reduction was noted in use of abdominal and pelvic computed tomography (CT) scans, head CT scans, chest radiographs, intravenous (IV) antibiotics, and IV ondansetron (P , .01 for all). Hospital admission rates decreased from 7.4% to 6.7% (P = .002). The median ED LOS for the 4 included conditions decreased from 112 minutes to 108 minutes (P , .001). The 72-hour return rate changed from 2.2% to 2.0%; though not statistically significant, the rate did not increase as use of tests and therapies during the initial visit decreased. Table 3 summarizes the trends over time for significant findings in Table 2. Statistically significant decrease in trends over time was noted for use of laboratory tests (P , .001) and for use of IV antibiotics (P = .02); admission trends showed an increase (P = .02). Other resource categories did not show statistically significant change in trends over time during the study period. Figure 3 shows trends over time as monthly estimates of rates of laboratory tests performed, IV antibiotic use, hospital admissions, IV ondansetron use, abdominal CT scans, and abdominal radiographs performed before and after the introduction of physician-level scorecards. DISCUSSION Our study shows reduction in resource use for several commonly seen conditions in the pediatric ED after physicians were provided with feedback on practice patterns, including resource use and quality metrics, relative to their peers. There was a small improvement in efficiency (LOS), and reduced resource use did not adversely affect quality of care (return rate). e673 FIGURE 2 Scorecard for 1 physician for all 4 conditions (fever, gastroenteritis-type illness, respiratory illness, and head injury) during the entire study period. Abx, antibiotics; KUB, kidneys, ureter, bladder; pts, patients; XR, radiograph. However, results were not consistent, and some trends predated the intervention. e674 Traditional ED quality measures include operational metrics such as turnaround time and LOS, timeliness measures such as time to antibiotics in newborns or patients with sickle cell disease and fever, and outcome JAIN et al FIGURE 2 Continued. measures such as patient satisfaction and left-without-being-seen rates.33–36 Many of these measures do not reflect physician-related clinical decision-making in the care provided to the patient and are often not directly controlled by ED physicians. Stang et al20 recently developed evidence-based quality PEDIATRICS Volume 136, number 3, September 2015 indicators for high-acuity pediatric conditions that can be applied to ED settings where children are seen. Most of the measures they suggest are efficiency measures. Our primary goal was to study whether providing feedback to ED physicians, on factors that they control (testing and treatment decisions for common midacuity presentations for which practice guidelines are available), can affect resource use. We have previously shown wide variation in physician use of common resources in our ED; higher resource use was associated with greater LOS but did not reduce return to ED.4 In a recent study, e675 TABLE 2 Overall Preintervention and Postintervention Results for the Quality, Efficiency, and Resource Use Measures Reported on Scorecards Resource or Outcome Preintervention IV antibiotics, n (%) IV ondansetron, n (%) Abdomen or pelvis CT scan, n (%) Head CT scan, n (%) Chest radiograph, n (%) Abdominal radiograph, n (%) Laboratory tests per 100 patients, mean (SD) IV fluids, n (%) Hospital admission, n (%) LOS in minutes, median (interquartile range) 72-h return rate, n (%) 1218/10 114 562/4827 56/4827 534/2055 2470/7838 742/4827 71.13 1826/4827 1842/24 834 112 549/24 834 (12.0) (11.6) (1.2) (26.0) (31.5) (15.4) (124.9) (37.8) (7.4) (69–168) (2.2) Postintervention 937/8691 478/5937 37/5937 522/2740 2592/9248 972/5937 68.74 2292/5937 1788/26 616 108 542/26 616 (10.8) (8.1) (0.6) (19.1) (28) (16.4) (121.6) (38.6) (6.7) (66–166) (2.0) P .007 ,.001 .003 ,.001 ,.001 NS NS NS .002 ,.001 NS NS, not significant. Kharbanda et al14 observed variation in quality measures for patients presenting to 21 US pediatric EDs with common conditions and noted that higher costs (reflective of overall resource use) were not associated with lower hospitalization or ED revisit rates. Coon et al21 noted that practice variation may be an important indicator for overdiagnosis, which can lead to a number of downstream effects that do not improve patient outcomes but can cause harm. In our current study, we took advantage of variation in practice at the local level to highlight outlier practice patterns in both directions of high and low resource use. Such benchmarking of providers relative to peers can highlight opportunities for improvement; when it is done at the local level, it is likely to be more meaningful to clinicians practicing in the same setting. A unique feature of our scorecard is that it is balanced and comprehensive, not only providing data on resource use but also providing return rates as a balancing measure and LOS as an efficiency measure. We also adjusted for both acuity and patient complaint, and we used common ED conditions for which institutional guidelines were available to inform practice. We used 2 separate methods of analysis: an unadjusted preintervention–postintervention analysis to evaluate aggregate group-level differences in practice patterns and a rigorous, adjusted model to evaluate changes in true trends at the individual physician level. Whereas the former analysis can indicate whether overall resource use differed in the 2 time periods, the latter analysis can identify ongoing sustained changes in trends of practice. The latter analysis also controls for preexisting trends (eg, if resource use was already decreasing in the preintervention period) and is more robust in adjusting for physicianlevel random effects and repeated measures over time. TABLE 3 Change in Trends Before and After Intervention for Selected Resource Use Measures Reported on Scorecards Resource Use Measure Change in Trends (slope after intervention 2 slope before intervention) Interpretation of Change P Laboratory tests (per 100 patients) IV antibiotics, % Hospital admissions, % IV ondansetron, % Abdominal CT scans, % Abdominal radiographs, % 20.021 Decreasing trend in use ,.0001 20.027 0.020 20.021 0.047 0.0116 Decreasing trend in use Increasing trend in use No change No change No change .02 .02 .21 .37 .4 e676 After providing feedback for just over 15 months, we observed overall reduction in the use of several resources by using an aggregate comparison between the 2 time periods. Trend analysis showed similar reduction in resource use over time for some but not all of these categories. This reduction may have been caused by several factors. Some trends reversed in the preintervention and postintervention period; for example, trends for laboratories showed increasing use before the intervention and decreasing use after (Fig 3). Although overall use of laboratory tests did not change during the study, the change in trends reflects statistically significant change in practice (reduction in use) after the intervention (Table 3). Trends such as use of IV ondansetron and abdominal and pelvic CT scans predated the intervention (Fig 3); some of these may have resulted from overall awareness of issues related to costs of IV medication, radiation exposure of CT scans, and the like. Reduction in use of CT scans for minor head injury during 2005 to 2009 was noted in a recent study by Mannix et al.11 Our secondary analysis of trends helps elucidate long-term changes in behavior at the individual physician level and sustainability of practice changes for resources whose use is already changing. We conjecture that our intervention kept the momentum of change moving in the desired direction for most but not all resource use. For example, for admission to hospital we noted a true discrepancy between the statistically significant decrease in overall admission rates in the postintervention period and the statistically significant increase in trends in admission rates. Although it is conceivable that this increase in admission rate trends may be related to lower resource use, the decrease in overall admission rates does not JAIN et al FIGURE 3 Trends over time in rates of laboratory tests performed, IV antibiotic use, hospital admissions, IV ondansetron use, abdominal CT scans, and abdominal radiographs performed before and after introduction of physician-level scorecards. support that possibility. Use of abdominal radiographs was the only resource category that showed an actual increase in use after the intervention, with a somewhat PEDIATRICS Volume 136, number 3, September 2015 increasing trend in the postintervention period (Fig 3). Although neither the increase in overall use of abdominal radiographs nor the increase in trends is statistically significant, this change may be related to a reduction in use of abdominal CT scans, with physicians replacing 1 imaging modality with another. e677 Use of IV antibiotics showed a significant decrease after the intervention, both in overall rate and in significant change in trends (Tables 2 and 3, Fig 3). In these common, nonsevere pediatric conditions, use of antibiotics is often not indicated, and use of IV antibiotics may be a cautionary approach of some providers and therefore amenable to change in practice. Decreased use of broad-spectrum IV antibiotics would have additional downstream benefits including reducing costs, side effects, and resistance. A limitation of our study is that it is a single-center study and may not be generalizable to other settings. Given the increasing use of EMRs nationally, and of standard systems of triage acuity assignment that can be used for severity adjustment, along with nearly universal tracking of return rates and LOS in most EDs, we think scorecards similar to ours can be easily adapted for use in many EDs. Additionally, trends for resource use in some categories predated the intervention, making interpretation difficult. We did not make any other changes in the ED during the study period that would affect resource use. The EDs did get a new EMR in midJuly 2011; the only foreseeable impact of that change pertaining to this study would be an increase in LOS as providers learned a new system. We saw an increase in overall LOS for a few months after implementation of the new EMR (data available); however, over the entire study period, the LOS specific to the patients included in the scorecards showed a decrease. The modest and inconsistent effects of the scorecards in our study underscore the difficulty of changing physician behavior. It remains to be seen how other factors, such as monetary incentives and medicolegal issues, none of which changed during this study, can influence resource use in addition to comprehensive, balanced feedback to providers. e678 CONCLUSIONS Our study shows reductions in resource use for several commonly seen conditions in the pediatric ED after we provided ED physicians with feedback on practice patterns, including resource use and quality metrics, relative to peers. Reduced resource use did not adversely affect ED efficiency (LOS) or quality of care (return rate). Feedback on practice patterns relative to peers can influence provider practice for patients outside the middle acuity range and the 4 conditions we studied, thus having a broader impact. Also, feedback to physicians on clinical factors that they can control, and for which there is evidence to base practice on, may be more meaningful than quality measures that focus primarily on operational metrics. Such performance measurement can be used for activities such as the Joint Commission’s Ongoing Professional Practice Evaluation. With additional refinement, comprehensive, severityadjusted performance measures that encompass both quality and resource use can be used to measure value at the individual provider level and to identify high-value providers. ACKNOWLEDGMENTS We thank Robert Massey and Seton McRae for data acquisition and management. ABBREVIATIONS CT: computed tomography ED: emergency department EMR: electronic medical record IV: intravenous LOS: length of stay REFERENCES 1. Goldin AB, Garrison M, Christakis D. Variations between hospitals in antireflux procedures in children. Arch Pediatr Adolesc Med. 2009;163(7): 658–663 2. Goldman RD, Scolnik D, Chauvin-Kimoff L, et al Practice variations in the treatment of febrile infants among pediatric emergency physicians. Pediatrics. 2009; 124(2):439–445 3. Hampers LC, Faries SG. Practice variation in the emergency management of croup. Pediatrics. 2002;109(3):505–508 4. Jain S, Elon LK, Johnson BA, Frank G, Deguzman M. Physician practice variation in the pediatric emergency department and its impact on resource use and quality of care. Pediatr Emerg Care. 2010;26(12):902–908 5. Landrigan CP, Conway PH, Stucky ER, Chiang VW, Ottolini MC. Variation in pediatric hospitalists’ use of proven and unproven therapies. J Hosp Med. 2008; 3(4):292–298 6. Lantos JD, Meadow W. Variation in the treatment of infants born at the borderline of viability. Pediatrics. 2009; 123(6):1588–1590 7. Powell EC, Hampers LC. Physician variation in test ordering in the management of gastroenteritis in children. Arch Pediatr Adolesc Med. 2003;157(10):978–983 8. Song Y, Skinner J, Bynum J, Sutherland J, Wennberg JE, Fisher ES. Regional variations in diagnostic practices. N Engl J Med. 2010;363(1):45–53 9. Tieder JS, Cowan CA, Garrison MM, Christakis DA. Variation in inpatient resource utilization and management of apparent life threatening events. J Pediatr. 2008;152(5):629–635, e621–e622 10. Wennberg JE. Variation in use of Medicare services in regions and selected academic medical centers: is more better? 2005. Available at: http:// www.commonwealthfund.org/ publications/fund-reports/2005/dec/ variation-in-use-of-medicare-servicesamong-regions-and-selected-academicmedical-centers–is-more-b. Accessed October 28, 2010 11. Mannix R, Meehan WP, Monuteaux MC, Bachur RG. Computed tomography for minor head injury: variation and trends in major United States pediatric emergency departments. J Pediatr. 2012; 160(1):136–139, e131 JAIN et al 12. Tieder JS, McLeod L, Keren R, et al Variation in resource use and readmission for diabetic ketoacidosis in children’s hospitals. Pediatrics. 2013; 132(2):229–236 13. Fisher ES, Wennberg DE, Stukel TA, Gottlieb DJ, Lucas FL, Pinder EL. The implications of regional variations in Medicare spending. Part 2: health outcomes and satisfaction with care. Ann Intern Med. 2003;138(4):288–298 14. Kharbanda AB, Hall M, Shah SS, et al. Variation in resource utilization across a national sample of pediatric emergency departments. J Pediatr. 2013;163(1):230–236 15. Florin TA, French B, Zorc JJ, Alpern ER, Shah SS. Variation in emergency department diagnostic testing and disposition outcomes in pneumonia. Pediatrics. 2013;132(2):237–244 16. Knapp JF, Simon SD, Sharma V. Variation and trends in ED use of radiographs for asthma, bronchiolitis, and croup in children. Pediatrics. 2013;132(2):245–252 17. Choosing Wisely: an initiative of the ABIM Foundation. 2013. Available at: www. choosingwisely.org. Accessed March 15, 2014 18. Guttmann A, Razzaq A, Lindsay P, Zagorski B, Anderson GM. Development of measures of the quality of emergency department care for children using a structured panel process. Pediatrics. 2006;118(1):114–123 19. Knapp JF, Simon SD, Sharma V. Quality of care for common pediatric respiratory illnesses in United States emergency departments: analysis of 2005 National Hospital Ambulatory Medical Care Survey Data. Pediatrics. 2008;122(6):1165–1170 20. Stang AS, Straus SE, Crotts J, Johnson DW, Guttmann A. Quality indicators for high acuity pediatric conditions. Pediatrics. 2013;132(4):752–762 21. Coon ER, Quinonez RA, Moyer VA, Schroeder AR. Overdiagnosis: how our PEDIATRICS Volume 136, number 3, September 2015 compulsion for diagnosis may be harming children. Pediatrics. 2014; 134(5):1013–1023 22. Jamtvedt G, Young JM, Kristoffersen DT, O’Brien MA, Oxman AD. Does telling people what they have been doing change what they do? A systematic review of the effects of audit and feedback. Qual Saf Health Care. 2006;15(6):433–436 23. Forrest CB, Fiks AG, Bailey LC, et al. Improving adherence to otitis media guidelines with clinical decision support and physician feedback. Pediatrics. 2013; 131(4). Available at: www.pediatrics.org/ cgi/content/full/131/4/e1071 24. Ahwah I, Karpiel M. Using profiling for cost and quality management in the emergency department. Healthc Financ Manage. 1997;51(7):48, 50–53 25. Grant L, Proctor T. Measuring quality: how to empower staff to take control. Nurs Times. 2011;107(7):22–25 26. Kruskal JB, Reedy A, Pascal L, Rosen MP, Boiselle PM. Quality initiatives: lean approach to improving performance and efficiency in a radiology department. Radiographics. 2012;32(2):573–587 27. Meliones JN, Alton M, Mericle J, et al 10-year experience integrating strategic performance improvement initiatives: can the Balanced Scorecard, Six Sigma (R), and Team Training all thrive in a single hospital? In: Henriksen K, Battles JB, Keyes MA, Grady ML, eds. Advances in Patient Safety: New Directions and Alternative Approaches (Vol. 3: Performance and Tools). Rockville, MD: Agency for Healthcare Research and Quality (US); 2008 28. Wicks AM, St Clair L. Competing values in healthcare: balancing the (un)balanced scorecard. J Healthc Manage. 2007;52(5): 309–323; discussion 323–324 29. Durani Y, Brecher D, Walmsley D, Attia MW, Loiselle JM. The Emergency Severity Index Version 4: reliability in pediatric patients. Pediatr Emerg Care. 2009; 25(11):751–753 30. Gilboy N, Tanabe T, Travers D, Rosenau AM. Emergency Severity Index (ESI): A Triage Tool for Emergency Department Care, Version 4. Implementation Handbook 2012 Edition. Rockville, MD: Agency for Healthcare Research and Quality; 2011 31. Pitts S, Niska R, Xu J, Burt C. National Hospital Ambulatory Medical Care Survey: 2006 emergency department summary. National Health Stat Rep. 2008; 7:1–38 32. SAS Institute Inc. SAS/STAT 9.2 User’s Guide. Cary, NC: SAS Institute Inc; 2008 33. Hung GR, Chalut D. A consensusestablished set of important indicators of pediatric emergency department performance. Pediatr Emerg Care. 2008; 24(1):9–15 34. Welch SJ, Asplin BR, Stone-Griffith S, Davidson SJ, Augustine J, Schuur J; Emergency Department Benchmarking Alliance. Emergency department operational metrics, measures and definitions: results of the Second Performance Measures and Benchmarking Summit. Ann Emerg Med. 2011;58(1):33–40 35. Welch SJ, Stone-Griffith S, Asplin B, Davidson SJ, Augustine J, Schuur JD; Second Performance Measures and Benchmarking Summit; Emergency Department Benchmarking Alliance. Emergency department operations dictionary: results of the Second Performance Measures and Benchmarking Summit. Acad Emerg Med. 2011;18(5):539–544 36. Pham JC, Kirsch TD, Hill PM, DeRuggerio K, Hoffmann B. Seventy-two-hour returns may not be a good indicator of safety in the emergency department: a national study. Acad Emerg Med. 2011;18(4): 390–397 e679
© Copyright 2026 Paperzz