Implementation of a Preventive Services Bundle in Academic Pediatric Primary Care Centers Zeina Marcho Samaan, MD, Courtney M. Brown, MD, MS, John Morehous, MD, Alison A. Perkins, BA, Robert S. Kahn, MD, MPH, Mona E. Mansour, MD, MS BACKGROUND AND OBJECTIVES: Previous studies have documented poor rates of delivery of preventive services, 1 of the core services provided in the primary care medical home setting. We aimed to increase the reliability of delivering a bundle of preventive services to patients 0 to 14 months of age from 58% of patient visits to 95% of visits. The bundle includes administration of routine vaccinations, offering influenza vaccination, completed lead screening, completed developmental screening tool, screening for maternal depression and food insecurity, and documentation of gestational age. METHODS: The setting was 3 academic pediatric primary care clinics that serve 31 000 patients (>90% Medicaid). Quality improvement methodology was used and key driver diagram was determined. Patient “Ideal Visit Flow” and the Responsible, Accountable, Support, Consulted, and Informed Matrix were developed to drive accountability for components of the ideal flow. Plan, Do, Study, Act cycles were used to develop successful interventions. The percent of patients seen who received all bundle elements for which they were eligible was plotted weekly on a run chart, and statistical process control methods were used to determine a significant change in performance. RESULTS: The preintervention percentage of patient visits ages 0 to 14 months receiving all preventive service bundle elements was 58%. The postintervention percentage is 92%. CONCLUSIONS: Innovative redesign led to improvement in percentage of patients age 0 to 14 months who received the entire preventive services bundle. Key elements for success were multidisciplinary site-specific teams, redesigned visit flow, effective communication, and resources for data and project management. More than 10 years ago, the Institute of Medicine called attention to the suboptimal performance of the United States health care system, concluding: “Current care systems cannot do the job. Trying harder will not work. Changing care systems will.”1 In pediatrics, a 2007 publication revealed that patients receive only 41% of indicated preventive services.2 Past efforts to improve quality in pediatric primary care have focused mostly on individual services, such as improving immunization rates3 or lead screening.4–6 Yet, pediatric practices continue to struggle with comprehensively delivering preventive services in light of lengthy Downloaded from by guest on July 31, 2017 PEDIATRICS Volume 137, number 3, March 2016:e20143136 abstract Division of General and Community Pediatrics, Cincinnati Children’s Hospital Medical Center, Cincinnati, Ohio Dr Samaan was involved in conceptualization and design of the study, data analysis, and drafted the initial manuscript; Drs Brown, Morehous, Kahn, and Mansour also contributed to the conceptualization and design of the study, data analysis, and reviewed and revised the manuscript; Ms Perkins designed the electronic data retrieval program and analyzed the data; and all authors approved the final manuscript as submitted. DOI: 10.1542/peds.2014-3136 Accepted for publication Jul 13, 2015 Address correspondence to Zeina Marcho Samaan, MD, Cincinnati Children’s Hospital Medical Center, ML 2011, 3333 Burnet Ave, Cincinnati, OH 45229. E-mail: [email protected] PEDIATRICS (ISSN Numbers: Print, 0031-4005; Online, 1098-4275). Copyright © 2016 by the American Academy of Pediatrics FINANCIAL DISCLOSURE: The authors have indicated they have no financial relationships relevant to this article to disclose. FUNDING: This study was conducted with internal funding from the Cincinnati Children’s Hospital Medical Center and the Division of General and Community Pediatrics. POTENTIAL CONFLICT OF INTEREST: The authors have indicated they have no potential conflicts of interest to disclose. To cite: Samaan ZM, Brown CM, Morehous J, et al. Implementation of a Preventive Services Bundle in Academic Pediatric Primary Care Centers. Pediatrics. 2016;137(3):e20143136 QUALITY REPORT health maintenance guidelines and short visits.7–12 The challenges of delivering comprehensive well child care underscore the need for true system transformation rather than intense focus on a single measure. In other areas of medicine, “bundle measures” have been used to measure whether every patient receives every service every time.13–15 We identified 1 previous study of a bundle measure in pediatric primary care. This bundle included services related only to physical health (immunization, lead screening, anemia, and tuberculosis) and was implemented in nonacademic practices.16 This 2004 study demonstrated that it is possible to improve care across multiple measures through system improvement. However, changes in the scope of well child care in the past decade call for a more broadly defined measure that includes services beyond screening for physical illness and exposures.17 Locally, our medical center was redesigning primary care, with a global aim to improve health promotion, chronic disease outcomes, and patient and family experience, while decreasing cost of care at a population level. The intervention described in this article focuses on our goal to improve health promotion. The scope of this quality improvement (QI) study was to improve delivery of preventive services, including assessment of development and family well-being, during visits at our 3 primary care centers in the first 14 months of life. We aimed to measure and improve a “preventive service bundle” of 5 indicators of promoting an infant’s physical health, cognitive and social development, and family well-being. We used an all-or-none bundle measure to obtain a more accurate picture of the quality of care that individual patients received18 during office visits. We theorized that, e2 by monitoring performance on a bundle measure and using the data to drive small tests of change, we could optimize performance during each stage of an office visit (eg, registration, intake, order entry). We could then clearly define roles, standardize processes, and transform our system to achieve highly reliable preventive service delivery.19 We aimed to improve the percentage of visits at which patients received all bundle elements for which they were eligible from 58% to 95% within 1 year. METHODS Setting Cincinnati Children’s Hospital Medical Center (CCHMC) has 3 primary care centers involved in primary care redesign. Sites 1 and 2 (urban) and site 3 (suburban) are the medical homes for ~18 000, 7000, and 6000 patients, with ~37 000, 14 000, and 15 000 visits, respectively, per year. The payer mix is uniform across the centers: 90% Medicaid, 3% private insurance, and 7% self-pay/uninsured. All 3 sites train large number of learners. All sites have been using Epic, an electronic health record (EHR), since May 2011. In 2012, CCHMC and the Division of General and Community Pediatrics invested $250 000 to conduct a large-scale primary care redesign that aimed to improve health promotion and prevention, outcomes for children with chronic diseases, patient and family experience, and decrease cost of care. This article focuses on 1 phase of the redesign and describes the QI tools and processes used to improve health promotion and prevention for infants in our patient population. This study was granted exemption by the CCHMC Institutional Review Board. Study Population and Outcome We sought to design a comprehensive, efficient, and Downloaded from by guest on July 31, 2017 effective preventive service delivery system. We focused on the first 14 months of life when opportunities are greatest for engaging families in primary care.20 The following services were selected for inclusion in the bundle of preventive services that all infants should receive during the first 14 months of life: administration of routine immunizations, seasonal influenza vaccination offer, lead screening, standardized developmental screening using Ages and Stages Questionnaire, and screening for bio-psychosocial risk factors including gestational age, parental depression, and food insecurity. These services were chosen by the primary care redesign steering committee, explained below, based on American Academy of Pediatrics recommendations,21 with minor modifications. Team Structure A team structure was developed to support all phases of the primary care redesign. Some of the team’s time focused on preventive services for infants. The team included the following: (1) a steering committee that provided overall guidance and consisted of a project sponsor, project leader, physicians, patient services manager, business director, QI consultant, and advisors; (2) a project manager that guided resources to achieve specific goals; (3) support teams that provided content expertise on family engagement, training, finance, resident education, QI data analysis, information technology (IT), and communication; and (4) sitebased teams at each of 3 primary care clinics spearheaded the work while participating in a hands-on QI training program. Participating physicians did not have extra protected time. Nurses and medical assistants (MAs) received payment for the hours they spent in these planning and training activities. At SAMAAN et al a staff kick-off meeting, the project vision, mission, and goals were reviewed and feedback was collected. An office system inventory22 was used to provide a baseline assessment of our preventive services delivery system. Our redesign aligned with the National Committee for Quality Assurance/ Patient Centered Medical Home requirements.23 Development of Key Driver Diagram and Interventions Improvement work began in June 2012. Preventive service support and site-based teams met throughout the planning and implementation phases to develop a theory of change, plan tests of change, and review data. The Model for Improvement was used.24 This model is based on 3 fundamental questions: (1) What are we trying to accomplish? (2) How will we know that a change is an improvement? (3) What changes can we make that will result in improvement? The team organized its theory of change by using a key driver diagram25 (Fig 1). Interventions were tested by using Plan, Do, Study, Act cycles.24 Optimizing patient visit flow was expected to improve preventive service bundle delivery. The project manager, consultant, and site-based teams mapped the detailed existing patient visit flow for each clinic. Clinic staff roles and responsibilities were identified by using a Responsible, Accountable, Support, Consulted, and Informed (RASCI) Matrix. RASCI is an acronym derived from the 5 responsibilities most typically used: Responsible, Accountable, Support, Consulted, and Informed. This matrix describes the participation by various roles in completing tasks26 and allows for prioritization of work. The current process map and RASCI matrix were used to develop ideal patient visit flow with the goal of optimizing each care team member’s role to maximize individuals’ training and skills. Roles were redefined PEDIATRICS Volume 137, number 3, March 2016 FIGURE 1 Key driver diagram. aPreventive services bundle includes routine immunizations, offering flu vaccine, blood lead screening, standardized developmental screening, and screening for risk factors (gestational age, food insecurity, parental depression). SMART, Specific Measurable Actionable Relevant Time-bound. and reassigned, and Plan, Do, Study, Act cycles tested integrating reassignments into visit flow. In addition, the IT team optimized the EHR to display patients’ preventive service needs on the patient schedule. This information was referenced by providers and staff during relevant steps in the ideal flow. Several key changes were made to visit flow. First, 2 types of “huddles” were implemented. The nurse manager huddled with nurses and MAs in the morning to discuss potential flow challenges that day. In addition, at the beginning of each morning, afternoon, and evening clinic session, each provider huddled with their assigned MA to review the preventive service needs of each scheduled patient. Second, registration staff preassembled standardized, age-specific packets that included developmental and social screening forms.27,28 These were distributed at registration for parents to complete before rooming. Third, much of the informationgathering and EHR documentation for well visits was shifted from providers to intake staff. This ensured screening questions were documented and allowed physicians to focus the visit on areas of concern. Downloaded from by guest on July 31, 2017 Fourth, intake staff preordered immunizations and screening tests for providers to review and sign. Finally, the patient discharge process was standardized and included a staff member checking that all appropriate preventive services had been delivered. Clinic start time was not changed, and the number of visit was not reduced during ideal flow implementation. Methods of Evaluation A measure of the reliability of daily preventive service delivery was created, with the visit as the unit of analysis. This measure calculated the percentage of 0- to 14-monthold visits during which the patient received all elements of the bundle for which they were eligible that day. More details on the measure are provided in a separate publication.29 Figure 2 summarizes expectations for preventive service delivery at each age. For example, any visit for a 9-month-old patient would require administration of any overdue immunizations, developmental screening, lead testing, and influenza vaccine offered if seasonally appropriate. Although ill visits were included in our study, if a provider decided to defer immunizations e3 through visits and collect feedback that was used to improve processes. As a balancing measure, patient cycle time (time from patient sign-in to discharge) was collected pre- and postinterventions. RESULTS Over an 8-week baseline period (June through July 2012), the entire bundle of preventive services for which patients age 0 to 14 months were eligible was delivered at 58% of visits. Postintervention results revealed a median performance of 92%, achieved by May 2013 and sustained over 1 year (Fig 3). FIGURE 2 Visit-level measure scoring. aGestational age and screening for parental depression and food insecurity documented in EHR, if not already in patient chart. bMust get all listed immunizations. If patient did not get vaccine, check the visit date and count as a “yes” if the patient got the vaccines indicated due to the date of the visit and recommended interval from last dose of vaccine. cRotavirus #1 rule: If patient did not get vaccine but age is ≥ 106 days, score as yes (too old to start the series). dRotavirus #2 rule: If patient received first rotavirus vaccine but did not get second rotavirus vaccine and age is ≥ 224 days, score as yes (too old to get second dose). eThe American Academy of Pediatrics recommends screening at 9 months. Given low adherence to well child schedule in this patient population, clinic processes include screening at any well visit occurring between 6 and 14 months of age to ensure developmental delays are detected in a timely way for children who miss a well visit. Measure includes at least 1 completed screen. fLead screening to be completed once between 9 and 12 months of age. gRefusals count as yes if appropriately documented. ASQ, Ages and Stages Questionnaire; DTaP, diphtheria-tetanus-acellular pertussis, Hib, Haemophilus influenzae type b; IPV, inactivated polio vaccine; MMR, measles, mumps, rubella; PCV, pneumococcal vaccine. because of acute illness and documented this with an appropriate diagnosis code, this was not counted as a failure. Performance data on bundle delivery were retrieved directly from clinical documentation in the EHR. Reports on this measure were constructed through an iterative process. An automated data report was initially formulated, then compared with manual chart reviews to ensure interrater reliability and to identify unique situations for which decision rules needed to be created. The automated report was then modified and the process was repeated. In its final form, the report was automated with quality checks done manually by a data manager who followed a list of decision rules. e4 Daily data were collated into weekly rates and plotted on a run chart. Standard probability-based rules were used for interpretation of run charts.30 In accordance with these rules, the median was recalculated when a “shift” of 8 consecutive points above or below the median was observed. Initially, chart audits were performed by the data manager for all failures, and appropriate timely feedback was provided to staff and providers. Weekly data were shared with all. Throughout the QI process, qualitative data were collected from front-line staff regarding barriers to implementing interventions and impact of interventions on clinic flow, patient care, and resident education. QI consultants were on-site during implementation to shadow patients Downloaded from by guest on July 31, 2017 Interventions fell into 2 major categories: structural and functional. Dates of intervention implementation are annotated on the run chart (Fig 3). Structural interventions included team development, shared vision, expert consultation, QI training, divisional support, and project and data management. Functional interventions included prioritization of work, optimizing the EHR, previsit planning, improving communication, and implementation of ideal visit flow. Interventions were temporally associated with shifts on the run chart (Fig 3). Cycle time remained the same pre and postinterventions (average 75 minutes). Qualitative data collected from frontline staff indicated that early in our testing, a majority of failures were due to family refusal of preventive services or failure to document services that were completed or deferred due to acute patient illnesses. In addition, more failures happened on evenings or weekends, and when transient providers worked in our clinics. There was no difference in failures by age group. Failures were addressed by having QI consultants coach providers on proper documentation, observe and evaluate flow, and provide feedback and recommendations to optimize the ideal flow map. A SAMAAN et al separate in-depth study looked at the effects of primary care redesign on residents’ educational experience. It is presented in a separate publication currently under review. Overall, residents had improved perceptions of patient flow and physician/ nonphysician teamwork. Our bundle delivery rate was sustained at a median of 92% over a 1-year period. We planned for sustainability by developing a standardized orientation and training process for all newly hired staff and providers. We also performed frequent assessment and retraining of existing staff and providers to ensure the ideal flow protocol was being followed. DISCUSSION Between June 2012 and June 2013, we increased delivery of a preventive services bundle from 58% of visits to 92% of visits for patients 0 to 14 months of age, and we sustained the improvement for over 1 year. Interventions that coincided temporally with improvement in bundle delivery included prioritization of work, previsit planning, preclinic huddles, and implementation of ideal flow. Through improvement on the bundle measure, we now reliably address infants’ physical health, cognitive and social development, and family wellbeing during primary care visits. We have also transformed our system to create effective mechanisms for communication among staff members and more failsafe processes for identifying patients’ health maintenance needs. Our study was strengthened by rigorous use of QI methods, with on-site QI consultants who ensured fidelity to the interventions. We maximized the sustainability of our intervention by studying our system under various conditions, including during evening and weekend hours and clinic sessions staffed by PEDIATRICS Volume 137, number 3, March 2016 FIGURE 3 Percent of primary care patients seen 14 months and younger who have received the entire bundle of preventive services for which they are eligible. transient providers (moonlighters and learners). By collecting performance data directly from the EHR, we were able to include all visits in our data. Because we audited charts and coached providers on proper documentation, we are confident that our data closely reflect actual preventive service delivery. Our findings support those of other studies in which system-based improvements resulted in improved preventive service delivery. Shaw et al6 and Young et al31 showed that practices who set goals and used QI methods achieved improvement in at least 1 preventive service. In accordance with recommendations by Solberg et al,32 we used strategies that considered multiple characteristics of participating clinics. Our findings were consistent with those of Bordley et al,5 who achieved improvement in immunization rates, anemia, and lead screening through implementation of previsit planning, risk assessment forms, provider prompts, and redistribution of responsibilities among office staff. Downloaded from by guest on July 31, 2017 Given findings by Shojania et al33 that point-of-care EHR reminders alone achieved small effects, and Lanham et al34 that relationships and respect among office staff contribute to primary care quality and success of improvement initiatives, our initial structural interventions (eg, formation of site-based teams with nonphysician team leaders) likely played a key role in laying the groundwork for our success. To our knowledge, ours was the second study to measure delivery of a “bundle” of preventive services in pediatric primary care. Margolis et al16 measured a bundle of 4 preventive services (immunizations, and screening for tuberculosis, anemia, and lead) at 18 nonacademic practices and achieved improvement from 7% to 34%. Because Margolis et al16 used a patient-level measure rather than a visit-level measure, we cannot directly compare our success. However, our findings build upon the work by Margolis et al16 by expanding the bundle to include services related to developmental e5 and social well-being and also demonstrate that it is possible to achieve significant sustainable improvement in large academicaffiliated pediatric practices. Expansion to this setting is important in ensuring high quality care for low-income children, given the large percentage of publicly insured patients who are served by academicaffiliated safety net hospitals.35 Our study had some limitations. Because we did not consider decisions to defer immunizations because of illness a failure to deliver indicated services, our data do not capture potential missed opportunities to vaccinate in situations where the patient is ill but immunization deferral is not clinically indicated.36 Additionally, because we are no longer performing chart audits on all failures, it is possible that some preventive services are being delivered but not documented; therefore, our actual performance may be better than what is shown. In considering whether our findings are generalizable to other practices, we acknowledge context-specific factors that may have contributed to our success including the financial support for external consultation, data analysis, a project manager, QI consultants, and QI methodology training for front-line staff. We believe these resources were integral to our success given the size and complexity of our settings. However, because the fund was not entirely used for this particular phase of the redesign and the support staff had multiple other job responsibilities, it is difficult to quantify the exact cost of our preventive service improvement. It is notable that our improvements were achieved without hiring extra staff or reducing patient volume. We believe that small and less complex practices can learn from our experience to implement some changes with minimal project management resources. Universal e6 strategies from our work that may be applied without a lot of further testing include the following: (1) developing EHR-based prompts about services due; (2) creating standardized packets for each patient age with appropriate developmental and social screening forms,27,28 which can be distributed by registration staff and completed by parents during waiting time; (3) shifting some history-taking and EHR documentation to intake staff to maximize the time and skills of all clinical team members. Also, new health care models with permember per-month payment systems may soon incentivize investment in personnel to facilitate improvement work and quality measurement because funding these activities may provide returns on investment by improving outcomes for patients for which the health system is accountable. Reaching our original goal of 95% would require higher reliability interventions, such as increased automation of tasks.32 We will continue to work closely with our IT team to implement these types of changes. Future directions include improving preventive service delivery on a population level and measuring preventive service bundles for other patient age groups. CONCLUSIONS Using QI methods, we achieved and sustained a preventive service bundle delivery rate of 92% of primary care visits for patients 0 to 14 months of age. The bundle included administration of all routine immunizations, offering influenza vaccine to eligible infants, and screening for lead, developmental delay, gestational age, food insecurity, and parental depression at appropriate intervals. Our bundle is the first, to our knowledge, to include preventive services in multiple domains of child Downloaded from by guest on July 31, 2017 and family well-being. Our study demonstrates the dramatic increase in reliability that is possible with system transformation, even in complex settings serving low-income populations. ABBREVIATIONS CCHMC: Cincinnati Children’s Hospital Medical Center EHR: electronic health record IT: information technology MA: medical assistant QI: quality improvement RASCI: Responsible, Accountable, Support, Consulted, and Informed REFERENCES 1. Crossing the Quality Chasm. A New Health System for the 21st Century. Washington, DC: The National Academies Press; 2001 2. Mangione-Smith R, DeCristofaro AH, Setodji CM, et al. The quality of ambulatory care delivered to children in the United States. N Engl J Med. 2007;357(15):1515–1523 3. Fiks AG, Grundmeier RW, Biggs LM, Localio AR, Alessandrini EA. Impact of clinical alerts within an electronic health record on routine childhood immunization in an urban pediatric population. Pediatrics. 2007;120(4):707–714 4. Meyer PA, Staley F, Staley P, Curtis J, Blanton C, Brown MJ. Improving strategies to prevent childhood lead poisoning using local data. Int J Hyg Environ Health. 2005;208(1–2):15–20 5. Bordley WC, Margolis PA, Stuart J, Lannon C, Keyes L. Improving preventive service delivery through office systems. Pediatrics. 2001;108(3). Available at: www.pediatrics.org/cgi/ content/full/108/3/e41 6. Shaw JS, Wasserman RC, Barry S, et al. Statewide quality improvement outreach improves preventive services for young children. Pediatrics. 2006;118(4). Available at: www. SAMAAN et al pediatrics.org/cgi/content/full/118/4/ e1039 7. Boyle CA, Perrin JM, Moyer VA. Use of clinical preventive services in infants, children, and adolescents. JAMA. 2014;312(15):1509–1510 8. Ayres CG, Griffith HM. Perceived barriers to and facilitators of the implementation of priority clinical preventive services guidelines. Am J Manag Care. 2007;13(3):150–155 9. Cabana MD, Rand CS, Powe NR, et al. Why don’t physicians follow clinical practice guidelines? A framework for improvement. JAMA. 1999;282(15):1458–1465 10. Cochrane LJ, Olson CA, Murray S, Dupuis M, Tooman T, Hayes S. Gaps between knowing and doing: understanding and assessing the barriers to optimal health care. J Contin Educ Health Prof. 2007;27(2):94–102 11. Belamarich PF, Gandica R, Stein RE, Racine AD. Drowning in a sea of advice: pediatricians and American Academy of Pediatrics policy statements. Pediatrics. 2006;118(4). Available at: www.pediatrics.org/cgi/content/full/ 118/4/e964 12. Uner H, Nezami FG, Yildirim MB, Dong F, Wellner J, Bradham DD. Visit length in pediatric primary care: lessons from a pilot study. J Med Pract Manage. 2013;28(6):363–370 13. Bigham MT, Amato R, Bondurrant P, et al Ventilator-associated pneumonia in the pediatric intensive care unit: characterizing the problem and implementing a sustainable solution. J Pediatr. 2009;154:582–587 14. Guerin K, Wagner J, Rains K, Bessesen M. Reduction in central lineassociated bloodstream infections by implementation of a postinsertion care bundle. Am J Infect Control. 2010;38(6):430–433 Practice based education to improve delivery systems for prevention in primary care: randomised trial. BMJ. 2004;328(7436):388 17. Committee on Psychosocial Aspects of Child and Family Health. American Academy of Pediatrics. The new morbidity revisited: a renewed commitment to the psychosocial aspects of pediatric care. Pediatrics. 2001;108(5):1227–1230 18. Nolan T, Berwick DM. All-ornone measurement raises the bar on performance. JAMA. 2006;295(10):1168–1170 19. Chassin MR, Loeb JM. High-reliability health care: getting there from here. Milbank Q. 2013;91(3):459–490 20. Luman ET, Chu SY. When and why children fall behind with vaccinations: missed visits and missed opportunities at milestone ages. Am J Prev Med. 2009;36(2):105–111 21. Bright Futures/American Academy of Pediatrics. Recommendations for preventive pediatric healthcare. Available at: http://brightfutures. aap.org/pdfs/AAP%20Bright%20 Futures%20Periodicity%20Sched%20 101107.pdf. Accessed July 23, 2015 22. Commonwealth Fund. Office Systems Inventory. Available at: www.commonwealthfund.org/usr_ doc/office_systems_inventory.pdf. Accessed July 23, 2015 23. The National Committee for Quality Assurance- NCQA's Patient-Centered Medical Home (PCMH)2011 http:// www.ncqa.org/tabid/631/Default.aspx National Academic Press, Washington, DC 2011 24. Langley G, Nolan K, Norman C, Provost L, Nolan T. The Improvement Guide: A Practical Approach to Enhancing Organizational Performance. 2nd ed. New York, NY: Jossey-Bass; 1996 15. Shubrook JH Jr, Snow RJ, McGill SL, Brannan GD. “All-or-none” (bundled) process and outcome indicators of diabetes care. Am J Manag Care. 2010;16(1):25–32 25. Chung PJ, Baum RA, Soares NS, Chan E. Introduction to quality improvement part one: defining the problem, making a plan. J Dev Behav Pediatr. 2014;35(7):460–466 16. Margolis PA, Lannon CM, Stuart JM, Fried BJ, Keyes-Elstein L, Moore DE Jr. 26. Project Management Institute. A Guide to Project Management Body of PEDIATRICS Volume 137, number 3, March 2016 Downloaded from by guest on July 31, 2017 Knowledge (PMBOK Guide) Newtown Square, PA 2010 27. Beck AF, Klein MD, Kahn RS. Identifying social risk via a clinical social history embedded in the electronic health record. Clin Pediatr (Phila). 2012;51(10):972–977 28. Squires J, Bricker D, Potter L. Revision of a parent-completed development screening tool: Ages and Stages Questionnaires. J Pediatr Psychol. 1997;22(3):313–328 29. Brown CM, Samaan ZM, Morehous JF, Perkins AA, Kahn RS, Mansour ME. Development of a bundle measure for preventive service delivery to infants in primary care. J Eval Clin Pract. 2015 30. Perla RJ, Provost LP, Murray SK. The run chart: a simple analytical tool for learning from variation in healthcare processes. BMJ Qual Saf. 2011;20(1):46–51 31. Young PC, Glade GB, Stoddard GJ, Norlin C. Evaluation of a learning collaborative to improve the delivery of preventive services by pediatric practices. Pediatrics. 2006;117(5):1469–1476 32. Solberg LI, Brekke ML, Fazio CJ, et al. Lessons from experienced guideline implementers: attend to many factors and use multiple strategies. Jt Comm J Qual Improv. 2000;26(4):171–188 33. Shojania KG, Jennings A, Mayhew A, Ramsay C, Eccles M, Grimshaw J. Effect of point-of-care computer reminders on physician behaviour: a systematic review. CMAJ. 2010;182(5):E216–E225 34. Lanham HJ, McDaniel RR Jr, Crabtree BF, et al. How improving practice relationships among clinicians and nonclinicians can improve quality in primary care. Jt Comm J Qual Patient Saf. 2009;35(9):457–466 35. Cunningham P, May J. Medicaid Patients Increasingly Concentrated Among Physicians; The Center of Studying Health System Change (HSC) Tracking Report No. 16, 2006 36. Szilagyi PG, Rodewald LE. Missed opportunities for immunizations: a review of the evidence. J Public Health Manag Pract. 1996;2(1):18–25 e7 Implementation of a Preventive Services Bundle in Academic Pediatric Primary Care Centers Zeina Marcho Samaan, Courtney M. Brown, John Morehous, Alison A. Perkins, Robert S. Kahn and Mona E. Mansour Pediatrics 2016;137;; originally published online February 17, 2016; DOI: 10.1542/peds.2014-3136 Updated Information & Services including high resolution figures, can be found at: /content/137/3/e20143136.full.html References This article cites 27 articles, 6 of which can be accessed free at: /content/137/3/e20143136.full.html#ref-list-1 Subspecialty Collections This article, along with others on similar topics, appears in the following collection(s): Administration/Practice Management /cgi/collection/administration:practice_management_sub Quality Improvement /cgi/collection/quality_improvement_sub Preventive Medicine /cgi/collection/preventative_medicine_sub Permissions & Licensing Information about reproducing this article in parts (figures, tables) or in its entirety can be found online at: /site/misc/Permissions.xhtml Reprints Information about ordering reprints can be found online: /site/misc/reprints.xhtml PEDIATRICS is the official journal of the American Academy of Pediatrics. A monthly publication, it has been published continuously since 1948. PEDIATRICS is owned, published, and trademarked by the American Academy of Pediatrics, 141 Northwest Point Boulevard, Elk Grove Village, Illinois, 60007. Copyright © 2016 by the American Academy of Pediatrics. All rights reserved. Print ISSN: 0031-4005. Online ISSN: 1098-4275. Downloaded from by guest on July 31, 2017 Implementation of a Preventive Services Bundle in Academic Pediatric Primary Care Centers Zeina Marcho Samaan, Courtney M. Brown, John Morehous, Alison A. Perkins, Robert S. Kahn and Mona E. Mansour Pediatrics 2016;137;; originally published online February 17, 2016; DOI: 10.1542/peds.2014-3136 The online version of this article, along with updated information and services, is located on the World Wide Web at: /content/137/3/e20143136.full.html PEDIATRICS is the official journal of the American Academy of Pediatrics. A monthly publication, it has been published continuously since 1948. PEDIATRICS is owned, published, and trademarked by the American Academy of Pediatrics, 141 Northwest Point Boulevard, Elk Grove Village, Illinois, 60007. Copyright © 2016 by the American Academy of Pediatrics. All rights reserved. Print ISSN: 0031-4005. Online ISSN: 1098-4275. Downloaded from by guest on July 31, 2017
© Copyright 2026 Paperzz