Critical appraisal of (Systematic review) Meta-analysis 羅政勤 彰化秀傳紀念醫院 Objectives 1. To understand the different terminology of Meta-analysis, systematic review, 2. To understand the key criteria for critical appraisal 3. To select an appropriate checklist or other instrument to use for critical appraisal. Validity, Impact, Practicability (CASP) Terminology Review: ≧2 publication synthesise results + conclusions Overview(systematic literature review): a review strives to comprehensively identify and track down all literature on a given topic Meta-analysis: Specific statistical strategy assembling results of several studies into a single estimate Introduction Systematic reviews form a potential method for overcoming the barriers faced by clinicians when trying to access and interpret evidence to inform their practice Systematic reviews Concise summaries of best available evidence that addresses defined questions scientific tool used to appraise, summarise, and communicate results and implications of otherwise unmanageable quantities of research Systematic reviews Defining a question A good question will have four components: –Type of person involved –Type of exposure –Type of control –Outcomes SR and Meta-analysis Systematic reviews may or may not include a statistical synthesis called meta-analysis, whether the studies are similar enough so that combining their results is meaningful Meta-analysis Statistical method for combining the results of trials Most appropriate for randomized trials May also be appropriate for observational studies Results of a metaanalysis Forest plots of a meta-analysis of four randomized trials comparing no adjuvant chemotherapy with adjuvant chemotherapy in early-stage ovarian cancer for overall survival (A) and recurrence free survival (B). JNCI Cancer Spectrum 95(2):105-112 Advantages of metaanalysis Allows pooling of several studies = increase sample size Gathers literature in one place Provides a quantitative summary (possibly less bias than a narrative) Generate hypotheses Provide information for future trials Disadvantages of metaanalysis Even randomized studies often differ significantly in their design, outcome, exposure measures Publication bias Studies differ in quality Time trends Health studies tend to be (comparatively) few Interpreting the results of a meta-analysis Was process valid (question, search strategy, reproducible)? Are studies comparable? Are results similar? What is the estimate and precision of the estimate? Conclusion Systematic reviews : top of hierarchy of evidence Caution before accepting findings of any systematic review without first appraising it Cautious Attention paid to patient selection group , intervention, or search strategy; SR combined studies in meta-analysis pooled in different intervention or participants included 3 reasons validity finding 1) Chance 2) Bias 3) Confounding Chance Random variation Chance: statistical analysis (hypothesis testing and estimation.) Avoid random variation : adequate sample size Bias Systematic (non-random) error in estimation of population characteristic e.g. effect of treatment compared to control in a population Systematic means … Classification of sources of bias in analytical studies Allocation Performance Placebo-effect Attrition Detection Analytical Reporting Selection Measurement Analysis 1. Allocation bias Any treatment allocation method that causes a systematic difference in participant characteristics at the start of a trial (baseline) – independent prognostic characteristics (confounders) – failure to plan e.g. confounding by indication – failure to execute 2. Performance bias Systematic differences in the care of the two groups, other than the intervention being investigated – nursing & supportive care – monitoring for adverse effects 3. Placebo-effect bias Placebo-effect - a beneficial effect gained because the participant believes he is receiving effective therapy (includes satisfying pat-doc relationship as well as medicinal intervention) In trials with a “no-treatment” arm, confounding due to a differential placeboeffect may occur if the subjects are aware they are not receiving active therapy Reasons for bias Confounding When a non-causal association due to a common cause of both T and H prevents us from quantifying any causal association Confounding – measured & unmeasured common causes Random variation (chance) imprecise Systematic variation (bias) inaccurate Confounder : factor prognostically linked to outcome and unevenly distributed btw study groups Known confounders : stratify resultsUnknown confounders: randomisation Confounding – measured & unmeasured common causes Non-causal assoc drug cancer Smoking Supportive care Placebo-effect 4. Attrition bias All clinical trials have a period of follow-up, attrition occurs when subjects do not complete the follow-up process (loss to follow-up) This is harmful because attrition causes loss of information and hence less precise estimates of the treatment effect, if too many subjects cannot be analyzed Systematic differences in the loss of participants to follow up between groups may cause bias if the analysis is improper e.g. analyzing only participants who had complete follow-up or who were fully compliant (per protocol analysis) 5. Detection bias Systematic differences in outcome assessment btw groups –measurement method –follow-up frequency for outcomes MY DOCUMENTS.lnk 6. Analytical bias Bias arising because of the method of analysis –choice of subjects to analyze the analysis dataset –choice of statistical estimators biased & unbiased estimators –choice of multivariate models 7. Reporting bias Selective reporting of –clinical outcomes e.g. surrogate, subgroups –time-points e.g. early Use of composite endpoints –component events not equally significant What is Apprasial? A technique to increase effectiveness of reading by exclude research studies too poorly designed to inform practice. Why appraisal? To free time of concentrate on a more systematic evaluation of studies cross quality threshold and extract salient points How to Appraise? Appraising a Secondary studies(Review) 1. Validity 2. Impact(Results) 3. Practicability(Application) 4. Instruments tools such as CASP Critical Appraisal Skills Programme (CASP) http://www.phru.nhs.uk/pages/PH D/CASP.htm Appraisal tools for Systematic review 10 questions to help you make sense of reviews Is the study valid? What are the results? Will the results help locally? 10 questions adapted from Oxman AD, Cook DJ, Guyatt GH, Users’ guides to medical literature. VI. How to use an overview. JAMA 1994; 272 (17): 1367-1371 Screening question First 2 questions Screening questions can be answered quickly. Worth proceeding If answer to both is “yes”, Screening question 1. Did the review ask a clearly-focused question? Yes Can’t tell No Focused : – the population studied – the intervention given or exposure – the outcomes considered 2. Did the review include the right type of study? Yes Can’t tell No included studies: – address the review’s question – have an appropriate study design Is it worth continuing? 3. Did the reviewers try to identify all relevant studies? Yes Can’t tell No Consider: – which bibliographic databases were used – if there was follow-up from reference lists – if there was personal contact with experts –searched for unpublished studies –searched for non-English-language studies 4. Did the reviewers assess the quality of the Yes Can’t tell No i– if a clear, pre-determined strategy was used to determine which studies were included. Look for: – a scoring system – more than one assessor 5. If the results of the studies have been combined, was it reasonable to do so? Consider – the results of each study are clearly displayed – the results were similar from study to study (look for tests of heterogeneity) – the reasons for any variations in results are discussed 6. How are the results presented and what is the main result? Consider: – how the results are expressed (e.g. odds ratio,relative risk, etc.) – how large this size of result is and how 7. How precise are these results? Consider: – if a confidence interval were reported. Would your decision about whether or not to use this intervention be the same at the upper confidence limit as at the lower confidence limit? – if a p-value is reported where confidence 8. Can the results be applied to the local Yes Can’t tell No population? Consider whether: – the population sample covered by the review could be different from your population in ways that would produce different results – your local setting differs much from that of the review – you can provide the same intervention in your setting 9. Were all important outcomes considered? Yes 10. Should policy or practice change as a result of Yes Can’t tell No the evidence contained in this review? Consider: – whether any benefit reported outweighs any harm and/or cost. If this information is not
© Copyright 2024 Paperzz