2015 Joint Congress on Medical Imaging and Radiation Sciences

2015 Joint Congress on
Medical Imaging and Radiation Sciences
Montreal, QC
May 30, 2015
George Wells
Comparative & Cost Effectiveness
Related to Diagnostic Testing
Learning Objectives
• To be aware of the role of comparative effectiveness in comparing
diagnostic tests
• To understand the steps in conducting a cost-effectiveness studies
• To appreciate the cautions associated with designing, conducting
and interpreting comparative and cost-effectiveness studies
Comparative
effectiveness
Cost
effectiveness
Test
performance
INTRODUCTION
Motivation
Medical imaging has transformed medical care
Evidence gap - growth in using many types of medical imaging has
occurred without convincing evidence of additional clinical benefits or
improved cost-effectiveness over existing care
Assessment need - with increasing cost pressures and wide variation in
usage, these gaps in evidence create a compelling argument for more
assessment of medical imaging to support health care decisions
Challenges – levels of evidence
A framework to evaluate diagnostic technologies was developed in 1970s
by categorizing evidence into six levels
Level of Evidence
Examples
Technical
Pixels per millimeter, Section thickness
Diagnostic accuracy
Sensitivity, Specificity, Area under ROC curve
Impact of diagnostic thinking
Percentage of cases in which the clinician believes that the test
alters the diagnosis
Impact of therapeutic actions
Impact on patient outcomes
Impact on societal outcomes
Test performance
Build upon test performance with each level
Percentage of cases in which the choice of therapies is changed
representing
a the
distinct
type of evidence that
after information from
test is provided
might,
or might not, be available when
Differences in mortality, morbidity, or quality of life between patients
evaluating
a test
diagnostic
test without it
managed with the
and those managed
Cost-effectiveness of the improvement in patient outcomes, such
as cost per life-year saved, calculated from a societal perspective
Pearson SD, Knudsen AB, Scherer RW, Weissberg J, Gazelle GS,
Assessing The Comparative Effectiveness Of A Diagnostic Technology:
CT Colonography. Health Affairs 27, no. 6 (2008): 1503–1514
Challenges – some features that complicate
creation or assessment of evidence
Level of Evidence
Examples
Constant change
• Imaging devices and patient preparation techniques evolve, and
different versions can be in use or under study at any one time
• Assessments hindered by questions of whether evidence has
much relevance to the most recent version of the technique
Role of interpretation
• Diagnostic accuracy of an imaging technique is linked with
practitioners’ experience and skill in interpreting results
Variability in published
results
• Sensitivity and specificity may vary due to differences in the
spectrum of patients – higher in patients with more typical or
severe; lower among patients with atypical or less severe
Transferability
• “Best hands” problem - how transferable are results coming
from studies of the most advanced devices, in the hands of
experts, among highly selected patients at academic health
centers TO devices, practitioners, and patients in the community
Screening intervals
• Evidence comparing the onetime use of a new technique with an
established gold standard produces no evidence to guide
decisions on intervals for screening or follow-up – clinical and
cost effectiveness may be highly dependent on frequency
Impact on patient
outcomes
• Evidence on a medical imaging technique often does not include
direct evidence of its impact on patient outcomes
Framework for levels of evidence
Level of Evidence
Examples
Technical
Pixels per millimeter, Section thickness
Diagnostic accuracy
Sensitivity, Specificity, Area under ROC curve
Impact of diagnostic thinking
Percentage of cases in which the clinician believes that the test
alters the diagnosis
Impact of therapeutic actions
Percentage of cases in which the choice of therapies is changed
after information from the test is provided
Impact on patient outcomes
Differences in mortality, morbidity, or quality of life between
patients managed with the test and those managed without it
Impact on societal outcomes
Cost-effectiveness of the improvement in patient outcomes, such
as cost per life-year saved, calculated from a societal perspective
META-ANALYSIS
Systematic reviews and meta-analysis
Systematic Review: application of scientific strategies that limit bias
to the systematic assembly, critical appraisal and synthesis of all relevant
studies on a specific topic
Meta-Analysis: a systematic review that employs statistical methods
to combine and summarize the results of several studies
Steps in a systematic review
Clearly formulated
question
Comprehensive data
search
Unbiased selection and
extraction process
Critical appraisal of data
Analysis of data
Sensitivity and subgroup
analyses
Interpretation of results
Steps in a systematic review
Clearly formulated
question
Pose question specifying operational definitions for: Population, Index tests,
Comparator tests, Outcomes
Criteria for considering inclusion of studies: PICO
P: Type of Participants / Population
I: Types of Index diagnostic tests
C: Types of comparator diagnostic tests
O: Types of outcome measures
Comprehensive data
search
Unbiased selection and
extraction process
Critical appraisal of data
Analysis of data
Sensitivity and subgroup
analyses
Interpretation of results
Steps in a systematic review
Clearly formulated question
Comprehensive data
search
Unbiased selection and
extraction process
Critical appraisal of data
Analysis of data
Sensitivity and subgroup
analyses
Interpretation of results
•
•
•
•
•
Need a well formulated and coordinated effort
Requirements for comprehensiveness of search depends on question
Usually begin with searches of bibliographic databases
Publications retrieved and references therein searched
As a step to eliminate publication bias need unpublished research
Steps in a systematic review
Clearly formulated question
Comprehensive data search
Unbiased selection and
extraction process
Study Selection
• 2 independent reviewers select studies
• Selection of studies based on PICO
• Level of agreement: kappa; differences resolved by consensus
• Specify reasons for excluding studies
Data Extraction
• 2 independent reviewers extract data using predetermined forms
• PICO characteristics, study design/methods, study results, quality items
• Level of agreement: kappa; differences resolved by consensus
Critical appraisal of data
Analysis of data
Sensitivity and subgroup
analyses
Interpretation of results
Steps in a systematic review
Clearly formulated question
Comprehensive data search
Unbiased selection and extraction process
Critical appraisal of data
Description of Studies
• Table of characteristics of included studies
• Reference, design, PICO elements, quality assessment
Quality Assessment
• Report quality assessment instrument used
(e.g. QUADAS – Quality Assessment of Diagnostic Accuracy Studies)
• Quality assessment instrument - transparent, parsimonious, reproducibility
Analysis of data
Sensitivity and subgroup
analyses
Interpretation of results
Steps in a systematic review
Clearly formulated question
Comprehensive data search
Unbiased selection and extraction process
Critical appraisal of data
Analysis of data
Effect estimates (EE): sensitivity, specificity, likelihood ratio (LR), diagnostic odds ratio (DOR)
Heterogeneity: clinical (PICO), methodological (design/conduct)
Procedures: do not pool, meta-regression, sensitivity, subgroup
Fixed effects model (FE); Random effects model (RE)
Procedures: subgroup
Overall estimates and analyses
Display: Forest Plots, Summary of Findings (SoF) tables
Summary receiver operating characteristic (SROC) curves with area under the curve
(AUC) as a measure of diagnostic accuracy - shape of curve on basis of changes in
DOR using Moses’s constant-of-linear model
Steps in a systematic review
Clearly formulated question
Comprehensive data search
Unbiased selection and extraction process
Critical appraisal of data
Analysis of data
Forest Plot
Sample ROC curve for diagnostic accuracy
Steps in a systematic review
Clearly formulated question
Comprehensive data search
Unbiased selection and extraction process
Critical appraisal of data
Analysis of data
Sensitivity and subgroup
analyses
Subgroup Analysis:
• Pre-specify hypothesis-testing subgroup analyses; label others as a posteriori
Sensitivity Analysis:
• Test robustness of results relative to key features of the studies and key
assumptions and decision
Interpretation of results
Steps in a systematic review
Clearly formulated question
Comprehensive data search
Unbiased selection and extraction process
Critical appraisal of data
Analysis of data
Sensitivity and subgroup analyses
Interpretation of results
•
•
•
•
•
•
Interpret results in context of current health care
State methodological limitations of studies and review
Consider size of effect and review consistency
Interpret results in light of other available evidence
Make recommendations clear and practical
Propose future research agenda
Illustration - Rb-82 PET
Background:
• Myocardial perfusion imaging is widely used in assessing patients with known or
suspected CAD
• PET using Rb-82 has potential advantages over SPECT that may make it more
accurate and reduce radiation exposure, has increased costs
• Comparisons of technologies are relevant for policy makers and practice guidelines
Illustration - Rb-82 PET
Objective:
To evaluate the accuracy of rubidium (Rb)-82 PET for the diagnosis of
obstructive coronary artery disease (CAD) in comparison to SPECT
Method:
• Systematic review of studies
• Where either Rb-82 PET or technetium-99m SPECT with both
attenuation correction and electrocardiography-gating were used as
a diagnostic test for obstructive CAD with invasive coronary
angiogram as a reference standard
Illustration - Rb-82 PET
Flow Diagram Showing the Results of the Systematic Review
Illustration - Rb-82 PET
Forest Plots Showing Sensitivities and Specificities of: (A) Rb-82
PET Studies and (B) Tc-99m SPECT Studies
Illustration - Rb-82 PET
SROC Curves for Diagnostic Accuracy of Rb-82 PET and
Tc-99m SPECT
With ECG-Gating and Attenuation Correction
Area under the curve (AUC) was compared for Rb82 PET (A) and Tc-99m SPECT (B), with Rb-82 PET
showing superior accuracy (p 0.001).
Illustration - Rb-82 PET
Conclusion:
• Rb-82 PET is accurate for the detection of obstructive CAD and,
despite advances in SPECT technology, remains superior
• More widespread use of Rb-82 PET may be beneficial to improve
CAD detection
Framework for levels of evidence
Level of Evidence
Examples
Technical
Pixels per millimeter, Section thickness
Diagnostic accuracy
Sensitivity, Specificity, Area under ROC curve
Impact of diagnostic thinking
Percentage of cases in which the clinician believes that the test
alters the diagnosis
Impact of therapeutic actions
Percentage of cases in which the choice of therapies is changed
after information from the test is provided
Impact on patient outcomes
Differences in mortality, morbidity, or quality of life between
patients managed with the test and those managed without it
Impact on societal outcomes
Cost-effectiveness of the improvement in patient outcomes, such
as cost per life-year saved, calculated from a societal perspective
COMPARATIVE
EFFECTIVENESS
Endpoints for diagnostic test evaluation
• Diagnostic performance:
• Measures of accuracy
Accuracy • Measures of predictive value
Affects
care
Affects
outcomes
• Intermediate process of care:
• Diagnostic thinking/decision making
• Therapeutic thinking/decision making
• Patient outcomes:
• Mortality, morbidity, quality of life, costs
Fundamental Challenge
Cost effectiveness research calls
for informing the path from
medical imaging diagnostic
information to patients outcomes
•
Diagnostic imaging tests provide
information for use in selecting course of
care
•
Both long and short-term effects of tests
materialize in context of available health
care options, including therapeutic
interventions
Diagnostic
information
Comparative
effectiveness
Diagnostic
workup
Outcomes
Treatment
decisions
Patient
outcomes
Hierarchy of Studies – levels of evidence
RCTs
Cohort studies
Case-Control studies
Cross-sectional studies
Case Series/ Case Reports
•
Concept of moving from low to high quality evidence generating
studies
RANDOMIZED CONTROLLED
TRIAL
Structure: Randomized Controlled Trial
Target Population
Study Population
Participants
Non-participants
R
Imaging Test A
Imaging Test B
Outcome
Outcome
Structure: Randomized Controlled Trial
Target Population
Study Population
Participants
Non-participants
R
Imaging Test A
Imaging Test B
Outcome
Outcome
Structure: Randomized Controlled Trial
Target Population
Study Population
Participants
Non-participants
R
Imaging Test A
Imaging Test B
Outcome
Outcome
Structure: Randomized Controlled Trial
Target Population
Study Population
Participants
Non-participants
R
Imaging Test A
Imaging Test B
Outcome
Outcome
Structure: Randomized Controlled Trial
Target Population
Study Population
Participants
Non-participants
R
Imaging Test A
Imaging Test B
Outcome
Outcome
Structure: Randomized Controlled Trial
Target Population
Study Population
Participants
Non-participants
R
Imaging Test A
Imaging Test B
Outcome
Outcome
Structure: Randomized Controlled Trial
Target Population
Study Population
Participants
Non-participants
R
Imaging Test A
Imaging Test B
Performance bias
Attrition bias
Outcome
Outcome
Detection bias
Structure: Randomized Controlled Trial
Target Population
Study Population
Participants
Non-participants
R
Imaging Test A
Imaging Test B
Compliance
Contamination
Outcome
Outcome
Cointervention
Structure: Randomized Controlled Trial
Target Population
Study Population
Participants
Non-participants
R
Imaging Test A
Imaging Test B
Outcome
Outcome
Analysis Populations
Potential Weaknesses
Limited applicability due to ethic considerations and artificial
experimental setting
Differential rates of noncompliance, withdrawals and losses to follow-up
Costly and logistically time consuming
Sample size due to recruitment (volunteers, eligibility)
External validity
NON-RANDOMIZED STUDIES
Some bias domains through which bias might be
introduced into a non-randomized study
Bias Domain
Related Terms
Bias due to confounding
Selection bias; Allocation bias; Casemix bias; Channelling
bias
Bias in selection of
participants into the study
Selection bias; Inception bias; Lead-time bias; Immortal
time bias
Bias in measurement of
interventions
Misclassification bias; Information bias; Recall bias;
Measurement bias; Observer bias
Bias due to departures
from intended interventions
Performance bias; Time-varying confounding
Bias due to missing data
Attrition bias; Selection bias
Bias in measurement of
outcomes
Detection bias; Recall bias; Information bias;
Misclassification bias; Observer bias; Measurement bias
Bias in selection of the
reported result
Outcome reporting bias; Analysis reporting bias
Making Comparisons
Exposure
Yes vs No
Outcome
Background
factors
1. Design
2. Analysis
3. Randomize
-
match
stratified sampling
restrict inclusion
standardize
stratify (subgroups)
adjust (statistical model)
Adjusting for Confounding
In conducting comparative effectiveness analysis for nonrandomized studies, concern regarding the imbalance of
underlying confounding variables between comparison groups
exists
Methods for addressing this concern:
• Covariate adjustment
• Propensity score methods
• Instrumental variable analysis
Regression Models – Covariate adjustment
General Linear Model
Y=β0 + β1X1 + β2X2 + … + βkXk + ε
Independent variables
Dependent variable
Regression Models – Covariate adjustment
Method
Dependent
Independent
General purpose
Statistical tools
Multiple
regression
analysis
Continuous
Continuous
(in theory all)
To describe the extent, direction and
strength of the relationship between
several independent variables and a
continuous dependent variable
• F-tests
• t-tests
Logistic
regression
analysis
Binary
(typically)
Continuous
Discrete
To describe the extent, direction and
strength of the relationship between
several independent variables and a
binary dependent variable
• chi-square test
• odds ratio
Cox
proportional
hazards
model
Continuous
(time-toevent)
Continuous
(in theory all)
To describe the time to occurrence of the
event of interest and the relationship
between tine to event and several
independent variables
•
•
•
Poisson
regression
analysis
Counts
(rates)
Continuous
Discrete
To describe the extent, direction and
strength of the relationship between
several independent variables and a rate
• t-tests
• chi-square test
log rank test
chi-square texts
hazard ratio
Propensity Score Analysis (PS)
A PS is a probability of receiving a specified imaging technology conditioned
on underlying known and observed confounding variables
PS analysis mimics some of the conditions of an RCT
• For example, for data matched on the PS, the distribution of measured baseline
characteristics will be similar in the imaging technologies groups compared
A PS analysis can yield unbiased estimates of the effect (impact)
PS can be calculated using logistic regression models where the imaging
technology received is the outcome and the variables on which we wish to
balance the data are the explanatory variables
Propensity Score Analysis
Once the PS is generated, the adjusted analysis can be implemented using
several different approaches
Covariate
Match
Stratify
• Consider the PS as a covariate in the regression model
• Match patients on the PS
• Use conditional logistic regression for analyzing the
matched data
• Divide the PS into strata, and separate the patients by
these strata so patients are similar on the measured
variables within each stratum
• Compare patients within strata and an overall treatment
effect can be obtained by calculating a weighted mean
across the strata
Propensity Score Analysis - pros
and cons
Pros
Cons
More similar to randomized comparison
More steps then just adding covariates to model
individually
Able to assess overlap of imaging groups
Effects of individual covariates on outcome are lost
Possible to develop score while blinded to
outcomes
Can only adjust for measured variables, and one can
only be ensured that the data is balanced on these
measured variables
Only one score needed for all outcomes
If unmeasured variables exist that are strongly
associated with the outcome, then will not be able to
adjust for them and biased estimates of the effect
(impact) may result
Avoids multicollinearity
May give false sense of security if an important
confounder is not collected
Instrumental Variable Analysis (IV)
Propensity scores can be useful, but they rely on observable and measurable
variables
• If important variables are unobservable then the effect estimates obtained from
propensity score analysis can be biased
IV method attempts to account for such unmeasured and unobserved sources of
confounding
IV method identifies variables correlated with the
imaging technology assigned but not directly
correlated to the outcome
• When such variables are found they can be
incorporated into the analysis in an attempt
to remove allocation bias
Imaging
Instrumental Variable Analysis
Once an IV is identified:
• Test if the IV is predictive of the imaging technology (can use logistic
regression)
• Next assess relationship between the IV and the outcome (can use
correlational analysis)
Once an IV is evaluated, a two-step process can be followed for the analysis
First
Second
• The IV along with other
covariates of interest are
used to predict the
imaging technology
• The predicted imaging
technology from first
stage and other
covariates are used to
estimate the outcome
Instrumental Variables –
Examples (Wall Street Journal)
Illustration – PARR-2
Background:
• Patients with severe ventricular dysfunction and suspected coronary disease, may benefit
from revascularization, but have significant perioperative morbidity and mortality
• FDG PET can detect viable myocardium that might recover after revascularization
Illustration – PARR-2
Objective:
To assess the effectiveness of FDG PET-assisted management in patients with severe
ventricular dysfunction and suspected coronary disease compared to standard care
Methods:
• Included patients with severe left ventricular (LV) dysfunction and suspected coronary
disease being considered for revascularization, heart failure, or transplantation workups or in whom PET was considered potentially useful
• Patients were stratified according to recent angiography or not, then randomized to
management assisted by FDG PET (n= 218) or standard care (n= 212)
• Primary outcome was the composite of cardiac death, myocardial infarction, or
recurrent hospital stay for cardiac cause, within 1 year.
Illustration – PARR-2
• Time to first occurring outcome of
composite event
• HR 0.78 (95% CI 0.58 to 1.1), p 0.15
• Time to first cardiac death for patients
without recent angiography
• HR 0.4 (95% CI 0.17 to 0.96), p 0.035
• Time to cardiac death for all patients
• HR 0.72 ( 95% CI 0.4 to 1.3), p 0.25
• Adherence to PET subgroup
• Time to first occurring outcome of the
composite event
• Adjusted HR 0.62 (95% CI 0.42 to
0.93), p 0.019.
Illustration – PARR-2
Conclusion:
• This study did not demonstrate a significant reduction in cardiac events in patients with
LV dysfunction and suspected coronary disease for FDG PET-assisted management
versus standard care
• In those who adhered to PET recommendations and in patients without recent
angiography, significant benefits were observed
• The utility of FDG PET is best realized in this subpopulation and when adherence to
recommendations can be achieved
Framework for levels of evidence
Level of Evidence
Examples
Technical
Pixels per millimeter, Section thickness
Diagnostic accuracy
Sensitivity, Specificity, Area under ROC curve
Impact of diagnostic thinking
Percentage of cases in which the clinician believes that the test
alters the diagnosis
Impact of therapeutic actions
Percentage of cases in which the choice of therapies is changed
after information from the test is provided
Impact on patient outcomes
Differences in mortality, morbidity, or quality of life between
patients managed with the test and those managed without it
Impact on societal outcomes
Cost-effectiveness of the improvement in patient outcomes, such
as cost per life-year saved, calculated from a societal perspective
COST EFFECTIVENESS
Concepts
• Economics is the study of unlimited needs/wants
constrained by a limited number of resources (scarcity)
• Choices need to be made
• For each choice that is made there is an opportunity cost
associated with it – something else that is given up
Opportunity Costs in Health Care
Since we are not able to pay for all therapies, we need to
make choices
In doing so, we need to know whether a therapy is worth the
cost
$
Clinical
benefits
Opportunity Costs in Health Care
To make choices we compare:
Costs
• Cost of the therapy
• Cost of health care resources (more or less because of the
therapy)
• Cost of ‘other things’ - lost time attending appointments
Benefits
• Makes me feel better
• Makes therapy more convenient
• Makes some ‘other things’ better
Example – Comparing Therapies
• When comparing two therapies, a therapy is a
good option (economically attractive) if:
• It does what the other does but costs less
• It costs the same as the other but does better
• It costs less and does better than the other therapy
• What if it:
• Costs less and does less? OR
• Costs more and does more?
Decrease in
total costs(-)
Consider
cost effectiveness
Status quo better than
new therapy

0
Increase in
total costs (+)
Comparing Therapies
?
?
O

Consider
cost effectiveness
Reduced clinical benefit (-)
New therapy better than
status quo
0
Improved clinical benefits (+)
Economic Evaluations
• Provide a measure of “value for money”
• Comprised of two concepts:
1. Cost
2. Clinical / health effects
• Systematic way to compare therapies
Types of Economic Evaluations
• Typically reported as a ratio (cost effectiveness):
Incremental
Cost Effectiveness of
=
Therapy (1) vs Therapy (2)
Total Cost (1) – Total Cost (2)
Effect (1) – Effect (2)
ICER – Incremental Cost Effectiveness Ratio
Types of Economic Evaluations
• Typically reported as a ratio (cost effectiveness):
Incremental
Cost Effectiveness of
=
Therapy (1) vs Therapy (2)
Total Cost (1) – Total Cost (2)
Effect (1) – Effect (2)
Type of analysis largely
depends on clinical effects
Cost Minimization
Focus on total cost associated with
therapies (therapies, health care cost,
administration, monitoring)
Incremental
Cost Effectiveness of
=
Therapy (1) vs Therapy (2)
Total Cost (1) – Total Cost (2)
Effect (1) = Effect (2)
Similar safety and effectiveness
established (clinical efficacy, safety,
compliance, patient satisfaction)
Costs
Opportunity cost is based on the value that would be gained from using
resources elsewhere
– In practice, assume price paid reflects the opportunity cost
Distinguish costs associated with a therapy between:
– Direct costs: medical (drugs, procedures, staff time, equipment);
patient (transportation, out-of pocket expenses)
– Indirect costs: production losses, other uses of time
– Intangibles: pain; suffering; adverse effects
Essential to specify which costs included in a cost-effectiveness analysis
Cost Effectiveness
Incremental
Cost Effectiveness of
=
Therapy (1) vs Therapy (2)
Total Cost (1) – Total Cost (2)
Effect (1) – Effect (2)
-
In terms of clinically meaningful outcomes
e.g., survival, clinical event avoided
-
Where “outcomes of uncertain clinical
significance” are used, interpretation of
results challenging
Cost Utility
Incremental
Cost Effectiveness of
=
Therapy (1) vs Therapy (2)
-
Total Cost (1) – Total Cost (2)
QALY (1) – QALY (2)
Therapy impacts patients quality of life or
meaningful outcomes that in turn affect
quality of life
What is a utility?
• A utility is a quantitative expression of an individuals’
preference for a particular health state
• Utilities can be measured by direct or indirect techniques
Reference: What is a QALY? London:
Hayward Medical Communications, 2001
What is a Quality Adjusted Life Year
(QALY)?
• Outcome measure that incorporates both quantity of life
(mortality) and health-related quality of life (morbidity)
• Quantity – how long person lives
• Quality – factor that represents a preference for a health
state
• one year of perfect health = one QALY
• one year less than perfect health < one QALY
• death = zero
Types of Analyses
Form of Analysis
Measurement
of Costs
Measurement of
Benefit
Synthesis of Costs and
Benefits
Cost Minimization
Analysis (CMA)
Dollars
None
Incremental cost
Cost Effectiveness
Analysis (CEA)
Dollars
Single dimension of
effectiveness
(e.g. life years gained)
Incremental cost
effectiveness;
Incremental cost per unit
gained
Cost Utility Analysis
(CUA)
Dollars
Utility gained
(e.g. QALYs – quality
adjusted life years)
Incremental cost
effectiveness;
Incremental cost per QALY
gained
Cost Benefit
Analysis (CBA)
Dollars
Monetary value of
benefits gained
Net benefit gained
Advantages of Economic Evaluations
• Provides a consistent framework to compare therapies
• Enables assessment or comparison of different therapies
(i.e., past therapies reviewed by CADTH, NICE etc.)
– League table - comprises a list of healthcare
interventions in ascending order (from low to high) of
their ICER (e.g., from high priority (low cost per QALY)
to low priority (high cost per QALY)).
• Reports costs and benefits in a single metric
• Provides a framework to assess uncertainty (i.e., sensitivity
analyses) to gain a better understanding of where true cost
effectiveness may fall
Rationale of CEA in diagnostic imaging
• CEA (cost-effectiveness analysis) – a comparative analysis of
alternative courses of action in terms of both their costs and
consequences
• When assessing the cost-effectiveness of DI, the initial question
is whether adding an imaging test in a medical pathway does
improve medical decision-making
• In many clinical situations imaging is already part of the
particular disease management; so CEA is used to compare
potential new imaging technologies to each other as well as to
the current reference standard
Steps for CEA in Diagnostic Imaging
Defining the decision
problem
Developing the decision
model
Selecting model input
parameters
Analysis of values and
uncertainty
Interpretation of the
results
Transferability and
validation
Sailer AM, van Zwam WH, Wildberger JE, Grutters JOC,
Cost-effectiveness modelling in diagnostic imaging: a
stepwise approach. European Radiology, Published online
May 24 2015
Steps for CEA in Diagnostic Imaging
Defining the decision
problem
Developing the decision
model
Selecting model input
parameters
Analysis of values and
uncertainty
Interpretation of the
results
Transferability and
validation
• Clear statement of imaging decision problem, modelling objectives and the
scope of the model
• Factors to consider are:
• decision maker (e.g. government or medical doctor)
• perspective (e.g. societal perspective or health care perspective)
• comparators (which imaging strategies should be compared in the analysis)
• outcomes of model (e.g. which health consequences)
• target population (for which patients is the decision relevant)
Steps for CEA in Diagnostic Imaging
Defining the decision
problem
Developing the decision
model
• Common practice in imaging studies to combine a decision tree for the shortterm diagnostic accuracy and treatment decision with a Markov model for the
longer-term disease consequences
• Decision Tree
• Flow model to visualize and calculate the effects of imaging on costs and
health outcome in a static situation and short time frame
• A hypothetical cohort of patients passes through the model and is divided
over the different pathways in the decision tree, according to the assigned
probabilities
Steps for CEA in Diagnostic Imaging
Defining the decision
problem
Developing the decision
model
• The Markov Model
• In non-static situations, such as in chronic diseases, health states of the
patients may alter depending on imaging and treatment options
• In such situations, a Markov model can be used to reflect alterations in
health states during the period of CEA
• In a Markov model, the clinical situation is described in terms of the
conditions that patients can be in (‘health states’), how they can move in
such states (‘transitions’), and how likely such moves are (‘transition
probabilities’)
Steps for CEA in Diagnostic Imaging
Defining the decision
problem
Developing the decision
model
Selecting model input
parameters
Analysis of values and
To predict costs and health outcome as effects of DI, both the direct and indirect
effects of the imaging test itself on health outcome and costs should be
considered as model input parameters
Steps for CEA in Diagnostic Imaging
Defining the decision
problem
Developing the decision
model
Selecting model input
parameters
Analysis of values and
uncertainty
• Cost-effectiveness modelling serves two purposes:
• Estimate expected costs and outcomes for medical decision-making
• Assess uncertainties around these estimates and its range of validity
• Expected costs and outcomes can be obtained by multiplying the probabilities
with their relevant costs and health outcomes
• Parameter uncertainty reflects the uncertainty of the model input parameters
because we do not know the precise values - can be addressed by a:
• Deterministic sensitivity analysis
• Probabilistic sensitivity analysis
Interpretation of the
results
Transferability and
validation
Steps for CEA in Diagnostic Imaging
Defining the decision
problem
Developing the decision
model
Selecting model input
parameters
Analysis of values and
uncertainty
Interpretation of the
results
Cost-effectiveness graph - total estimated costs of diagnostic test, treatments and
health state are compared to estimate health outcome over the chosen period of
time
Steps for CEA in Diagnostic Imaging
Defining the decision
problem
Developing the decision
model
Selecting model input
parameters
Analysis of values and
uncertainty
Interpretation of the
results
Cost-effectiveness acceptability curves (CEACs) - probability of cost-effectiveness
of investigated imaging tests is plotted against the willingness to pay for a quality
adjusted life year
Steps for CEA in Diagnostic Imaging
Defining the decision
problem
Developing the decision
model
Selecting model input
parameters
Analysis of values and
uncertainty
Interpretation of the
results
Transferability and
validation
When developing the model, it is important to consider the decision make and the
jurisdiction - the value of input parameters may well differ between countries, and
the chosen estimates should be relevant for the given society
A decision model that has been developed for one country might need adaptation
to support decision-making in another.
Important that input parameters of a model are transparently described; this
allows others to explore whether the inputs are relevant for their country or
decision problem
A transparent description of all inputs and choices is also important for reasons of
validity
Level of Evidence
Examples
Technical
Pixels per millimeter, Section thickness
Diagnostic accuracy
Sensitivity, Specificity, Area under ROC curve
Impact of diagnostic thinking
Percentage of cases in which the clinician believes that the test
alters the diagnosis
Impact of therapeutic actions
Percentage of cases in which the choice of therapies is changed
after information from the test is provided
Impact on patient outcomes
Differences in mortality, morbidity, or quality of life between patients
managed with the test and those managed without it
Impact on societal outcomes
Cost-effectiveness of the improvement in patient outcomes, such
as cost per life-year saved, calculated from a societal perspective
Questions or Thoughts