Simulation and Uncertainty

Download Report

Transcript Simulation and Uncertainty

Simulation and Uncertainty
Tony O’Hagan
University of Sheffield
6 July 2007
I-Sim Workshop, Fontainebleau
1
Outline
Uncertainty
Example – bovine tuberculosis
Uncertainty analysis
Elicitation
Case study 1 – inhibiting platelet aggregation
Propagating uncertainty
Case study 2 – cost-effectiveness
Conclusions
6 July 2007
I-Sim Workshop, Fontainebleau
2
Two kinds of uncertainty
Aleatory (randomness)
Number of heads in 10 tosses of a fair coin
Mean of a sample of 25 from a N(0,1) distribution
Epistemic (lack of knowledge)
Atomic weight of Ruthenium
Number of deaths at Agincourt
Often, both arise together
Number of patients who respond to a drug in a trial
Mean height of a sample of 25 men in Fontainebleau
6 July 2007
I-Sim Workshop, Fontainebleau
3
Two kinds of probability
Frequency probability
Long run frequency in many repetitions
Appropriate only for purely aleatory uncertainty
Subjective (or personal) probability
Degree of belief
Appropriate for both aleatory and epistemic (and
mixed) uncertainties
Consider, for instance
Probability that next president of USA is Republican
6 July 2007
I-Sim Workshop, Fontainebleau
4
Uncertainty and statistics
Data are random
Repeatable
Parameters are uncertain but not random
Unique
Uncertainty in data is mixed
But aleatory if we condition on (fix) the parameters
E.g. likelihood function
Uncertainty in parameters is epistemic
If we condition on the data, nothing aleatory remains
6 July 2007
I-Sim Workshop, Fontainebleau
5
Two kinds of statistics
Frequentist
Based on frequency probability
Confidence intervals, significance tests etc
Inferences valid only in long run repetition
Does not make probability statements about
parameters
Bayesian
Based on personal probability
Inferences conditional on the actual data obtained
Makes probability statements about parameters
6 July 2007
I-Sim Workshop, Fontainebleau
6
Example: bovine tuberculosis
Consider a model for
the spread of
tuberculosis (TB) in
cows
In the UK, TB is
primarily spread by
badgers
Model in order to assess
reduction of TB in cows
if we introduce local
culling (i.e. killing) of
badgers
6 July 2007
I-Sim Workshop, Fontainebleau
7
How the model might look
Simulation model components
Location of badger setts, litter size and fecundity
Spread of badgers
Rates of transmission of disease
Success rate of culling
6 July 2007
I-Sim Workshop, Fontainebleau
8
Uncertainty in the TB model
Simulation
Replicate runs give different outcomes (aleatory)
Parameter uncertainty
E.g. mean (and distribution of) litter size, dispersal
range, transmission rates (epistemic)
Structural uncertainty
Alternative modelling assumptions (epistemic)
Interest in properties of simulation distribution
E.g. probability of reducing bovine TB incidence
below threshold (with optimal culling)
All are functions of parameters and model structure
6 July 2007
I-Sim Workshop, Fontainebleau
9
General structure
Uncertain model parameters (structure) X
With known distribution
True value XT
Object of interest YT = Y(XT)
Possibly optimised over control parameters
Model output Z(X), related to Y(X)
E.g. Z(X) = Y(X) + error
Can run model for any X
Uncertainty about YT due to two sources
We don’t know XT (epistemic)
Even if we knew XT,can only observe Z(XT) (aleatory)
6 July 2007
I-Sim Workshop, Fontainebleau
10
Uncertainty analysis
Find the distribution of YT
Challenges:
Specifying distribution of X
Computing Z(X)
Identifying distribution of Z(X) given Y(X)
Propagating uncertainty in X
6 July 2007
I-Sim Workshop, Fontainebleau
11
Parameter distributions
Necessarily personal
Even if we have data
E.g. sample of badger litter sizes
Expert judgement generally plays a part
May be formal or informal
Formal elicitation of expert knowledge
A seriously non-trivial business
Substantial body of literature, particularly in
psychology
6 July 2007
I-Sim Workshop, Fontainebleau
12
Case study 1
A pharmaceutical company is developing a new
drug to reduce platelet aggregation for patients
with acute coronary syndrome (ACS)
Primary comparator is clopidogrel
Case study concerns elicitation of expert
knowledge prior to reporting of Phase 2a trial
Required in order to do Bayesian clinical trial
simulation
5 elicitation sessions with several experts over a total
of about 3 days
Analysis revisited after Phase 2a and 2b trials
6 July 2007
I-Sim Workshop, Fontainebleau
13
Simulating SVEs
Patient enters
Randomise to
new/clopidogrel
Generate mean
IPA for each drug
Generate patient
IPA
Generate IPASVE relationship
Generate whether
patient has SVE
Patient loop
6 July 2007
I-Sim Workshop, Fontainebleau
SVE = Secondary
vascular event
IPA = Inhibition of
platelet aggregation
14
Distributions elicited
Many distributions were actually elicited
1.
2.
3.
4.
5.
Mean IPA (efficacy on biomarker) for each drug and dose
Patient-level variation in IPA around mean
Relative risk of SVE conditional on individual patient IPA
Baseline SVE risk
Other things to do with side effects
We will just look here at elicitation of the distribution of
mean IPA for a high dose of the new drug
Judgements made at the time
Knowledge now is of course quite different!
But decisions had to be made then about Phase 2b trial
Whether to go ahead or drop the drug
Size of sample, how many doses, etc
6 July 2007
I-Sim Workshop, Fontainebleau
15
Elicitation record
Elicitation title
Session
Date
Start time
Attendance and roles
ORIENTATION
Purpose of elicitation
This record
This document will form an auditable record of the elicitation process.
Participants’ expertise
Nature of uncertainty
Effective elicitation
Strengths/weaknesses
PRACTICE
Objective
Practice elicitation. Eliciting knowledge about the population of Portugal. Etc.
IPA DISTRIBUTIONS
Objective
To elicit a joint probability distribution for individual patient IPA Etc.
Definitions
IPA is platelet aggregation inhibition, on a scale of 0 to 100 (i.e. 0% to 100%) Etc.
Evidence
Healthy volunteer data, about 150 volunteers in total. Etc.
Structuring
Session ended
6 July 2007
I-Sim Workshop, Fontainebleau
16
Eliciting one distribution
Mean IPA (%) for high dose
Range: 80 to 100
Median: 92
Probabilities: P(over 95) = 0.4, P(under 85) = 0.2
Chosen distribution:
Beta(11.5, 1.2)
Median 93
P(over 95) = 0.36,
P(under 85) = 0.20,
P(under 80) = 0.11
6 July 2007
I-Sim Workshop, Fontainebleau
17
Propagating uncertainty
Usual approach is by Monte Carlo
Randomly draw parameter sets Xi, i = 1, 2, …, N from
distribution of X
Run model for each parameter set to get outputs Yi =
Y(Xi), i = 1, 2, …, N
Assume for now that we can do big enough runs to ignore
the difference between Z(X) and Y(X)
These are a sample from distribution of YT
Use sample to make inferences about this distribution
Generally frequentist but fundamentally epistemic
Impractical if computing each Yi is
computationally intensive
6 July 2007
I-Sim Workshop, Fontainebleau
18
Optimal balance of resources
Consider the situation where each Z(Xi) is an
average over n individuals
And Y(Xi) could be got by using very large n
Then total computing effort is Nn individuals
Simulation within simulation
Suppose
The variance between individuals is v
The variance of Y(X) is w
We are interested in E(Y(X)) and w
Then optimally n = 1 + v/w (approx)
Of order 36 times more efficient than large n
6 July 2007
I-Sim Workshop, Fontainebleau
19
Emulation
When even this efficiency gain is not enough
Or when we the conditions don’t hold
We may be able to propagate uncertainty
through emulation
An emulator is a statistical model/approximation
for the function Y(X)
Trained on a set of model runs Yi = Y(Xi) or Zi = Z(Xi)
But Xis not chosen randomly (inference is now Bayesian)
Runs much faster than the original simulator
Think neural net or response surface, but better!
6 July 2007
I-Sim Workshop, Fontainebleau
20
Gaussian process
The emulator represents Y(.) as a Gaussian
process
Prior distribution embodies only a belief that Y(X) is a
smooth, continuous function of X
Condition on training set to get posterior GP
Posterior mean function is a fast approximation to Y(.)
Posterior variance expresses additional uncertainty
Unlike neural net or response surface, the GP
emulator correctly encodes the training data
6 July 2007
I-Sim Workshop, Fontainebleau
21
2 code runs
Consider one input and one output
Emulator estimate interpolates data
Emulator uncertainty grows between data points
dat2
10
5
0
0
1
2
3
4
5
6
x
6 July 2007
I-Sim Workshop, Fontainebleau
22
3 code runs
Adding another point changes estimate and
reduces uncertainty
dat3
10
5
0
0
1
2
3
4
5
6
x
6 July 2007
I-Sim Workshop, Fontainebleau
23
5 code runs
And so on
9
8
7
dat5
6
5
4
3
2
1
0
0
1
2
3
4
5
6
x
6 July 2007
I-Sim Workshop, Fontainebleau
24
Then what?
Given enough training data points we can
emulate any model accurately
So that posterior variance is small “everywhere”
Typically, this can be done with orders of magnitude
fewer model runs than traditional methods
Use the emulator to make inference about other
things of interest
E.g. uncertainty analysis, calibration, optimisation
Conceptually very straightforward in the
Bayesian framework
But of course can be computationally hard
6 July 2007
I-Sim Workshop, Fontainebleau
25
Case study 2
Clinical trial simulation coupled to economic
model
Simulation within simulation
Outer simulation of clinical trials, producing trial
outcome results
In the form of posterior distributions for drug efficacy
Incorporating parameter uncertainty
Inner simulation of cost-effectiveness (NICE decision)
For each trial outcome simulate patient outcomes with those
efficacy distributions (and many other uncertain parameters)
Like the “optimal balance of resources” slide
But complex clinical trial simulation replaces simply drawing
from distribution of X
6 July 2007
I-Sim Workshop, Fontainebleau
26
Emulator solution
5 emulators built
Means and variances of (population mean)
incremental costs and QALYs, and their covariance
Together these characterised the Cost
Effectiveness Acceptability Curve
Which was basically our Y(X)
For any given trial design and drug development
protocols, we could assess the uncertainty (due to all
causes) regarding whether the final Phase 3 trial
would produce good enough results for the drug to be
1.
2.
6 July 2007
Licensed for use
Adopted as cost-effective by the UK National Health Service
I-Sim Workshop, Fontainebleau
27
Conclusions
The distinction between epistemic and aleatory
uncertainty is useful
Recognising that uncertainty about parameters
of a model (and structural assumptions) is
epistemic is useful
Expert judgement is an integral part of specifying
distributions
Uncertainty analysis of a stochastic simulation
model is conceptually a nested simulation
Optimal balance of sample sizes
More efficient computation using emulators
6 July 2007
I-Sim Workshop, Fontainebleau
28
References
On elicitation
O’Hagan, A. et al (2006). Uncertain Judgements: Eliciting Expert
Probabilities. Wiley
www.shef.ac.uk/beep
On optimal resource allocation
O’Hagan, A., Stevenson, M.D. and Madan, J. (2007). Monte
Carlo probabilistic sensitivity analysis for patient level simulation
models: Efficient estimation of mean and variance using ANOVA.
Health Economics (in press)
Download from tonyohagan.co.uk/academic
On emulators
O'Hagan, A. (2006). Bayesian analysis of computer code
outputs: a tutorial. Reliability Engineering and System Safety 91,
1290-1300.
mucm.group.shef.ac.uk
6 July 2007
I-Sim Workshop, Fontainebleau
29