Important Note: These materials have been superseded by materials provided as part of the as part of the SISMID short course on Simulation-based Inference.


Licensed under the Creative Commons Attribution-NonCommercial license. Please share and remix noncommercially, mentioning its origin.
CC-BY_NC This document has its origins in the SISMID short course on Simulation-based Inference given by Aaron King and Edward Ionides.

Produced with R version 3.3.3 and pomp version 1.12.1.1.


Introduction

This tutorial covers likelihood estimation via the method of iterated filtering. It presupposes familiarity with building partially observed Markov process (POMP) objects in the R package pomp (King et al. 2016). pomp is available from CRAN and github. This tutorial follows on from the topic of carrying out particle filtering (also known as sequential Monte Carlo) via pfilter in pomp.

We have the following goals:

  1. Review the available options for inference on POMP models, to put iterated filtering in context.
  2. Understand how iterated filtering algorithms carry out repeated particle filtering operations, with randomly perturbed parameter values, in order to maximize the likelihood.
  3. Gain experience carrying out statistical investigations using iterated filtering in a relatively simple situation (fitting an SIR model to a boarding school flu outbreak).

Many, many statistical methods have been proposed for inference on POMP models (He et al. 2010,King et al. (2016)). The volume of research indicates both the importance and the difficulty of the problem. Let’s start by considering three criteria to categorize inference methods: the plug-and-play property; full-information or feature-based; frequentist or Bayesian.

Plug-and-play (also called simulation-based) methods

  • Inference methodology that calls rprocess but not dprocess is said to be plug-and-play. All popular modern Monte Carlo methods fall into this category.
  • Simulation-based is equivalent to plug-and-play.
    • Historically, simulation-based meant simulating forward from initial conditions to the end of the time series.
    • However, particle filtering methods instead consider each observation interval sequentially. They carry out multiple, carefully selected, simulations over each interval.
    • Plug-and-play methods can call dmeasure. A method that uses only rprocess and rmeasure is called “doubly plug-and-play”.
  • Two non-plug-and-play methods (expectation-maximization (EM) algorithms and Markov chain Monte Carlo (MCMC)) have theoretical convergence problems for nonlinear POMP models. The failures of these two workhorses of statistical computation have prompted development of alternative methodology.

Full-information and feature-based methods

  • Full-information methods are defined to be those based on the likelihood function for the full data (i.e., likelihood-based frequentist inference and Bayesian inference).
  • Feature-based methods either consider a summary statistic (a function of the data) or work with an an alternative to the likelihood.
  • Asymptotically, full-information methods are statistically efficient and feature-based methods are not.
  • In some cases, loss of statistical efficiency might be an acceptable tradeoff for advantages in computational efficiency.
  • However:
    • Good low-dimensional summary statistics can be hard to find.
    • When using statistically inefficient methods, it can be hard to know how much information you are losing.
    • Intuition and scientific reasoning can be inadequate tools to derive informative low-dimensional summary statistics (Shrestha et al. 2011,Ionides (2011)).

Bayesian and frequentist methods

  • Recently, plug-and-play Bayesian methods have been discovered:
    • particle Markov chain Monte Carlo (PMCMC) (Andrieu et al. 2010).
    • approximate Bayesian computation (ABC) (Toni et al. 2009).
  • Prior belief specification is both the strength and weakness of Bayesian methodology:
    • The likelihood surface for nonlinear POMP models often contains nonlinear ridges and variations in curvature.
    • These situations bring into question the appropriateness of independent priors derived from expert opinion on marginal distributions of parameters.
    • They also are problematic for specification of “flat” or “uninformative” prior beliefs.
    • Expert opinion can be treated as data for non-Bayesian analysis. However, our primary task is to identify the information in the data under investigation, so it can be helpful to use methods that do not force us to make our conclusions dependent on quantification of prior beliefs.

Full-information, plug-and-play, frequentist methods

  • Iterated filtering methods (Ionides et al. 2006,Ionides et al. (2015)) are the only currently available, full-information, plug-and-play, frequentist methods for POMP models.
  • Iterated filtering methods have been shown to solve likelihood-based inference problems for epidemiological situations which are computationally intractable for available Bayesian methodology (Ionides et al. 2015).

Summary of POMP inference methodologies

Frequentist Bayesian
Plug-and-play Full-information iterated filtering particle MCMC
Feature-based simulated moments ABC
synthetic likelihood (SL) SL-based MCMC
nonlinear forecasting  
Not plug-and-play Full-information EM algorithm MCMC
Kalman filter  
Feature-based Yule-Walker1 extended Kalman filter2
extended Kalman filter2  
  1. Yule-Walker is the method of moments for ARMA, a linear Gaussian POMP.
  2. The Kalman filter gives the exact likelihood for a linear Gaussian POMP. The extended Kalman filter gives an approximation for nonlinear models that can be used for quasi-likelihood or quasi-Bayesian inference.

An iterated filtering algorithm (IF2)

We focus on the IF2 algorithm of Ionides et al. (2015). In this algorithm:

In theory, this procedure converges toward the region of parameter space maximizing the maximum likelihood.
In practice, we can test this claim on examples.

IF2 algorithm pseudocode

Input:
Simulators for \(f_{X_0}(x_0;\theta)\) and \(f_{X_n|X_{n-1}}(x_n| x_{n-1}; \theta)\);
evaluator for \(f_{Y_n|X_n}(y_n| x_n;\theta)\);
data, \(y^*_{1:N}\)

Algorithmic parameters:
Number of iterations, \(M\);
number of particles, \(J\);
initial parameter swarm, \(\{\Theta^0_j, j=1,\dots,J\}\);
perturbation density, \(h_n(\theta|\varphi;\sigma)\);
perturbation scale, \(\sigma_{1{:}M}\)

Output:
Final parameter swarm, \(\{\Theta^M_j, j=1,\dots,J\}\)

Procedure:

  1. \(\quad\) For \(m\) in \(1{:} M\)
  2. \(\quad\quad\quad\) \(\Theta^{F,m}_{0,j}\sim h_0(\theta|\Theta^{m-1}_{j}; \sigma_m)\) for \(j\) in \(1{:} J\)
  3. \(\quad\quad\quad\) \(X_{0,j}^{F,m}\sim f_{X_0}(x_0 ; \Theta^{F,m}_{0,j})\) for \(j\) in \(1{:} J\)
  4. \(\quad\quad\quad\) For \(n\) in \(1{:} N\)
  5. \(\quad\quad\quad\quad\quad\) \(\Theta^{P,m}_{n,j}\sim h_n(\theta|\Theta^{F,m}_{n-1,j},\sigma_m)\) for \(j\) in \(1{:} J\)
  6. \(\quad\quad\quad\quad\quad\) \(X_{n,j}^{P,m}\sim f_{X_n|X_{n-1}}(x_n | X^{F,m}_{n-1,j}; \Theta^{P,m}_j)\) for \(j\) in \(1{:} J\)
  7. \(\quad\quad\quad\quad\quad\) \(w_{n,j}^m = f_{Y_n|X_n}(y^*_n| X_{n,j}^{P,m} ; \Theta^{P,m}_{n,j})\) for \(j\) in \(1{:} J\)
  8. \(\quad\quad\quad\quad\quad\) Draw \(k_{1{:}J}\) with \(P[k_j=i]= w_{n,i}^m\Big/\sum_{u=1}^J w_{n,u}^m\)
  9. \(\quad\quad\quad\quad\quad\) \(\Theta^{F,m}_{n,j}=\Theta^{P,m}_{n,k_j}\) and \(X^{F,m}_{n,j}=X^{P,m}_{n,k_j}\) for \(j\) in \(1{:} J\)
  10. \(\quad\quad\quad\) End For
  11. \(\quad\quad\quad\) Set \(\Theta^{m}_{j}=\Theta^{F,m}_{N,j}\) for \(j\) in \(1{:} J\)
  12. \(\quad\) End For

Remarks:

  • The \(N\) loop (lines 4 through 10) is a basic particle filter applied to a model with stochastic perturbations to the parameters.
  • The \(M\) loop repeats this particle filter with decreasing perturbations.
  • The superscript \(F\) in \(\Theta^{F,m}_{n,j}\) and \(X^{F,m}_{n,j}\) denote solutions to the filtering problem, with the particles \(j=1,\dots,J\) providing a Monte Carlo representation of the conditional distribution at time \(n\) given data \(y^*_{1:n}\) for filtering iteration \(m\).
  • The superscript \(P\) in \(\Theta^{P,m}_{n,j}\) and \(X^{P,m}_{n,j}\) denote solutions to the prediction problem, with the particles \(j=1,\dots,J\) providing a Monte Carlo representation of the conditional distribution at time \(n\) given data \(y^*_{1:n-1}\) for filtering iteration \(m\).
  • The weight \(w^m_{n,j}\) gives the likelihood of the data at time \(n\) for particle \(j\) in filtering iteration \(m\).

Analogy with evolution by natural selection

  • The parameters characterize the genotype.
  • The swarm of particles is a population.
  • The likelihood, a measure of the compatibility between the parameters and the data, is the analogue of fitness.
  • Each successive observation is a new generation.
  • Since particles reproduce in each generation in proportion to their likelihood, the particle filter acts like natural selection.
  • The artificial perturbations augment the “genetic” variance and therefore correspond to mutation.
  • IF2 increases the fitness of the population of particles.
  • However, because our scientific interest focuses on the model without the artificial perturbations, we decrease the intensity of the latter with successive iterations.

Applying IF2 to the boarding school influenza outbreak

For a relatively simple epidemiological example of IF2, we consider fitting a stochastic SIR model to an influenza outbreak in a British boarding school (Anonymous 1978). Reports consist of the number of children confined to bed for each of the 14 days of the outbreak. The total number of children at the school was 763, and a total of 512 children spent time away from class. Only one adult developed influenza-like illness, so adults are omitted from the data and model. First, we read in the data:

bsflu_data <- read.table("http://kingaa.github.io/sbied/stochsim/bsflu_data.txt")

Our model is a variation on a basic SIR Markov chain, with state \(X(t)=(S(t),I(t),R_1(t),R_2(t),R_3(t))\) giving the numbers of individuals in the susceptible and infectious categories, and three stages of recovery. The recovery stages, \(R_1\), \(R_2\) and \(R_3\), are all modeled to be non-contagious. \(R_1\) consists of individuals who are bed-confined if they showed symptoms; \(R_2\) consists of individuals who are convalescent if they showed symptoms; \(R_3\) consists of recovered individuals who have returned to school-work if they were symtomatic. The observation on day \(n\) of the observed epidemic (with \(t_1\) being 22 January) consists of the numbers of children who are bed-confined and convalescent. Ten individuals received antibiotics for secondary infections, and they had longer bed-confinement and convalescence times. Partly for this reason, and because our primary interest is in parameters related to transmission, we’ll narrow our focus to the bed-confinement numbers, \(B_n\), modeling these as \(B_n\sim{\mathrm{Poisson}\left(\rho R_1(t_n)\right)}\), where \(\rho\) is a reporting rate corresponding to the chance an infected boy is symptomatic.

Model flow diagram.

Model flow diagram.

The index case for the epidemic was purportedly a boy recently returned from Hong Kong, who was reported to have a transient febrile illness from 15 to 18 January. It would therefore be reasonable to initialize the epidemic at \(t_0=-6\) with \(I(t_0)=1\). This is a little tricky to reconcile with the rest of the data; for the moment, we avoid this issue by instead initializing with \(I(t_0)=1\) at \(t_0=0\). All other individuals are modeled to be initially susceptible.

Our Markov transmission model is that each individual in \(S\) transitions to \(I\) at rate \(\beta\,I(t)/N\); each individual in \(I\) transitions at rate \(\mu_I\) to \(R_1\). Subsequently, the individual moves from \(R_1\) to \(R_2\) at rate \(\mu_{R_1}\), and finally from \(R_2\) to \(R_3\) at rate \(\mu_{R_2}\). Therefore, \(1/\mu_I\) is the mean infectious time prior to bed-confinement; \(1/\mu_{R_1}\) is the mean duration of bed-confinement for symptomatic cases; \(1/\mu_{R_2}\) is the mean duration of convalescence for symptomatic cases. All rates have units \(\mathrm{day}^{-1}\).

This model has limitations and weaknesses, but writing down and fitting a model is a starting point for data analysis, not an end point. In particular, having fit one model, one should certainly try variations on that model. For example, one could include a latency period for infections, or one could modify the model to give a better description of the bed-confinement and convalescence processes.

We do not need a representation of \(R_3\) since this variable has consequences neither for the dynamics of the state process nor for the data. Since we are confining ourselves for the present to fitting only the \(B_n\) data, we need not track \(R_2\). We enumerate the state variables (\(S\), \(I\), \(R_1\)) and the parameters (\(\beta\), \(\mu_I\), \(\rho\), \(\mu_{R_1}\)) as follows:

statenames <- c("S","I","R1")
paramnames <- c("Beta","mu_I","mu_R1","rho")

In the codes below, we’ll refer to the data variables by their names (\(B\), \(C\)), as given in the bsflu_data data-frame:

colnames(bsflu_data)
## [1] "B"   "C"   "day"

Now, we write the model code:

dmeas <- Csnippet("
  lik = dpois(B,rho*R1+1e-6,give_log);
")

rmeas <- Csnippet("
  B = rpois(rho*R1+1e-6);
")

rproc <- Csnippet("
  double N = 763;
  double t1 = rbinom(S,1-exp(-Beta*I/N*dt));
  double t2 = rbinom(I,1-exp(-mu_I*dt));
  double t3 = rbinom(R1,1-exp(-mu_R1*dt));
  S  -= t1;
  I  += t1 - t2;
  R1 += t2 - t3;
")

init <- Csnippet("
 S = 762;
 I = 1;
 R1 = 0;
")

fromEst <- Csnippet("
 TBeta = exp(Beta);
 Tmu_I = exp(mu_I);
 Trho = expit(rho);
")

toEst <- Csnippet("
 TBeta = log(Beta);
 Tmu_I = log(mu_I);
 Trho = logit(rho);
")

Note that, in our measurement model, we’ve added a small positive number (\(10^{-6}\)) to the expected number of cases. Why is this useful? What complications does it introduce in the interpretation of results?

The fromEst and toEst C snippets implement parameter transformations that we’ll want soon.

Now we build the pomp object:

library(pomp)

pomp(
  data=subset(bsflu_data,select=-C),
  times="day",t0=0,
  rmeasure=rmeas,dmeasure=dmeas,
  rprocess=euler.sim(rproc,delta.t=1/12),
  initializer=init,
  fromEstimationScale=fromEst,toEstimationScale=toEst,
  statenames=statenames,
  paramnames=paramnames
) -> bsflu
plot(bsflu,main="")

Testing the codes.

To develop and debug code, it is useful to have testing codes that run quickly and fail if the codes are not working correctly. As such a test, here we run some simulations and a particle filter. We’ll use the following parameters, derived from our earlier explorations:

params <- c(Beta=2,mu_I=1,rho=0.9,mu_R1=1/3,mu_R2=1/2)

Now to run and plot some simulations:

y <- simulate(bsflu,params=params,nsim=10,as.data.frame=TRUE)
library(ggplot2)
theme_set(theme_bw())
library(reshape2)

ggplot(data=y,mapping=aes(x=time,y=B,group=sim))+
    geom_line()

Before engaging in iterated filtering, it is a good idea to check that the basic particle filter is working since iterated filtering builds on this technique. The simulations above check the rprocess and rmeasure codes; the particle filter depends on the rprocess and dmeasure codes and so is a check of the latter.

pf <- pfilter(bsflu,params=params,Np=1000)
plot(pf)

The above plot shows the data (B), along with the effective sample size of the particle filter (ess) and the log likelihood of each observation conditional on the preceding ones (cond.logLik).

Setting up the estimation problem.

Let’s treat \(\mu_{R_1}\) and \(\mu_{R_2}\) as known, and fix these parameters at the empirical means of the bed-confinement and convalescence times for symptomatic cases, respectively:

(fixed_params <- with(bsflu_data,c(mu_R1=1/(sum(B)/512),mu_R2=1/(sum(C)/512))))
##     mu_R1     mu_R2 
## 0.3324675 0.5541126

We will estimate \(\beta\), \(\mu_I\), and \(\rho\).

It will be helpful to parallelize most of the computations. Most machines nowadays have multiple cores and using this computational capacity is as simple as:

  1. letting R know you plan to use multiple processors;
  2. using the parallel for loop provided by the foreach package; and
  3. paying proper attention to the use of parallel random number generators.

For example:

library(foreach)
library(doParallel)
registerDoParallel()

The first two lines above load the foreach and doParallel packages, the latter being a “backend” for the foreach package. The next line tells foreach that we will use the doParallel backend. By default, R will guess how many cores are available and will run about half this number of concurrent R processes.

Running a particle filter.

We proceed to carry out replicated particle filters at an initial guess of \(\beta=2\), \(\mu_I=1\), and \(\rho=0.9\).

library(doRNG)
registerDoRNG(625904618)
bake(file="pf.rds",{
  foreach(i=1:10,.packages='pomp',
    .export=c("bsflu","fixed_params")
  ) %dopar% {
    pfilter(bsflu,params=c(Beta=2,mu_I=1,rho=0.9,fixed_params),Np=10000)
  }
}) -> pf
(L_pf <- logmeanexp(sapply(pf,logLik),se=TRUE))
##                      se 
## -87.4037461   0.8487475

In 1.35 seconds, using 4 cores, we obtain an unbiased likelihood estimate of -87.4 with a Monte Carlo standard error of 0.85.

Building up a picture of the likelihood surface

Given a model and a set of data, the likelihood surface is well defined, though it may be difficult to visualize. We can develop a progressively more complete picture of this surface by storing likelihood estimates whenever we compute them. In particular, it is a very good idea to set up a database within which to store the likelihood of every point for which we have an estimated likelihood. This will become larger and more complete as our parameter-space search goes on and will be a basis for a variety of explorations. At this point, we’ve computed the likelihood at a single point. Let’s store this point, together with the estimated likelihood and our estimate of the standard error on that likelihood, in a CSV file:

results <- as.data.frame(as.list(c(coef(pf[[1]]),loglik=L_pf[1],loglik=L_pf[2])))
write.csv(results,file="bsflu_params.csv",row.names=FALSE)

A local search of the likelihood surface

Let’s carry out a local search using mif2 around this point in parameter space. To do so, we need to choose the rw.sd and cooling.fraction.50 algorithmic parameters. Since \(\beta\) and \(\mu_I\) will be estimated on the log scale, and we expect that multiplicative perturbations of these parameters will have roughly similar effects on the likelihood, we’ll use a perturbation size of \(0.02\), which we imagine will have a small but non-negligible effect. For simplicity, we’ll use the same perturbation size on \(\rho\). We fix cooling.fraction.50=0.5, so that after 50 mif2 iterations, the perturbations are reduced to half their original magnitudes.

registerDoRNG(482947940)
bake(file="box_search_local.rds",{
  foreach(i=1:20,
    .packages='pomp',
    .combine=c, 
    .export=c("bsflu","fixed_params")
  ) %dopar%  
  {
    mif2(
      bsflu,
      start=c(Beta=2,mu_I=1,rho=0.9,fixed_params),
      Np=2000,
      Nmif=50,
      cooling.type="geometric",
      cooling.fraction.50=0.5,
      transform=TRUE,
      rw.sd=rw.sd(Beta=0.02,mu_I=0.02,rho=0.02)
    )
  }
}) -> mifs_local

We obtain some diagnostic plots with the plot command applied to mifs_local. Here is a way to get a prettier version:

ggplot(data=melt(conv.rec(mifs_local)),
       aes(x=iteration,y=value,group=L1,color=factor(L1)))+
  geom_line()+
  guides(color=FALSE)+
  facet_wrap(~variable,scales="free_y")+
  theme_bw()

No filtering failures (nfail) are generated at any point, which is comforting. In general, we expect to see filtering failures whenever our initial guess (start) is incompatible with one or more of the observations. Filtering failures at the MLE are an indication that the model, at its best, is incompatible with one or more of the data.

We see that the likelihood generally increases as the iterations proceed, though there is considerable variability due to the stochastic nature of this Monte Carlo algorithm. Although the filtering carried out by mif2 in the final filtering iteration generates an approximation to the likelihood at the resulting point estimate, this is not usually good enough for reliable inference. Partly, this is because parameter perturbations are applied in the last filtering iteration, so that the likelhood shown here is not identical to that of the model of interest. Partly, this is because mif2 is usually carried out with fewer particles than are needed for a good likelihood evaluation: the errors in mif2 average out over many iterations of the filtering. Therefore, we evaluate the likelihood, together with a standard error, using replicated particle filters at each point estimate:

registerDoRNG(900242057)
bake(file="lik_local.rds",{
  foreach(mf=mifs_local,.packages='pomp',.combine=rbind) %dopar% 
  {
    evals <- replicate(10, logLik(pfilter(mf,Np=20000)))
    ll <- logmeanexp(evals,se=TRUE)
    c(coef(mf),loglik=ll[1],loglik=ll[2])
  }
}) -> results_local
results_local <- as.data.frame(results_local)

This investigation took 12 sec for the maximization and 18 sec for the likelihood evaluation. These repeated stochastic maximizations can also show us the geometry of the likelihood surface in a neighborhood of this point estimate:

pairs(~loglik+Beta+mu_I+rho,data=results_local,pch=16)

Although this plot some hints of ridges in the likelihood surface (cf. the \(\beta\)-\(\mu_I\) panel), the sampling is still too sparse to give a clear picture.

We add these newly explored points to our database:

results <- rbind(results,results_local[names(results)])
write.csv(results,file="bsflu_params.csv",row.names=FALSE)

A global search of the likelihood surface using randomized starting values

When carrying out parameter estimation for dynamic systems, we need to specify beginning values for both the dynamic system (in the state space) and the parameters (in the parameter space). To avoid confusion, we use the term “initial values” to refer to the state of the system at \(t_0\) and “starting values” to refer to the point in parameter space at which a search is initialized.

Practical parameter estimation involves trying many starting values for the parameters. One way to approach this is to choose a large box in parameter space that contains all remotely sensible parameter vectors. If an estimation method gives stable conclusions with starting values drawn randomly from this box, this gives some confidence that an adequate global search has been carried out.

For our flu model, a box containing reasonable parameter values might be

params_box <- rbind(
  Beta=c(1,5),
  mu_I=c(0.5,3),
  rho = c(0.5,1)
)

We are now ready to carry out likelihood maximizations from diverse starting points.

registerDoRNG(1270401374)
guesses <- as.data.frame(apply(params_box,1,function(x)runif(300,x[1],x[2])))
mf1 <- mifs_local[[1]]
bake(file="box_search_global.rds",{
  foreach(guess=iter(guesses,"row"), 
    .packages='pomp', 
    .combine=rbind,
    .options.multicore=list(set.seed=TRUE),
    .export=c("mf1","fixed_params")
  ) %dopar% 
  {
    mf <- mif2(mf1,start=c(unlist(guess),fixed_params))
    mf <- mif2(mf,Nmif=100)
    ll <- replicate(10,logLik(pfilter(mf,Np=100000)))
    ll <- logmeanexp(ll,se=TRUE)
    c(coef(mf),loglik=ll[1],loglik=ll[2])
  }
}) -> results_global
results_global <- as.data.frame(results_global)
results <- rbind(results,results_global[names(results)])
write.csv(results,file="bsflu_params.csv",row.names=FALSE)

The above codes run one search from each of 300 starting values. Each search consissts of an initial run of 51 IF2 iterations, followed by another 100 iterations. These codes exhibit a general pomp behavior: re-running a command on an object (i.e., mif2 on mf1) created by the same command preserves the algorithmic arguments. In particular, running mif2 on the result of a mif2 computation re-runs IF2 from the endpoint of the first run. In the second computation, by default, all algorithmic parameters are preserved; here we overrode the default choice of Nmif.

Following the mif2 computations, the particle filter is used to evaluate the likelihood, as before. In contract to the local-search codes above, here we return only the endpoint of the search, together with the likelihood estimate and its standard error in a named vector. The best result of this search had a likelihood of -73.2 with a standard error of 1.38. This took 22.6 minutes altogether using 4 processors.

Again, we attempt to visualize the global geometry of the likelihood surface using a scatterplot matrix. In particular, here we plot both the starting values (grey) and the IF2 estimates (red).

library(plyr)
all <- ldply( list(guess=guesses, result=subset(results, loglik > max(loglik)-50) ), .id="type")
pairs(~loglik+Beta+mu_I+rho, data=all, col=ifelse(all$type=="guess", grey(0.5), "red"), pch=16)

We see that optimization attempts from diverse remote starting points converge on a particular region in parameter space. Moreover, the estimates have comparable likelihoods, despite their considerable variability. This gives us some confidence in our maximization procedure.

Exercises


Basic Exercise: Fitting the SEIR3 model

Following the template above, estimate the parameters and likelihood of the SEIR3 model you implemented in the earlier lessons. Specifically, first conduct a local search and then a global search using the multi-stage, multi-start method displayed above.

How does the maximized likelihood compare with what we obtained for the SIR3 model? How do the parameter estimates differ?

You will need to tailor the intensity of your search to the computational resources at your disposal. In particular, choose the number of starts, number of particles employed, and the number of IF2 iterations to perform in view of the size and speed of your machine.


Extra Exercise: Modify the measurement model

The Poisson measurement model used here may not seem easy to interpret. Formulate an alternative measurement model and maximize the likelihood to compare the alternative model.


Extra Exercise: Construct a profile likelihood

How strong is the evidence about the contact rate, \(\beta\), given this model and data? Use mif2 to construct a profile likelihood. Due to time constraints, you may be able to compute only a preliminary version.

It is also possible to profile over the basic reproduction number, \(R_0=\beta /\mu_I\). Is this more or less well determined that \(\beta\) for this model and data?


Extra Exercise: Checking the source code

Check the source code for the bsflu pomp object. Does the code implement the model described?

For various reasons, it can be surprisingly hard to make sure that the written equations and the code are perfectly matched. Here are some things to think about:

  1. Papers should be written to be readable. Code must be written to run successfully. People rarely choose to clutter papers with numerical details which they hope and believe are scientifically irrelevant. What problems can arise due to the conflict between readability and reproducibility? What solutions are available?
  2. Suppose that there is an error in the coding of rprocess and suppose that plug-and-play statistical methodology is used to infer parameters. As a conscientious researcher, you carry out a simulation study to check the soundness of your inference methodology on this model. To do this, you use simulate to generate realizations from the fitted model and checking that your parameter inference procedure recovers the known parameters, up to some statistical error. Will this procedure help to identify the error in rprocess? If not, how might you debug rprocess? What research practices help minimize the risk of errors in simulation code?

Choosing the algorithmic settings for IF2

More exercises


Optional Exercise: Assessing and improving algorithmic parameters

Develop your own heuristics to try to improve the performance of mif2 in the previous example. Specifically, for a global optimization procedure carried out using random starting values in the specified box, let \(\hat\Theta_{\mathrm{max}}\) be a random Monte Carlo estimate of the resulting MLE, and let \(\hat\theta\) be the true (unknown) MLE. We can define the maximization error in the log likelihood to be \[e = \ell(\hat\theta) - E[\ell(\hat\Theta_{\mathrm{max}})].\] We cannot directly evaluate \(e\), since there is also Monte Carlo error in our evaluation of \(\ell(\theta)\), but we can compute it up to a known precision. Plan some code to estimates \(e\) for a search procedure using a computational effort of \(JM=2\times 10^7\), comparable to that used for each mif computation in the global search. Discuss the strengths and weaknesses of this quantification of optimization success. See if you can choose \(J\) and \(M\) subject to this constraint, together with choices of rw.sd and the cooling rate, cooling.fraction.50, to arrive at a quantifiably better procedure. Computationally, you may not be readily able to run your full procedure, but you could run a quicker version of it.


Optional Exercise: Finding sharp peaks in the likelihood surface

Even in this small, 3 parameter example, it takes a considerable amount of computation to find the global maximum (with values of \(\beta\) around 0.004) starting from uniform draws in the specified box. The problem is that, on the scale on which “uniform” is defined, the peak around \(\beta\approx 0.004\) is very narrow. Propose and test a more favorable way to draw starting parameters for the global search, with better scale invariance properties.


Back to course homepage

R codes for this document


References

Andrieu, C., A. Doucet, and R. Holenstein. 2010. Particle Markov chain Monte Carlo methods. Journal of the Royal Statistical Society, Series B 72:269–342.

Anonymous. 1978. Influenza in a boarding school. British medical journal 1:587.

He, D., E. L. Ionides, and A. A. King. 2010. Plug-and-play inference for disease dynamics: Measles in large and small populations as a case study. Journal of the Royal Society, Interface 7:271–283.

Ionides, E. L. 2011. Discussion on “Feature matching in time series modeling” by Y. Xia and H. Tong. Statistical Science 26:49–52.

Ionides, E. L., C. Bretó, and A. A. King. 2006. Inference for nonlinear dynamical systems. Proceedings of the National Academy of Sciences of the U.S.A. 103:18438–18443.

Ionides, E. L., D. Nguyen, Y. Atchadé, S. Stoev, and A. A. King. 2015. Inference for dynamic and latent variable models via iterated, perturbed Bayes maps. Proceedings of the National Academy of Sciences of the U.S.A. 112:719–724.

King, A. A., D. Nguyen, and E. L. Ionides. 2016. Statistical inference for partially observed Markov processes via the R package pomp. Journal of Statistical Software 69:1–43.

Shrestha, S., A. A. King, and P. Rohani. 2011. Statistical inference for multi-pathogen systems. PLoS Computational Biology 7:e1002135.

Toni, T., D. Welch, N. Strelkowa, A. Ipsen, and M. P. Stumpf. 2009. Approximate Bayesian computation scheme for parameter inference and model selection in dynamical systems. Journal of the Royal Society Interface 6:187–202.