
rbmi: Statistical Specifications
Alessandro Noci, Craig Gower-Page, and Marcel Wolbers
Source:vignettes/stat_specs.Rmd
stat_specs.Rmd
1 Scope of this document
This document describes the statistical methods implemented in the rbmi
R package for standard and reference-based multiple imputation of continuous longitudinal outcomes.
The package implements three classes of multiple imputation (MI) approaches:
Conventional MI methods based on Bayesian (or approximate Bayesian) posterior draws of model parameters combined with Rubin’s rules to make inferences as described in Carpenter, Roger, and Kenward (2013) and Cro et al. (2020).
Conditional mean imputation methods combined with re-sampling techniques as described in Wolbers et al. (2022).
Bootstrapped MI methods as described in von Hippel and Bartlett (2021).
The document is structured as follows: we first provide an informal introduction to estimands and corresponding treatment effect estimation based on MI (section 2). The core of this document consists of section 3 which describes the statistical methodology in detail and also contains a comparison of the implemented approaches (section 3.10). The link between theory and the functions included in package rbmi
is described in section 4. We conclude with a comparison of our package to some alternative software implementations of reference-based imputation methods (section 5).
2 Introduction to estimands and estimation methods
2.1 Estimands
The ICH E9(R1) addendum on estimands and sensitivity analyses describes a systematic approach to ensure alignment among clinical trial objectives, trial execution/conduct, statistical analyses, and interpretation of results (ICH E9 working group (2019)). As per the addendum, an estimand is a precise description of the treatment effect reflecting the clinical question posed by the trial objective which summarizes at a population-level what the outcomes would be in the same patients under different treatment conditions being compared. One important attribute of an estimand is a list of possible intercurrent events (ICEs), i.e. of events occurring after treatment initiation that affect either the interpretation or the existence of the measurements associated with the clinical question of interest, and the definition of appropriate strategies to deal with ICEs. The three most relevant strategies for the purpose of this document are the hypothetical strategy, the treatment policy strategy, and the composite strategy. For the hypothetical strategy, a scenario is envisaged in which the ICE would not occur. Under this scenario, endpoint values after the ICE are not directly observable and treated using models for missing data. For the treatment policy strategy, the treatment effect in the presence of the ICEs is targeted and analyses are based on the observed outcomes regardless whether the subject had an ICE or not. For the composite strategy, the ICE itself is included as a component of the endpoint.
2.2 Alignment between the estimand and the estimation method
The ICH E9(R1) addendum distinguishes between ICEs and missing data (ICH E9 working group (2019)). Whereas ICEs such as treatment discontinuations reflect clinical practice, the amount of missing data can be minimized in the conduct of a clinical trial. However, there are many connections between missing data and ICEs. For example, it is often difficult to retain subjects in a clinical trial after treatment discontinuation and a subject’s dropout from the trial leads to missing data. As another example, outcome values after ICEs addressed using a hypothetical strateg are not directly observable under the hypothetical scenario. Consequently, any observed outcome values after such ICEs are typically discarded and treated as missing data.
The addendum proposes that estimation methods to address the problem presented by missing data should be selected to align with the estimand. A recent overview of methods to align the estimator with the estimand is Mallinckrodt et al. (2020). A short introduction on estimation methods for studies with longitudinal endpoints can also be found in Wolbers et al. (2022). One prominent statistical method for this purpose is multiple imputation (MI), which is the target of the rbmi
package.
2.2.1 Missing data prior to ICEs
Missing data may occur in subjects without an ICE or prior to the occurrence of an ICE. As such missing outcomes are not associated with an ICE, it is often plausible to impute them under a missing-at-random (MAR) assumption using a standard MMRM imputation model of the longitudinal outcomes. Informally, MAR occurs if the missing data can be fully accounted for by the baseline variables included in the model and the observed longitudinal outcomes, and if the model is correctly specified.
2.2.2 Implementation of the hypothetical strategy
The MAR imputation model described above is often also a good starting point for imputing data after an ICE handled using a hypothetical strategy (Mallinckrodt et al. (2020)).
Informally, this assumes that unobserved values after the ICE would have been similar to the observed data from subjects who did not have the ICE and remained under follow-up.
However, in some situations, it may be more reasonable to assume that missingness is “informative” and indicates a systematically better or worse outcome than in observed subjects. In such situations, MNAR imputation with a rbmi
only fixed
2.2.3 Implementation of the treatment policy strategy
Ideally, data collection continues after an ICE handled with a treatment policy strategy and no missing data arises. Indeed, such post-ICE data are increasingly systematically collected in RCTs. However, despite best efforts, missing data after an ICE such as study treatment discontinuation may still occur because the subject drops out from the study after discontinuation. It is difficult to give definite recommendations regarding the implementation of the treatment policy strategy in the presence of missing data at this stage because the optimal method is highly context dependent and a topic of ongoing statistical research.
For ICEs which are thought to have a negligible effect on efficacy outcomes, standard MAR-based imputation which ignores whether an outcome is observed pre- or post-ICE may be appropriate.
In contrast, an ICE such as treatment discontinuation may be expected to have a more substantial impact on efficacy outcomes. In such settings, the MAR assumption may still be plausible after conditioning on the subject’s time-varying treatment status (Guizzaro et al. (2021)). In this case, one option is to impute missing post-discontinuation data based on subjects who also discontinued treatment but continued to be followed up. Another option which may require somewhat less post-discontinuation data is to include all subjects in the imputation procedure but to model post-discontinuation data by using time-varying treatment status indicators (Guizzaro et al. (2021), Polverejan and Dragalin (2020), Noci et al. (2023), Drury et al. (2024), Bell et al. (2024)). In this approach, post-ICE outcomes are included in every step of the analysis, including in the fitting of the imputation model. It assumes that ICEs may impact post-ICE outcomes but that otherwise missingness is non-informative. The approach also assumes that the time-varying covariates do not contain missing values, deviations in outcomes after the ICE are correctly modeled by these time-varying covariates, and that sufficient post-ICE data are available to inform the regression coefficients of the time-varying covariates.
The resulting imputation models are called “retrieved dropout models” in the statistical literature. These models tend to have less bias than alternative analysis approaches based on imputation under a basic MAR assumption or a reference-based missing data assumption. However, retrieved dropout models have been associated with inflated standard errors of associated treatment effect estimators which has a detrimental effect on study power. In particular, it has been observed that once the post-ICE observation percentages falls below 50%, the power loss can be quite dramatic (Bell et al. 2024). We illustrate the implementation of retrieved dropout models in the vignette “Implementation of retrieved-dropout models using rbmi” (vignette(topic = "retrieved_dropout", package = "rbmi")
).
In some trial settings, only few subjects discontinue the randomized treatment. In other settings, treatment discontinuation rates are higher but it is difficult to retain subjects in the trial after treatment discontinuation leading to sparse data collection after treatment discontinuation. In both settings, the amount of available data after treatment discontinuation may be insufficient to inform an imputation model which explicitly models post-discontinuation data. Depending on the disease area and the anticipated mechanism of action of the intervention, it may be plausible to assume that subjects in the intervention group behave similarly to subjects in the control group after the ICE treatment discontinuation. In this case, reference-based imputation methods are an option (Mallinckrodt et al. (2020)). Reference-based imputation methods formalize the idea to impute missing data in the intervention group based on data from a control or reference group. For a general description and review of reference-based imputation methods, we refer to Carpenter, Roger, and Kenward (2013), Cro et al. (2020), I. White, Royes, and Best (2020) and Wolbers et al. (2022). For a technical description of the implemented statistical methodology for reference-based imputation, we refer to section 3 (in particular section 3.4).
2.2.4 Implementation of the composite strategy
The composite strategy is typically applied to binary or time-to-event outcomes but it can also be used for continuous outcomes by ascribing a suitably unfavorable value to patients who experience ICEs for which a composite strategy has been defined. One possibility to implement this is to use MI with a
3 Statistical methodology
3.1 Overview of the imputation procedure
Analyses of datasets with missing data always rely on missing data assumptions. The methods described here can be used to produce valid imputations under a MAR assumption or under reference-based imputation assumptions. MNAR imputation based on fixed
Three general imputation approaches are implemented in rbmi
:
Conventional MI based on Bayesian (or approximate Bayesian) posterior draws from the imputation model combined with Rubin’s rules for inference as described in Carpenter, Roger, and Kenward (2013) and Cro et al. (2020).
Conditional mean imputation based on the REML estimate of the imputation model combined with resampling techniques (the jackknife or the bootstrap) for inference as described in Wolbers et al. (2022).
Bootstrapped MI methods based on REML estimates of the imputation model as described in von Hippel and Bartlett (2021).
3.1.1 Conventional MI
Conventional MI approaches include the following steps:
- Base imputation model fitting step (Section 3.3)
Fit a Bayesian multivariate normal mixed model for repeated measures (MMRM) to the observed longitudinal outcomes after exclusion of data after ICEs for which reference-based missing data imputation is desired (Section 3.3.3). Draw
posterior samples of the estimated parameters (regression coefficients and covariance matrices) from this model.Alternatively,
approximate posterior draws from the posterior distribution can be sampled by repeatedly applying conventional restricted maximum-likelihood (REML) parameter estimation of the MMRM model to nonparametric bootstrap samples from the original dataset (Section 3.3.4).
- Imputation step (Section 3.4)
Take a single sample
( from the posterior distribution of the imputation model parameters.For each subject, use the sampled parameters and the defined imputation strategy to determine the mean and covariance matrix describing the subject’s marginal outcome distribution for all longitudinal outcome assessments (i.e. observed and missing outcomes).
For each subjects, construct the conditional multivariate normal distribution of their missing outcomes given their observed outcomes (including observed outcomes after ICEs for which a reference-based assumption is desired).
For each subject, draw a single sample from this conditional distribution to impute their missing outcomes leading to a complete imputed dataset.
For sensitivity analyses, a pre-defined
-adjustment may be applied to the imputed data prior to the analysis step. (Section 3.5).
- Analysis step (Section 3.6)
- Analyze the imputed dataset using an analysis model (e.g. ANCOVA) resulting in a point estimate and a standard error (with corresponding degrees of freedom) of the treatment effect.
- Pooling step for inference (Section 3.7)
- Repeat steps 2. and 3. for each posterior sample
, resulting in complete datasets, point estimates of the treatment effect, and standard errors (with corresponding degrees of freedom). Pool the treatment effect estimates, standard errors, and degrees of freedom using the rules by Barnard and Rubin to obtain the final pooled treatment effect estimator, standard error, and degrees of freedom.
3.1.2 Conditional mean imputation
The conditional mean imputation approach includes the following steps:
- Base imputation model fitting step (Section 3.3)
- Fit a conventional multivariate normal/MMRM model using restricted maximum likelihood (REML) to the observed longitudinal outcomes after exclusion of data after ICEs for which reference-based missing data imputation is desired (Section 3.3.2).
- Imputation step (Section 3.4)
For each subject, use the fitted parameters from step 1. to construct the conditional distribution of missing outcomes given observed outcomes (including observed outcomes after ICEs for which reference-based missing data imputation is desired) as described above.
For each subject, impute their missing data deterministically by the mean of this conditional distribution leading to a complete imputed dataset.
For sensitivity analyses, a pre-defined
-adjustment may be applied to the imputed data prior to the analysis step. (Section 3.5).
- Analysis step (Section 3.6)
- Apply an analysis model (e.g. ANCOVA) to the completed dataset resulting in a point estimate of the treatment effect.
- Jackknife or bootstrap inference step (Section 3.8)
- Inference for the treatment effect estimate from 3. is based on re-sampling techniques. Both the jackknife and the bootstrap are supported. Importantly, these methods require repeating all steps of the imputation procedure (i.e. imputation, conditional mean imputation, and analysis steps) on each of the resampled datasets.
3.1.3 Bootstrapped MI
The bootstrapped MI approach includes the following steps:
- Base imputation model fitting step (Section 3.3)
- Apply conventional restricted maximum-likelihood (REML) parameter estimation of the MMRM model to
nonparametric bootstrap samples from the original dataset using the observed longitudinal outcomes after exclusion of data after ICEs for which reference-based missing data imputation is desired.
- Imputation step (Section 3.4)
Take a bootstrapped dataset
( and its corresponding imputation model parameter estimates.For each subject (from the bootstrapped dataset), use the parameter estimates and the defined strategy for dealing with their ICEs to determine the mean and covariance matrix describing the subject’s marginal outcome distribution for all longitudinal outcome assessments (i.e. observed and missing outcomes).
For each subjects (from the bootstrapped dataset), construct the conditional multivariate normal distribution of their missing outcomes given their observed outcomes (including observed outcomes after ICEs for which reference-based missing data imputation is desired).
For each subject (from the bootstrapped dataset), draw
samples from this conditional distributions to impute their missing outcomes leading to complete imputed dataset for bootstrap sample .For sensitivity analyses, a pre-defined
-adjustment may be applied to the imputed data prior to the analysis step. (Section 3.5).
- Analysis step (Section 3.6)
- Analyze each of the
imputed datasets using an analysis model (e.g. ANCOVA) resulting in point estimates of the treatment effect.
- Pooling step for inference (Section 3.9)
- Pool the
treatment effect estimates as described in von Hippel and Bartlett (2021) to obtain the final pooled treatment effect estimate, standard error, and degrees of freedom.
3.2 Setting, notation, and missing data assumptions
Assume that the data are from a study with rbmi
implementation.
Denote the observed outcome vector of length rbmi
. Therefore, if missing data following an ICE are to be handled using MAR imputation, this is compatible with the default assumption. As discussed in Section 2, the MAR assumption is often a good starting point for implementing a hypothetical strategy. But also note that observed outcome data after an ICE handled using a hypothetical strategy is not compatible with this strategy. Therefore, we assume that all post-ICE data after ICEs handled using a hypothetical strategy are already set to NA in rbmi
functions. However, any observed outcomes after ICEs handled using a treatment policy strategy should be included in
Subjects may also experience up to one ICE after which missing data imputation according to a reference-based imputation method is foreseen. For a subject
MNAR
3.3 The base imputation model
3.3.1 Included data and model specification
The purpose of the imputation model is to estimate (covariate-dependent) mean trajectories and covariance matrices for each group in the absence of ICEs handled using reference-based imputation methods. Conventionally,
publications on reference-based imputation methods have implicitly assumed that the corresponding post-ICE
data is missing for all subjects (Carpenter, Roger, and Kenward (2013)). We also allow the situation where post-ICE data
is available for some subjects but needs to be imputed using reference-based methods for others. However,
any observed data after ICEs for which reference-based imputation methods are specified is not compatible
with the imputation model described below and they are therefore removed and considered as missing for
the purpose of estimating the imputation model, and for this purpose only. For example, if a patient has an ICE addressed with a reference-based method but outcomes after the ICE are collected, these post-ICE outcomes will be excluded when fitting the base imputation model (but they will be included again in the following steps).
That is, the base imputation model is fitted to
Observed post-ICE outcomes in the control or reference group are also excluded from the base imputation model if the user specifies a reference-based imputation strategy for such ICEs. This ensures that an ICE has the same impact on the data included in the imputation model regardless whether the ICE occurred in the control or the intervention group. On the other hand, imputation in the reference group is based on a MAR assumption even for reference-based imputation methods and it may be preferable in some settings to include post-ICE data from the control group in the base imputation model. This can be implemented by specifying a MAR
strategy for the ICE in the control group and a reference-based strategy for the same ICE in the intervention group.
The base imputation model of the longitudinal outcomes
Denote the
Typically, a common unstructured covariance matrix for all subjects is assumed for
3.3.2 Restricted maximum likelihood estimation (REML)
Frequentist parameter estimation for the base imputation is based on REML. The use of REML as an improved alternative to maximum likelihood (ML) for covariance parameter estimation was originally proposed by Patterson and Thompson (1971). Since then, it has become the default method for parameter estimation in linear mixed effects models. rbmi
allows to choose between ML and REML methods to estimate the model parameters, with REML being the default option.
3.3.3 Bayesian model fitting
The Bayesian imputation model is fitted with the R package rstan
(Stan Development Team (2020)). rstan
is the R interface of Stan. Stan is a powerful and flexible statistical software developed by a dedicated team and implements Bayesian inference with state-of-the-art MCMC sampling procedures. The multivariate normal model with missing data specified in section 3.3.1 can be considered a generalization of the models described in the Stan user’s guide (see Stan Development Team (2020, sec. 3.5)).
The same prior distributions as in the SAS implementation of the “five macros” are used (Roger (2021)), i.e. an improper flat priors for the regression coefficients and a weakly informative inverse Wishart prior for the covariance matrix (or matrices). Specifically, let
As in the “five macros”, the MCMC algorithm is initialized at the parameters from a frequentist REML fit (see section 3.3.2). As described above, we are using only weakly informative priors for the parameters. Therefore, the Markov chain is essentially starting from the targeted stationary posterior distribution and only a minimal amount of burn-in of the chain is required.
The initial values and other expert options for the MCMC sampling can be defined via the control
argument to method_bayes
, which is simplified with the corresponding control_bayes()
function call.
With the default initial values, i.e. when using init = "mmrm"
in the control
list, then for obtaining reproducible results an external set.seed()
call is required before running draws()
. In particular, please note that it is not sufficient to just set the seed
option in the control
list.
If more than one chain is being run, the chains are currently initialized at "random"
values as per rstan
default. Note that for consistency with rstan
, the burn-in can be customized via the warmup
argument, and the burn-between can be customized via the thin
argument.
Independent of the number of chains used, the total number of returned samples is always n_samples
. In the case that more than one chain is used, these samples are distributed across the multiple chains appropriately.
3.3.4 Approximate Bayesian posterior draws via the bootstrap
Several authors have suggested that a stabler way to get Bayesian posterior draws from the imputation model is to bootstrap the incomplete data and to calculate REML estimates for each bootstrap sample (Little and Rubin (2002), Efron (1994), Honaker and King (2010), von Hippel and Bartlett (2021)). This method is proper in that the REML estimates from the bootstrap samples are asymptotically equivalent to a sample from the posterior distribution and may provide additional robustness to model misspecification (Little and Rubin (2002, sec. 10.2.3, part 6), Honaker and King (2010)). In order to retain balance between treatment groups and stratification factors across bootstrap samples, the user is able to provide stratification variables for the bootstrap in the rbmi
implementation.
3.4 Imputation step
3.4.1 Marginal imputation distribution for a subject - MAR case
For each subject
3.4.2 Marginal imputation distribution for a subject - reference-based imputation methods
For each subject
Based on these means and covariance matrices, the subject’s marginal imputation distribution for the reference-based imputation methods is then calculated as detailed in Carpenter, Roger, and Kenward (2013, sec. 4.3).
Denote the mean and covariance matrix of this marginal imputation distribution by
Jump to reference (JR): the patient’s outcome distribution is normally distributed with the following mean:
The covariance matrix is constructed as follows. First, we partition the covariance matrices and in blocks according to the time of the ICE : We want the covariance matrix to match for the pre-deviation measurements, and for the conditional components for the post-deviation given the pre-deviation measurements. The solution is derived in Carpenter, Roger, and Kenward (2013, sec. 4.3) and is given by:Copy increments in reference (CIR): the patient’s outcome distribution is normally distributed with the following mean:
The covariance matrix is derived as for the JR method.Copy reference (CR): the patient’s outcome distribution is normally distributed with mean and covariance matrix taken from the reference group:
Last mean carried forward (LMCF): the patient’s outcome distribution is normally distributed with the following mean:
and covariance matrix:
3.4.3 Imputation of missing outcome data
The joint marginal multivariate normal imputation distribution of subject
The conditional distribution used for the imputation is again a multivariate normal distribution and explicit formulas for the conditional mean and covariance are readily available. For completeness, we report them here with the notation and terminology of our setting. The marginal distribution for the outcome of patient
Conventional random imputation consists in sampling from this conditional multivariate normal distribution. Conditional mean imputation imputes missing values with the deterministic conditional expectation
3.5 -adjustment
A marginal
The implementation provides full flexibility regarding the specific implementation of the
3.6 Analysis step
After data imputation, a standard analysis model can be applied to the completed data resulting in a treatment effect estimate. As the imputed data no longer contains missing values, the analysis model is often simple. For example, it can be an analysis of covariance (ANCOVA) model with the outcome (or the change in the outcome from baseline) at a specific visit j as the dependent variable, the randomized treatment group as the primary covariate and, typically, adjustment for the same baseline covariates as for the imputation model.
3.7 Pooling step for inference of (approximate) Bayesian MI and Rubin’s rules
Assume that the analysis model has been applied to
Rubin’s rules are used for pooling the treatment effect estimates and corresponding variances estimates from the analysis steps across the
Confidence intervals and tests of the null hypothesis
3.8 Bootstrap and jackknife inference for conditional mean imputation
3.8.1 Point estimate of the treatment effect
The point estimator is obtained by applying the analysis model (Section 3.6) to a single conditional mean imputation of the missing data (see Section 3.4.3) based on the REML estimator of the parameters of the imputation model (see Section 3.3.2). We denote this treatment effect estimator by
As demonstrated in Wolbers et al. (2022) (Section 2.4), this treatment effect estimator is valid if the analysis model is an ANCOVA model or, more generally, if the treatment effect estimator is a linear function of the imputed outcome vector. Indeed, if this is the case, then the estimator is identical to the pooled treatment effect across multiple random REML imputation with an infinite number of imputations and corresponds to a computationally efficient implementation of a proposal by von Hippel and Bartlett (2021). We expect that the conditional mean imputation method is also applicable to some other analysis models (e.g. for general MMRM analysis models) but this has not been formally justified.
3.8.2 Jackknife standard errors, confidence intervals (CI) and tests for the treatment effect
For a dataset containing
Then, the jackknife standard error is defined as
A simulation study reported in Wolbers et al. (2022) demonstrated exact protection of the type I error for jackknife-based inference with a relatively low sample size (n = 100 per group) and a substantial amount of missing data (>25% of subjects with an ICE).
3.8.3 Bootstrap standard errors, confidence intervals (CI) and tests for the treatment effect
As an alternative to the jackknife, the bootstrap has also been implemented in rbmi
(Efron and Tibshirani (1994), Davison and Hinkley (1997)).
Two different bootstrap methods are implemented in rbmi
: Methods based on the bootstrap standard error and the normal approximation and percentile bootstrap methods. Denote the treatment effect estimates from rbmi
via inversion of the confidence interval. Explicit formulas for bootstrap inference as implemented in the rbmi
package and some considerations regarding the required number of bootstrap samples are included in the Appendix of Wolbers et al. (2022).
A simulation study reported in Wolbers et al. (2022) demonstrated a small inflation of the type I error rate for inference based on the bootstrap standard error (up to
3.9 Pooling step for inference of the bootstrapped MI methods
Assume that the analysis model has been applied to
The final estimate of the treatment effect is calculated as the sample mean over the
where
The degrees of freedom are estimated with the following formula (von Hippel and Bartlett (2021), formula 8.6):
Confidence intervals and tests of the null hypothesis
3.10 Comparison between the implemented approaches
3.10.1 Treatment effect estimation
All approaches provide consistent treatment effect estimates for standard and reference-based imputation methods in case the analysis model of the completed datasets is a general linear model such as ANCOVA. Methods other than conditional mean imputation should also be valid for other analysis models. The validity of conditional mean imputation has only been formally demonstrated for analyses using the general linear model (Wolbers et al. (2022, sec. 2.4)) though it may also be applicable more widely (e.g. for general MMRM analysis models).
Treatment effects based on conditional mean imputation are deterministic. All other methods are affected by Monte Carlo sampling error and the precision of estimates depends on the number of imputations or bootstrap samples, respectively.
3.10.2 Standard errors of the treatment effect
All approaches for imputation under a MAR assumption provide consistent estimates of the frequentist standard error.
For reference-based imputation methods, the situation is more complicated and two different types of variance estimators have been proposed in the statistical literature (Bartlett (2023)). The first is the frequentist variance which describes the actual repeated sampling variability of the estimator. If the reference-based missing data assumption is correctly specified, then the resulting inference based on this variance is correct in the frequentist sense, i.e. hypothesis tests have asymptotically correct type I error control and confidence intervals have correct coverage probabilities under repeated sampling (Bartlett (2023), Wolbers et al. (2022)). Reference-based missing data assumptions are strong and borrow information from the reference arm for imputation in the active arm. As a consequence, the size of frequentist standard errors for treatment effects may decrease with increasing amounts of missing data. The second proposal is the so-called “information-anchored” variance which was originally proposed in the context of sensitivity analyses (Cro, Carpenter, and Kenward (2019)). This variance estimator is based on disentangling point estimation and variance estimation altogether. The information-anchoring principle described in Cro, Carpenter, and Kenward (2019) states that the relative increase in the variance of the treatment effect estimator under MAR imputation with increasing amounts of missing data should be preserved for reference-based imputation methods. The resulting information-anchored variance is typically very similar to the variance under MAR imputation and typically increases with increasing amounts of missing data. However, the information-anchored variance does not reflect the actual variability of the reference-based estimator under repeated sampling and the resulting inference is highly conservative resulting in a substantial power loss (Wolbers et al. (2022)). Moreover, to date, no Bayesian or frequentist framework has been developed under which the information-anchored variance provides correct inference for reference-based missingness assumptions, nor is it clear whether such a framework can even be developed.
Reference-based conditional mean imputation (method_condmean()
) and bootstrapped likelihood-based multiple methods (method = method_bmlmi()
) obtain standard errors via resampling and hence target the frequentist variance (Wolbers et al. (2022), von Hippel and Bartlett (2021)).
For finite samples, simulations for a sample size of method_condmean(type = "jackknife")
) provided exact protection of the type one error rate whereas the bootstrap (method_condmean(type = "bootstrap")
) was associated with a small type I error inflation (between 5.1% to 5.3% for a nominal level of 5%).
For reference-based conditional mean imputation, an alternative information-anchored variance can be obtained by following a proposal by Lu (2021).
The basic idea of Lu (2021) is to obtain the information-anchored variance via a MAR imputation combined with a delta-adjustment where delta is selected in a data-driven way to match the reference-based estimator.
For conditional mean imputation, the proposal by Lu (2021) can be implemented by choosing the delta-adjustment as the difference between the conditional mean imputation under the chosen reference-based assumption and MAR on the original dataset.
An illustration of how the different variances can be obtained for conditional mean imputation in rbmi
is provided in the vignette “Frequentist and information-anchored inference for reference-based conditional mean imputation” (vignette(topic = "CondMean_Inference", package = "rbmi")
).
Reference-based Bayesian (or approximate Bayesian) multiple imputation methods combined with Rubin’s rules (method_bayes()
and method_approxbayes()
) target the information-anchored variance (Cro, Carpenter, and Kenward (2019)).
A frequentist variance for these methods could in principle be obtained via bootstrap or jackknife re-sampling of the treatment effect estimates but this would be very computationally intensive and is not directly supported by rbmi
.
Our view is that for primary analyses, accurate type I error control (which can be obtained by using the frequentist variance) is more important than adherence to the information anchoring principle which, to us, is not fully compatible with the strong reference-based assumptions. In any case, if reference-based imputation is used for the primary analysis, it is critical that the chosen reference-based assumption can be clinically justified, and that suitable sensitivity analyses are conducted to stress-test these assumptions.
Conditional mean imputation combined with the jackknife is the only method which leads to deterministic standard error estimates and, consequently, confidence intervals and
3.10.3 Computational complexity
Bayesian MI methods rely on the specification of prior distributions and the usage of Markov chain Monte Carlo (MCMC) methods.
All other methods based on multiple imputation or bootstrapping require no other tuning parameters than the specification of the number of imputations
In our rbmi
implementation, the fitting of the MMRM imputation model via REML is computationally most expensive. MCMC sampling using rstan
(Stan Development Team (2020)) is typically relatively fast in our setting and requires only a small warmup phase and thinning of the chains. In addition, the number of random imputations for reliable inference using Rubin’s rules is often smaller than the number of resamples required for the jackknife or the bootstrap (see e.g. the discussions in I. R. White, Royston, and Wood (2011, sec. 7) for Bayesian MI and the Appendix of Wolbers et al. (2022) for the bootstrap). Thus, for many applications, we expect that conventional MI based on Bayesian posterior draws will be fastest, followed by conventional MI using approximate Bayesian posterior draws and conditional mean imputation combined with the jackknife. Conditional mean imputation combined with the bootstrap and bootstrapped MI methods will typically be most computationally demanding. Of note, all implemented methods are conceptually straightforward to parallelise and some parallelisation support is provided by rbmi
.
4 Mapping of statistical methods to rbmi
functions
For a full documentation of the rbmi
package functionality we refer to the help pages of all functions and to the other package vignettes. Here we only give a brief overview of how the different steps of the imputation procedure are mapped to rbmi
functions:
- The base imputation model fitting step is implemented in the function
draws()
. The chosen MI approach can be set using the argumentmethod
and should be one of the following:- Bayesian posterior parameter draws from the imputation model are obtained via the argument
method = method_bayes()
. - Approximate Bayesian posterior parameter draws from the imputation model are obtained via argument
method = method_approxbayes()
. - ML or REML parameter estimates of the imputation model parameters for the original dataset and all leave-one-subject-out datasets (as required for the jackknife) are obtained via argument
method = method_condmean(type = "jackknife")
. - ML or REML parameter estimates of the imputation model parameters for the original dataset and bootstrapped datasets are obtained via argument
method = method_condmean(type = "bootstrap")
. - Bootstrapped MI methods are obtained via argument
method = method_bmlmi(B=B, D=D)
where refers to the number of bootstrap samples and to the number of random imputations for each bootstrap sample.
- Bayesian posterior parameter draws from the imputation model are obtained via the argument
- The imputation step using random imputation or deterministic conditional mean imputation, respectively, is implemented in function
impute()
. Imputation can be performed assuming the already implemented imputation strategies as presented in section 3.4. Additionally, user-defined imputation strategies are also supported. - The analysis step is implemented in function
analyse()
and applies the analysis model to all imputed datasets. By default, the analysis model (argumentfun
) is theancova()
function but alternative analysis functions can also be provided by the user. Theanalyse()
function also allows for -adjustments to the imputed datasets prior to the analysis via argumentdelta
. - The inference step is implemented in function
pool()
which pools the results across imputed datasets. The Rubin and Bernard rule is applied in case of (approximate) Bayesian MI. For conditional mean imputation, jackknife and bootstrap (normal approximation or percentile) inference is supported. For BMLMI, the pooling and inference steps are performed viapool()
which in this case implements the method described in Section 3.9.
5 Comparison to other software implementations
An established software implementation of reference-based imputation in SAS are the so-called “five macros” by James Roger (Roger (2021)). An alternative R
implementation which is also currently under development is the R package RefBasedMI
(McGrath and White (2021)).
rbmi
has several features which are not supported by the other implementations:
In addition to the Bayesian MI approach implemented also in the other packages, our implementation provides three alternative MI approaches: approximate Bayesian MI, conditional mean imputation combined with resampling, and bootstrapped MI.
rbmi
allows for the usage of data collected after an ICE. For example, suppose that we want to adopt a treatment policy strategy for the ICE “treatment discontinuation”. A possible implementation of this strategy is to use the observed outcome data for subjects who remain in the study after the ICE and to use reference-based imputation in case the subject drops out. In our implementation, this is implemented by excluding observed post ICE data from the imputation model which assumes MAR missingness but including them in the analysis model. To our knowledge, this is not directly supported by the other implementations.RefBasedMI
fits the imputation model to data from each treatment group separately which implies covariate-treatment group interactions for all covariates for the pooled data from both treatment groups. In contrast, Roger’s five macros assume a joint model including data from all the randomized groups and covariate-treatment interactions covariates are not allowed. We also chose to implement a joint model but use a flexible model for the linear predictor which may or may not include an interaction term between any covariate and the treatment group. In addition, our imputation model also allows for the inclusion of time-varying covariates.In our implementation, the grouping of the subjects for the purpose of the imputation model (and the definition of the reference group) does not need to correspond to the assigned treatment groups. This provides additional flexibility for the imputation procedure. It is not clear to us whether this feature is supported by Roger’s five macros or
RefBasedMI
.We believe that our R-based implementation is more modular than
RefBasedMI
which should facilitate further package enhancements.
In contrast, the more general causal model introduced by I. White, Royes, and Best (2020) is available in the other implementations but is currently not supported by ours.