This function takes multiple imputed datasets (as generated by
the impute()
function) and runs an analysis function on
each of them.
Arguments
- imputations
An
imputations
object as created byimpute()
.- fun
An analysis function to be applied to each imputed dataset. See details.
- delta
A
data.frame
containing the delta transformation to be applied to the imputed datasets prior to runningfun
. See details.- ...
Additional arguments passed onto
fun
.- ncores
The number of parallel processes to use when running this function. Can also be a cluster object created by
make_rbmi_cluster()
. See the parallisation section below.- .validate
Should
inputations
be checked to ensure it conforms to the required format (default =TRUE
) ? Can gain a small performance increase if this is set toFALSE
when analysing a large number of samples.
Details
This function works by performing the following steps:
Extract a dataset from the
imputations
object.Apply any delta adjustments as specified by the
delta
argument.Run the analysis function
fun
on the dataset.Repeat steps 1-3 across all of the datasets inside the
imputations
object.Collect and return all of the analysis results.
The analysis function fun
must take a data.frame
as its first
argument. All other options to analyse()
are passed onto fun
via ...
.
fun
must return a named list with each element itself being a
list containing a single
numeric element called est
(or additionally se
and df
if
you had originally specified method_bayes()
or method_approxbayes()
)
i.e.:
myfun <- function(dat, ...) {
mod_1 <- lm(data = dat, outcome ~ group)
mod_2 <- lm(data = dat, outcome ~ group + covar)
x <- list(
trt_1 = list(
est = coef(mod_1)[[group]],
se = sqrt(vcov(mod_1)[group, group]),
df = df.residual(mod_1)
),
trt_2 = list(
est = coef(mod_2)[[group]],
se = sqrt(vcov(mod_2)[group, group]),
df = df.residual(mod_2)
)
)
return(x)
}
Please note that the vars$subjid
column (as defined in the original call to
draws()
) will be scrambled in the data.frames that are provided to fun
.
This is to say they will not contain the original subject values and as such
any hard coding of subject ids is strictly to be avoided.
By default fun
is the ancova()
function.
Please note that this function
requires that a vars
object, as created by set_vars()
, is provided via
the vars
argument e.g. analyse(imputeObj, vars = set_vars(...))
. Please
see the documentation for ancova()
for full details.
Please also note that the theoretical justification for the conditional mean imputation
method (method = method_condmean()
in draws()
) relies on the fact that ANCOVA is
a linear transformation of the outcomes.
Thus care is required when applying alternative analysis functions in this setting.
The delta
argument can be used to specify offsets to be applied
to the outcome variable in the imputed datasets prior to the analysis.
This is typically used for sensitivity or tipping point analyses. The
delta dataset must contain columns vars$subjid
, vars$visit
(as specified
in the original call to draws()
) and delta
. Essentially this data.frame
is merged onto the imputed dataset by vars$subjid
and vars$visit
and then
the outcome variable is modified by:
Please note that in order to provide maximum flexibility, the delta
argument
can be used to modify any/all outcome values including those that were not
imputed. Care must be taken when defining offsets. It is recommend that you
use the helper function delta_template()
to define the delta datasets as
this provides utility variables such as is_missing
which can be used to identify
exactly which visits have been imputed.
Parallelisation
To speed up the evaluation of analyse()
you can use the ncores
argument to enable parallelisation.
Simply providing an integer will get rbmi
to automatically spawn that many background processes
to parallelise across. If you are using a custom analysis function then you need to ensure
that any libraries or global objects required by your function are available in the
sub-processes. To do this you need to use the make_rbmi_cluster()
function for example:
my_custom_fun <- function(...) <some analysis code>
cl <- make_rbmi_cluster(
4,
objects = list("my_custom_fun" = my_custom_fun),
packages = c("dplyr", "nlme")
)
analyse(
imputations = imputeObj,
fun = my_custom_fun,
ncores = cl
)
parallel::stopCluster(cl)
Note that there is significant overhead both with setting up the sub-processes and with
transferring data back-and-forth between the main process and the sub-processes. As such
parallelisation of the analyse()
function tends to only be worth it when you have
> 2000
samples generated by draws()
. Conversely using parallelisation if your samples
are smaller than this may lead to longer run times than just running it sequentially.
It is important to note that the implementation of parallel processing within analyse()
has
been optimised around the assumption that the parallel processes will be spawned on the same
machine and not a remote cluster. One such optimisation is that the required data is saved to
a temporary file on the local disk from which it is then read into each sub-process. This is
done to avoid the overhead of transferring the data over the network. Our assumption is that
if you are at the stage where you need to be parallelising your analysis over a remote cluster
then you would likely be better off parallelising across multiple rbmi
runs rather than within
a single rbmi
run.
Finally, if you are doing a tipping point analysis you can get a reasonable performance
improvement by re-using the cluster between each call to analyse()
e.g.
See also
extract_imputed_dfs()
for manually extracting imputed
datasets.
delta_template()
for creating delta data.frames.
ancova()
for the default analysis function.
Examples
if (FALSE) { # \dontrun{
vars <- set_vars(
subjid = "subjid",
visit = "visit",
outcome = "outcome",
group = "group",
covariates = c("sex", "age", "sex*age")
)
analyse(
imputations = imputeObj,
vars = vars
)
deltadf <- data.frame(
subjid = c("Pt1", "Pt1", "Pt2"),
visit = c("Visit_1", "Visit_2", "Visit_2"),
delta = c( 5, 9, -10)
)
analyse(
imputations = imputeObj,
delta = deltadf,
vars = vars
)
} # }