16 results on '"Morris, Tim P."'
Search Results
2. Comments on 'standard and reference‐based conditional mean imputation': Regulators and trial statisticians be aware!
- Author
-
Cro, Suzie, Morris, Tim P., Roger, James H., and Carpenter, James R.
- Subjects
- *
MISSING data (Statistics) , *CLINICAL trials , *TREATMENT effectiveness , *STATISTICIANS - Abstract
Accurate frequentist performance of a method is desirable in confirmatory clinical trials, but is not sufficient on its own to justify the use of a missing data method. Reference‐based conditional mean imputation, with variance estimation justified solely by its frequentist performance, has the surprising and undesirable property that the estimated variance becomes smaller the greater the number of missing observations; as explained under jump‐to‐reference it effectively forces the true treatment effect to be exactly zero for patients with missing data. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. A four-step strategy for handling missing outcome data in randomised trials affected by a pandemic
- Author
-
Cro, Suzie, Morris, Tim P., Kahan, Brennan C., Cornelius, Victoria R., and Carpenter, James R.
- Published
- 2020
- Full Text
- View/download PDF
4. How are missing data in covariates handled in observational time-to-event studies in oncology? A systematic review
- Author
-
Carroll, Orlagh U., Morris, Tim P., and Keogh, Ruth H.
- Published
- 2020
- Full Text
- View/download PDF
5. Population-calibrated multiple imputation for a binary/categorical covariate in categorical regression models
- Author
-
Pham, Tra My, Carpenter, James R, Morris, Tim P, Wood, Angela M, Petersen, Irene, Pham, Tra My [0000-0003-0528-6303], Morris, Tim P [0000-0001-5850-3610], Petersen, Irene [0000-0002-0037-7524], and Apollo - University of Cambridge Repository
- Subjects
FOS: Computer and information sciences ,Models, Statistical ,multiple imputation ,missing not at random ,Methodology (stat.ME) ,missing data ,electronic health records ,Logistic Models ,sensitivity analysis ,Diabetes Mellitus, Type 2 ,Research Design ,Data Interpretation, Statistical ,Ethnicity ,Prevalence ,Humans ,Research Articles ,Statistics - Methodology ,Research Article - Abstract
Multiple imputation (MI) has become popular for analyses with missing data in medical research. The standard implementation of MI is based on the assumption of data being missing at random (MAR). However, for missing data generated by missing not at random (MNAR) mechanisms, MI performed assuming MAR might not be satisfactory. For an incomplete variable in a given dataset, its corresponding population marginal distribution might also be available in an external data source. We show how this information can be readily utilised in the imputation model to calibrate inference to the population, by incorporating an appropriately calculated offset termed the `calibrated-$\delta$ adjustment'. We describe the derivation of this offset from the population distribution of the incomplete variable and show how in applications it can be used to closely (and often exactly) match the post-imputation distribution to the population level. Through analytic and simulation studies, we show that our proposed calibrated-$\delta$ adjustment MI method can give the same inference as standard MI when data are MAR, and can produce more accurate inference under two general MNAR missingness mechanisms. The method is used to impute missing ethnicity data in a type 2 diabetes prevalence case study using UK primary care electronic health records, where it results in scientifically relevant changes in inference for non-White ethnic groups compared to standard MI. Calibrated-$\delta$ adjustment MI represents a pragmatic approach for utilising available population-level information in a sensitivity analysis to explore potential departure from the MAR assumption.
- Published
- 2019
6. Handling Missing Values in Interrupted Time Series Analysis of Longitudinal Individual-Level Data
- Author
-
Bazo-Alvarez, Juan Carlos, Morris, Tim P, Pham, Tra My, Carpenter, James R, and Petersen, Irene
- Subjects
Mixed effects models ,multiple imputation ,Interrupted time series analysis ,Missing data ,Big data ,missing data ,Segmented regression ,electronic health records ,big data ,Multiple imputation ,Electronic health records ,mixed effects models ,Clinical Epidemiology ,interrupted time series analysis ,segmented regression ,Original Research - Abstract
Juan Carlos Bazo-Alvarez,1,2 Tim P Morris,3 Tra My Pham,3 James R Carpenter,3,4 Irene Petersen1,5 1Research Department of Primary Care and Population Health, University College London (UCL), London, UK; 2Instituto de Investigación, Universidad Católica Los Ángeles de Chimbote, Chimbote, Peru; 3MRC Clinical Trials Unit at UCL, London, UK; 4Department of Medical Statistics, London School of Hygiene and Tropical Medicine, London, UK; 5Department of Clinical Epidemiology, Aarhus University, Aarhus, DenmarkCorrespondence: Juan Carlos Bazo-AlvarezResearch Department of Primary Care and Population Health, University College London (UCL), Rowland Hill Street, London NW3 2PF, UKTel +44 7376076260Email juan.alvarez.16@ucl.ac.ukBackground: In the interrupted time series (ITS) approach, it is common to average the outcome of interest at each time point and then perform a segmented regression (SR) analysis. In this study, we illustrate that such ‘aggregate-level’ analysis is biased when data are missing at random (MAR) and provide alternative analysis methods.Methods: Using electronic health records from the UK, we evaluated weight change over time induced by the initiation of antipsychotic treatment. We contrasted estimates from aggregate-level SR analysis against estimates from mixed models with and without multiple imputation of missing covariates, using individual-level data. Then, we conducted a simulation study for insight about the different results in a controlled environment.Results: Aggregate-level SR analysis suggested a substantial weight gain after initiation of treatment (average short-term weight change: 0.799kg/week) compared to mixed models (0.412kg/week). Simulation studies confirmed that aggregate-level SR analysis was biased when data were MAR. In simulations, mixed models gave less biased estimates than SR analysis and, in combination with multilevel multiple imputation, provided unbiased estimates. Mixed models with multiple imputation can be used with other types of ITS outcomes (eg, proportions). Other standard methods applied in ITS do not help to correct this bias problem.Conclusion: Aggregate-level SR analysis can bias the ITS estimates when individual-level data are MAR, because taking averages of individual-level data before SR means that data at the cluster level are missing not at random. Avoiding the averaging-step and using mixed models with or without multilevel multiple imputation of covariates is recommended.Keywords: interrupted time series analysis, segmented regression, missing data, multiple imputation, mixed effects models, electronic health records, big data
- Published
- 2020
- Full Text
- View/download PDF
7. A comparison of methods for analyzing a binary composite endpoint with partially observed components in randomized controlled trials.
- Author
-
Pham, Tra My, White, Ian R., Kahan, Brennan C., Morris, Tim P., Stanworth, Simon J., and Forbes, Gordon
- Subjects
RANDOMIZED controlled trials ,MISSING data (Statistics) ,DIRECTLY observed therapy - Abstract
Composite endpoints are commonly used to define primary outcomes in randomized controlled trials. A participant may be classified as meeting the endpoint if they experience an event in one or several components (eg, a favorable outcome based on a composite of being alive and attaining negative culture results in trials assessing tuberculosis treatments). Partially observed components that are not missing simultaneously complicate the analysis of the composite endpoint. An intuitive strategy frequently used in practice for handling missing values in the components is to derive the values of the composite endpoint from observed components when possible, and exclude from analysis participants whose composite endpoint cannot be derived. Alternatively, complete record analysis (CRA) (excluding participants with any missing components) or multiple imputation (MI) can be used. We compare a set of methods for analyzing a composite endpoint with partially observed components mathematically and by simulation, and apply these methods in a reanalysis of a published trial (TOPPS). We show that the derived composite endpoint can be missing not at random even when the components are missing completely at random. Consequently, the treatment effect estimated from the derived endpoint is biased while CRA results without the derived endpoint are valid. Missing at random mechanisms require MI of the components. We conclude that, although superficially attractive, deriving the composite endpoint from observed components should generally be avoided. Despite the potential risk of imputation model mis‐specification, MI of missing components is the preferred approach in this study setting. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
8. Combining fractional polynomial model building with multiple imputation
- Author
-
Morris, Tim P, White, Ian R, Carpenter, James R, Stanworth, Simon J, and Royston, Patrick
- Subjects
Likelihood Functions ,Models, Statistical ,multivariable fractional polynomials ,multiple imputation ,fractional polynomials ,Prognosis ,missing data ,Multivariate Analysis ,Linear Models ,Statistics::Methodology ,Humans ,Regression Analysis ,Computer Simulation ,Registries ,Research Articles ,Research Article - Abstract
Multivariable fractional polynomial (MFP) models are commonly used in medical research. The datasets in which MFP models are applied often contain covariates with missing values. To handle the missing values, we describe methods for combining multiple imputation with MFP modelling, considering in turn three issues: first, how to impute so that the imputation model does not favour certain fractional polynomial (FP) models over others; second, how to estimate the FP exponents in multiply imputed data; and third, how to choose between models of differing complexity. Two imputation methods are outlined for different settings. For model selection, methods based on Wald‐type statistics and weighted likelihood‐ratio tests are proposed and evaluated in simulation studies. The Wald‐based method is very slightly better at estimating FP exponents. Type I error rates are very similar for both methods, although slightly less well controlled than analysis of complete records; however, there is potential for substantial gains in power over the analysis of complete records. We illustrate the two methods in a dataset from five trauma registries for which a prognostic model has previously been published, contrasting the selected models with that obtained by analysing the complete records only. © 2015 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd.
- Published
- 2015
9. Current Practices in Missing Data Handling for Interrupted Time Series Studies Performed on Individual-Level Data: A Scoping Review in Health Research.
- Author
-
Bazo-Alvarez, Juan Carlos, Morris, Tim P, Carpenter, James R, and Petersen, Irene
- Subjects
MISSING data (Statistics) ,PUBLIC health research ,TIME series analysis ,STATISTICAL models - Abstract
abstracts indexed on MEDLINE. Results: From 732 studies identified, we finally reviewed 60. Reporting of missing data was rare. Data aggregation, statistical tools for modelling population-level data and complete case analyses were preferred, but these can lead to bias when data are missing at random. Seasonality and other time-dependent confounders were rarely accounted for and, when they were, missing data implications were typically ignored. Very few studies reflected on the consequences of missing data. Conclusion: Handling and reporting of missing data in recent ITS studies performed for health research have many shortcomings compared with best practice. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
10. Sensitivity analysis for clinical trials with missing continuous outcome data using controlled multiple imputation: A practical guide.
- Author
-
Cro, Suzie, Morris, Tim P., Kenward, Michael G., and Carpenter, James R.
- Subjects
- *
SENSITIVITY analysis , *CLINICAL trials , *EXPECTED returns , *REFERENCE values , *DATA analysis , *STATISTICS , *RESEARCH , *RESEARCH methodology , *MEDICAL cooperation , *EVALUATION research , *COMPARATIVE studies , *RESEARCH funding - Abstract
Missing data due to loss to follow-up or intercurrent events are unintended, but unfortunately inevitable in clinical trials. Since the true values of missing data are never known, it is necessary to assess the impact of untestable and unavoidable assumptions about any unobserved data in sensitivity analysis. This tutorial provides an overview of controlled multiple imputation (MI) techniques and a practical guide to their use for sensitivity analysis of trials with missing continuous outcome data. These include δ- and reference-based MI procedures. In δ-based imputation, an offset term, δ, is typically added to the expected value of the missing data to assess the impact of unobserved participants having a worse or better response than those observed. Reference-based imputation draws imputed values with some reference to observed data in other groups of the trial, typically in other treatment arms. We illustrate the accessibility of these methods using data from a pediatric eczema trial and a chronic headache trial and provide Stata code to facilitate adoption. We discuss issues surrounding the choice of δ in δ-based sensitivity analysis. We also review the debate on variance estimation within reference-based analysis and justify the use of Rubin's variance estimator in this setting, since as we further elaborate on within, it provides information anchored inference. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
11. Multiple imputation in Cox regression when there are time-varying effects of covariates.
- Author
-
Keogh, Ruth H. and Morris, Tim P.
- Abstract
In Cox regression, it is important to test the proportional hazards assumption and sometimes of interest in itself to study time-varying effects (TVEs) of covariates. TVEs can be investigated with log hazard ratios modelled as a function of time. Missing data on covariates are common and multiple imputation is a popular approach to handling this to avoid the potential bias and efficiency loss resulting from a "complete-case" analysis. Two multiple imputation methods have been proposed for when the substantive model is a Cox proportional hazards regression: an approximate method (Imputing missing covariate values for the Cox model in Statistics in Medicine (2009) by White and Royston) and a substantive-model-compatible method (Multiple imputation of covariates by fully conditional specification: accommodating the substantive model in Statistical Methods in Medical Research (2015) by Bartlett et al). At present, neither accommodates TVEs of covariates. We extend them to do so for a general form for the TVEs and give specific details for TVEs modelled using restricted cubic splines. Simulation studies assess the performance of the methods under several underlying shapes for TVEs. Our proposed methods give approximately unbiased TVE estimates for binary covariates with missing data, but for continuous covariates, the substantive-model-compatible method performs better. The methods also give approximately correct type I errors in the test for proportional hazards when there is no TVE and gain power to detect TVEs relative to complete-case analysis. Ignoring TVEs at the imputation stage results in biased TVE estimates, incorrect type I errors, and substantial loss of power in detecting TVEs. We also propose a multivariable TVE model selection algorithm. The methods are illustrated using data from the Rotterdam Breast Cancer Study. R code is provided. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
12. Reference-based sensitivity analysis via multiple imputation for longitudinal trials with protocol deviation.
- Author
-
Cro, Suzie, Morris, Tim P., Kenward, Michael G., and Carpenter, James R.
- Subjects
- *
RANDOMIZATION (Statistics) , *STATISTICS , *STANDARD deviations - Abstract
Randomized controlled trials provide essential evidence for the evaluation of new and existing medical treatments. Unfortunately, the statistical analysis is often complicated by the occurrence of protocol deviations, which mean we cannot always measure the intended outcomes for individuals who deviate, resulting in a missing-data problem. In such settings, however one approaches the analysis, an untestable assumption about the distribution of the unobserved data must be made. To understand how far the results depend on these assumptions, the primary analysis should be supplemented by a range of sensitivity analyses, which explore how the conclusions vary over a range of different credible assumptions for the missing data. In this article, we describe a new command, mimix, that can be used to perform reference-based sensitivity analyses for randomized controlled trials with longitudinal quantitative outcome data, using the approach proposed by Carpenter, Roger, and Kenward (2013, Journal of Biopharmaceutical Statistics 23: 1352-1371). Under this approach, we make qualitative assumptions about how individuals' missing outcomes relate to those observed in relevant groups in the trial, based on plausible clinical scenarios. Statistical analysis then proceeds using the method of multiple imputation. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
13. Tuning multiple imputation by predictive mean matching and local residual draws.
- Author
-
Morris, Tim P., White, Ian R., and Royston, Patrick
- Subjects
- *
MULTIPLE imputation (Statistics) , *MISSING data (Statistics) , *RESIDUAL charges , *SOCIAL statistics , *DATA security - Abstract
Background Multiple imputation is a commonly used method for handling incomplete covariates as it can provide valid inference when data are missing at random. This depends on being able to correctly specify the parametric model used to impute missing values, which may be difficult in many realistic settings. Imputation by predictive mean matching (PMM) borrows an observed value from a donor with a similar predictive mean; imputation by local residual draws (LRD) instead borrows the donor's residual. Both methods relax some assumptions of parametric imputation, promising greater robustness when the imputation model is misspecified. Methods We review development of PMM and LRD and outline the various forms available, and aim to clarify some choices about how and when they should be used. We compare performance to fully parametric imputation in simulation studies, first when the imputation model is correctly specified and then when it is misspecified. Results In using PMM or LRD we strongly caution against using a single donor, the default value in some implementations, and instead advocate sampling from a pool of around 10 donors. We also clarify which matching metric is best. Among the current MI software there are several poor implementations. Conclusions PMM and LRD may have a role for imputing covariates (i) which are not strongly associated with outcome, and (ii) when the imputation model is thought to be slightly but not grossly misspecified. Researchers should spend efforts on specifying the imputation model correctly, rather than expecting predictive mean matching or local residual draws to do the work. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
14. Choosing sensitivity analyses for randomised trials: principles.
- Author
-
Morris, Tim P., Kahan, Brennan C., and White, Ian R.
- Subjects
- *
SENSITIVITY analysis , *CLINICAL trials , *MISSING data (Statistics) , *MEDICAL research ,MATHEMATICAL models of uncertainty - Abstract
Background Sensitivity analyses are an important tool for understanding the extent to which the results of randomised trials depend upon the assumptions of the analysis. There is currently no guidance governing the choice of sensitivity analyses. Discussion We provide a principled approach to choosing sensitivity analyses through the consideration of the following questions: 1) Does the proposed sensitivity analysis address the same question as the primary analysis? 2) Is it possible for the proposed sensitivity analysis to return a different result to the primary analysis? 3) If the results do differ, is there any uncertainty as to which will be believed? Answering all of these questions in the affirmative will help researchers to identify relevant sensitivity analyses. Treating analyses as sensitivity analyses when one or more of the answers are negative can be misleading and confuse the interpretation of studies. The value of these questions is illustrated with several examples. Summary By removing unreasonable analyses that might have been performed, these questions will lead to relevant sensitivity analyses, which help to assess the robustness of trial results. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
15. Multiple imputation for an incomplete covariate that is a ratio.
- Author
-
Morris, Tim P., White, Ian R., Royston, Patrick, Seaman, Shaun R., and Wood, Angela M.
- Abstract
We are concerned with multiple imputation of the ratio of two variables, which is to be used as a covariate in a regression analysis. If the numerator and denominator are not missing simultaneously, it seems sensible to make use of the observed variable in the imputation model. One such strategy is to impute missing values for the numerator and denominator, or the log-transformed numerator and denominator, and then calculate the ratio of interest; we call this 'passive' imputation. Alternatively, missing ratio values might be imputed directly, with or without the numerator and/or the denominator in the imputation model; we call this 'active' imputation. In two motivating datasets, one involving body mass index as a covariate and the other involving the ratio of total to high-density lipoprotein cholesterol, we assess the sensitivity of results to the choice of imputation model and, as an alternative, explore fully Bayesian joint models for the outcome and incomplete ratio. Fully Bayesian approaches using Win bugs were unusable in both datasets because of computational problems. In our first dataset, multiple imputation results are similar regardless of the imputation model; in the second, results are sensitive to the choice of imputation model. Sensitivity depends strongly on the coefficient of variation of the ratio's denominator. A simulation study demonstrates that passive imputation without transformation is risky because it can lead to downward bias when the coefficient of variation of the ratio's denominator is larger than about 0.1. Active imputation or passive imputation after log-transformation is preferable. © 2013 The Authors. Statistics in Medicine published by John Wiley & Sons, Ltd. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
16. Multiple imputation for an incomplete covariate that is a ratio
- Author
-
Morris, Tim P, White, Ian R, Royston, Patrick, Seaman, Shaun R, and Wood, Angela M
- Subjects
Male ,Models, Statistical ,Statistics::Applications ,multiple imputation ,Bayes Theorem ,HIV Infections ,compatibility ,Quantitative Biology::Genomics ,3. Good health ,Body Mass Index ,CD4 Lymphocyte Count ,Cohort Studies ,missing data ,Hemoglobins ,South Africa ,Cholesterol ,Neoplasms ,Statistics::Methodology ,Humans ,Regression Analysis ,Computer Simulation ,Female ,ratios - Abstract
We are concerned with multiple imputation of the ratio of two variables, which is to be used as a covariate in a regression analysis. If the numerator and denominator are not missing simultaneously, it seems sensible to make use of the observed variable in the imputation model. One such strategy is to impute missing values for the numerator and denominator, or the log-transformed numerator and denominator, and then calculate the ratio of interest; we call this 'passive' imputation. Alternatively, missing ratio values might be imputed directly, with or without the numerator and/or the denominator in the imputation model; we call this 'active' imputation. In two motivating datasets, one involving body mass index as a covariate and the other involving the ratio of total to high-density lipoprotein cholesterol, we assess the sensitivity of results to the choice of imputation model and, as an alternative, explore fully Bayesian joint models for the outcome and incomplete ratio. Fully Bayesian approaches using Winbugs were unusable in both datasets because of computational problems. In our first dataset, multiple imputation results are similar regardless of the imputation model; in the second, results are sensitive to the choice of imputation model. Sensitivity depends strongly on the coefficient of variation of the ratio's denominator. A simulation study demonstrates that passive imputation without transformation is risky because it can lead to downward bias when the coefficient of variation of the ratio's denominator is larger than about 0.1. Active imputation or passive imputation after log-transformation is preferable.
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.