906 results
Search Results
2. Discussion of paper by C. B. Begg
- Author
-
Richard M. Royall
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Mathematics education ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1990
3. Discussion of paper by C. B. Begg
- Author
-
Oscar Kempthorne
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Mathematics education ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1990
4. Discussion of paper by C. B. Begg
- Author
-
Thomas R. Fleming
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Mathematics education ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1990
5. Discussion of paper by C. B. Begg
- Author
-
D. R. Cox
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Mathematics education ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1990
6. Discussion of paper by C. B. Begg
- Author
-
L. J. Wei
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Mathematics education ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1990
7. Discussion of paper by C. B. Begg
- Author
-
D. A. Sprott and Vernon T. Farewell
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Mathematics education ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1990
8. Comments on paper by J. D. Kalbfleisch: Some personal comments on sufficiency and conditionality
- Author
-
A. D. McLAREN
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Conditionality ,Statistics, Probability and Uncertainty ,Positive economics ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1975
9. Comments on a paper by I. Olkin and M. Vaeth on two-way analysis of variance with correlated errors
- Author
-
D. E. Walters and J. G. Rowell
- Subjects
Statistics and Probability ,Wishart distribution ,Covariance matrix ,Applied Mathematics ,General Mathematics ,Two-way analysis of variance ,Multivariate normal distribution ,Agricultural and Biological Sciences (miscellaneous) ,One-way analysis of variance ,Time factor ,Statistics ,Data analysis ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Row ,Mathematics - Abstract
Olkin & Vaeth (1981) consider a two-way classification model with k rows and p columns in which the residuals (ei1, ..., eip) for row i are independently distributed with a multivariate normal distribution with zero means and common covariance matrix ((ij)). They concentrate their attention particularly on the situation where the row classification corresponds to k subjects, each treated in one of two or more ways; the column classification corresponds to p repetitions, perhaps in time, from each subject, probably resulting in correlated errors. Replication of subjects within treatments enables the elements of the covariance matrix to be estimated from the data, which in turn enables maximum likelihood estimates of the row and column parameters to be computed and likelihood ratio tests to be carried out without having to make assumptions about the form of the covariance matrix. The careful analysis of data of this kind has been the subject of numerous papers, Wishart (1938) perhaps being the first to concentrate on this topic. Despite this, however, the use of erroneous methods in published material is widespread and this provided the stimulus for the publication of our earlier paper (Rowell & Walters, 1976), and also for the present communication. We note here that there is confusion in Olkin & Vaeth's row/column terminology in their worked example: the rows in their tables are referred to as columns in the text, and vice versa. In what follows, when referring to their numerical example, the rows classification will refer to the high/low factor, and the columns classification to the time factor.
- Published
- 1982
10. Comments on paper by J. D. Kalbfleisch
- Author
-
Allan Birnbaum
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematical economics ,Mathematics - Published
- 1975
11. Comments on paper by B. Efron and D. V. Hinkley
- Author
-
A. T. James
- Subjects
Statistics and Probability ,Discrete mathematics ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1978
12. Comments on paper by B. Efron and D. V. Hinkley
- Author
-
D. A. Sprott
- Subjects
Statistics and Probability ,Discrete mathematics ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1978
13. Comments on paper by P. D. Finch
- Author
-
Oscar Kempthorne
- Subjects
Statistics and Probability ,biology ,Applied Mathematics ,General Mathematics ,biology.animal ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Humanities ,Finch ,Mathematics - Published
- 1979
14. Comments on paper by M. Hollander and J. Sethuraman
- Author
-
William R. Schucany
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Humanities ,Mathematics - Published
- 1978
15. Comments on paper by J. D. Kalbfleisch
- Author
-
Ole E. Barndorff-Nielsen
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematical economics ,Mathematics - Published
- 1975
16. Comments on paper by J. D. Kalbfleisch
- Author
-
G. A. Barnard
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematical economics ,Mathematics - Published
- 1975
17. Comments on a paper by R. C. Geary on standardized mean deviation
- Author
-
K. O. Bowman, H. K. Lam, and L.R. Shenton
- Subjects
Statistics and Probability ,Absolute deviation ,Deviation ,Applied Mathematics ,General Mathematics ,Mean square weighted deviation ,Statistics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1979
18. Comments on paper by B. Efron and D. V. Hinkley
- Author
-
G. K. Robinson
- Subjects
Statistics and Probability ,Discrete mathematics ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Published
- 1978
19. Comments on paper by P. D. Finch
- Author
-
M. J. R. Healy
- Subjects
Statistics and Probability ,biology ,Applied Mathematics ,General Mathematics ,biology.animal ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Humanities ,Finch ,Mathematics - Published
- 1979
20. Causal inference with misspecified exposure mappings: separating definitions and assumptions
- Author
-
Fredrik Sävje
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Applied Mathematics ,General Mathematics ,Econometrics (econ.EM) ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Agricultural and Biological Sciences (miscellaneous) ,Methodology (stat.ME) ,FOS: Economics and business ,FOS: Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Statistics - Methodology ,Economics - Econometrics - Abstract
Summary Exposure mappings facilitate investigations of complex causal effects when units interact in experiments. Current methods require experimenters to use the same exposure mappings to define the effect of interest and to impose assumptions on the interference structure. However, the two roles rarely coincide in practice, and experimenters are forced to make the often questionable assumption that their exposures are correctly specified. This paper argues that the two roles exposure mappings currently serve can, and typically should, be separated, so that exposures are used to define effects without necessarily assuming that they are capturing the complete causal structure in the experiment. The paper shows that this approach is practically viable by providing conditions under which exposure effects can be precisely estimated when the exposures are misspecified. Some important questions remain open.
- Published
- 2023
21. Minimal dispersion approximately balancing weights: asymptotic properties and practical considerations
- Author
-
José R. Zubizarreta and Yixin Wang
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Mean squared error ,General Mathematics ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Statistics - Applications ,01 natural sciences ,Methodology (stat.ME) ,010104 statistics & probability ,Covariate ,FOS: Mathematics ,050602 political science & public administration ,Applied mathematics ,Applications (stat.AP) ,Statistical dispersion ,0101 mathematics ,Statistics - Methodology ,Mathematics ,Smoothness (probability theory) ,Applied Mathematics ,05 social sciences ,Estimator ,Function (mathematics) ,Agricultural and Biological Sciences (miscellaneous) ,0506 political science ,Weighting ,Inverse probability ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences - Abstract
Weighting methods are widely used to adjust for covariates in observational studies, sample surveys, and regression settings. In this paper, we study a class of recently proposed weighting methods which find the weights of minimum dispersion that approximately balance the covariates. We call these weights "minimal weights" and study them under a common optimization framework. The key observation is the connection between approximate covariate balance and shrinkage estimation of the propensity score. This connection leads to both theoretical and practical developments. From a theoretical standpoint, we characterize the asymptotic properties of minimal weights and show that, under standard smoothness conditions on the propensity score function, minimal weights are consistent estimates of the true inverse probability weights. Also, we show that the resulting weighting estimator is consistent, asymptotically normal, and semiparametrically efficient. From a practical standpoint, we present a finite sample oracle inequality that bounds the loss incurred by balancing more functions of the covariates than strictly needed. This inequality shows that minimal weights implicitly bound the number of active covariate balance constraints. We finally provide a tuning algorithm for choosing the degree of approximate balance in minimal weights. We conclude the paper with four empirical studies that suggest approximate balance is preferable to exact balance, especially when there is limited overlap in covariate distributions. In these studies, we show that the root mean squared error of the weighting estimator can be reduced by as much as a half with approximate balance., 41 pages
- Published
- 2019
22. Sparse envelope model: efficient estimation and response variable selection in multivariate linear regression
- Author
-
Zhihua Su, Yi Yang, Guangyu Zhu, and Xin Chen
- Subjects
0301 basic medicine ,Statistics and Probability ,Applied Mathematics ,General Mathematics ,Asymptotic distribution ,Estimator ,Feature selection ,01 natural sciences ,Agricultural and Biological Sciences (miscellaneous) ,Oracle ,010104 statistics & probability ,03 medical and health sciences ,030104 developmental biology ,Bayesian multivariate linear regression ,Linear predictor function ,Linear regression ,Statistics ,Statistics::Methodology ,Applied mathematics ,0101 mathematics ,Statistics, Probability and Uncertainty ,Invariant (mathematics) ,General Agricultural and Biological Sciences ,Mathematics - Abstract
The envelope model allows efficient estimation in multivariate linear regression. In this paper, we propose the sparse envelope model, which is motivated by applications where some response variables are invariant with respect to changes of the predictors and have zero regression coefficients. The envelope estimator is consistent but not sparse, and in many situations it is important to identify the response variables for which the regression coefficients are zero. The sparse envelope model performs variable selection on the responses and preserves the efficiency gains offered by the envelope model. Response variable selection arises naturally in many applications, but has not been studied as thoroughly as predictor variable selection. In this paper, we discuss response variable selection in both the standard multivariate linear regression and the envelope contexts. In response variable selection, even if a response has zero coefficients, it should still be retained to improve the estimation efficiency of the nonzero coefficients. This is different from the practice in predictor variable selection. We establish consistency and the oracle property and obtain the asymptotic distribution of the sparse envelope estimator.
- Published
- 2016
23. Designing dose-finding studies with an active control for exponential families
- Author
-
Frank Bretz, Holger Dette, and Katrin Kettelhake
- Subjects
Optimal design ,Statistics and Probability ,Mathematical optimization ,Applied Mathematics ,General Mathematics ,Design of experiments ,Dose-finding ,Regression analysis ,Articles ,Variance (accounting) ,Active control ,computer.software_genre ,Agricultural and Biological Sciences (miscellaneous) ,Dose finding ,Dose-response study ,Exponential family ,Data mining ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,computer ,Count data ,Mathematics - Abstract
In a recent paper Dette et al. (2014) introduced optimal design problems for dose finding studies with an active control. These authors concentrated on regression models with normal distributed errors (with known variance) and the problem of determining optimal designs for estimating the smallest dose, which achieves the same treatment effect as the active control. This paper discusses the problem of designing active-controlled dose finding studies from a broader perspective. In particular, we consider a general class of optimality criteria and models arising from an exponential family, which are frequently used analyzing count data. We investigate under which circumstances optimal designs for dose finding studies including a placebo can be used to obtain optimal designs for studies with an active control. Optimal designs are constructed for several situations and the differences arising from different distributional assumptions are investigated in detail. In particular, our results are applicable for constructing optimal experimental designs to analyze active-controlled dose finding studies with discrete data, and we illustrate the efficiency of the new optimal designs with two recent examples from our consulting projects.
- Published
- 2015
24. Statistical inference methods for recurrent event processes with shape and size parameters
- Author
-
Mei Cheng Wang and Chiung Yu Huang
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Contrast (statistics) ,Agricultural and Biological Sciences (miscellaneous) ,Censoring (statistics) ,Article ,Point process ,Statistics ,Statistical inference ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Rate function ,Random variable ,Statistical hypothesis testing ,Event (probability theory) ,Mathematics - Abstract
This paper proposes a unified framework to characterize the rate function of a recurrent event process through shape and size parameters. In contrast to the intensity function, which is the event occurrence rate conditional on the event history, the rate function is the occurrence rate unconditional on the event history, and thus it can be interpreted as a population-averaged count of events in unit time. In this paper, shape and size parameters are introduced and used to characterize the association between the rate function λ(⋅) and a random variable X. Measures of association between X and λ(⋅) are defined via shape- and size-based coefficients. Rate-independence of X and λ(⋅) is studied through tests of shape-independence and size-independence, where the shape- and size-based test statistics can be used separately or in combination. These tests can be applied when X is a covariable possibly correlated with the recurrent event process through λ(⋅) or, in the one-sample setting, when X is the censoring time at which the observation of N(⋅) is terminated. The proposed tests are shape- and size-based, so when a null hypothesis is rejected, the test results can serve to distinguish the source of violation.
- Published
- 2014
25. Characterization of the likelihood continual reassessment method
- Author
-
Xiaoyu Jia, Shing M. Lee, and Ying Kuen Cheung
- Subjects
Statistics and Probability ,Mathematical optimization ,Process (engineering) ,Calibration (statistics) ,Applied Mathematics ,General Mathematics ,Coherence (statistics) ,Initial sequence ,Agricultural and Biological Sciences (miscellaneous) ,Characterization (materials science) ,Continual reassessment method ,Econometrics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Mathematics - Abstract
This paper deals with the design of the likelihood continual reassessment method, which is an increasingly widely used model-based method for dose-finding studies. It is common to implement the method in a two-stage approach, whereby the model-based stage is activated after an initial sequence of patients has been treated. While this two-stage approach is practically appealing, it lacks a theoretical framework, and it is often unclear how the design components should be specified. This paper develops a general framework based on the coherence principle, from which we derive a design calibration process. A real clinical-trial example is used to demonstrate that the proposed process can be implemented in a timely and reproducible manner, while offering competitive operating characteristics. We explore the operating characteristics of different models within this framework and show the performance to be insensitive to the choice of dose-toxicity model.
- Published
- 2014
26. Evaluating causes of effects by posterior effects of causes
- Author
-
Zitong Lu, Zhi Geng, Wei Li, Shengyu Zhu, and Jinzhu Jia
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
Summary For the case with a single causal variable, Dawid et al. (2014) defined the probability of causation, and Pearl (2000) defined the probability of necessity to assess the causes of effects. For a case with multiple causes that could affect each other, this paper defines the posterior total and direct causal effects based on the evidence observed for post-treatment variables, which could be viewed as measurements of causes of effects. Posterior causal effects involve the probabilities of counterfactual variables. Thus, as with the probability of causation, the probability of necessity and direct causal effects, the identifiability of posterior total and direct causal effects requires more assumptions than the identifiability of traditional causal effects conditional on pre-treatment variables. We present assumptions required for the identifiability of posterior causal effects and provide identification equations. Further, when the causal relationships between multiple causes and an endpoint can be depicted by causal networks, we can simplify both the required assumptions and the identification equations of the posterior total and direct causal effects. Finally, using numerical examples, we compare the posterior total and direct causal effects with other measures for evaluating the causes of effects and the population attributable risks.
- Published
- 2022
27. Saddlepoint approximations for the normalizing constant of Fisher-Bingham distributions on products of spheres and Stiefel manifolds
- Author
-
Simon Preston, Andrew T. A. Wood, and Alfred Kume
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Normalizing constant ,Univariate ,Bingham distribution ,Cartesian product ,Agricultural and Biological Sciences (miscellaneous) ,Statistics::Computation ,Combinatorics ,Matrix (mathematics) ,symbols.namesake ,Distribution (mathematics) ,Joint probability distribution ,symbols ,Kent distribution ,Statistics::Methodology ,Applied mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Mathematics - Abstract
In an earlier paper Kume & Wood (2005) showed how the normalizing constant of the Fisher– Bingham distribution on a sphere can be approximated with high accuracy using a univariate saddlepoint density approximation. In this sequel, we extend the approach to a more general setting and derive saddlepoint approximations for the normalizing constants of multicomponent Fisher– Bingham distributions on Cartesian products of spheres, and Fisher–Bingham distributions on Stiefel manifolds. In each case, the approximation for the normalizing constant is essentially a multivariate saddlepoint density approximation for the joint distribution of a set of quadratic forms in normal variables. Both first-order and second-order saddlepoint approximations are considered. Computational algorithms, numerical results and theoretical properties of the approximations are presented. In the challenging high-dimensional settings considered in this paper the saddlepoint approximations perform very well in all examples considered. Some key words: Directional data; Fisher matrix distribution; Kent distribution; Orientation statistics.
- Published
- 2013
28. One-step TMLE for targeting cause-specific absolute risks and survival curves
- Author
-
H C W Rytgaard and M J Van Der Laan
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
This paper considers one-step targeted maximum likelihood estimation methodology for multi-dimensional causal parameters in general survival and competing risks settings where event times take place on the positive real line ℝ+ and are subject to right-censoring. We focus on effects of baseline treatment decisions possibly confounded by pre-treatment covariates, but remark that our work generalizes to settings with time-varying treatment regimes and time-dependent confounding. We point out two overall contributions of our work. First, our methods can be used to obtain simultaneous inference for treatment effects on multiple absolute risks in competing risks settings. Second, our methods can be used to achieve inference for the full survival curve, or a full absolute risk curve, across time. The one-step targeted maximum likelihood procedure is based on a one-dimensional universal least favourable submodel for each cause-specific hazard that we implement in recursive steps along a corresponding non-universal multivariate least favourable submodel. Our empirical study demonstrates the practical use of the methods.
- Published
- 2023
29. Statistical summaries of unlabelled evolutionary trees
- Author
-
Rajanala Samyak and Julia A Palacios
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
SUMMARY Rooted and ranked phylogenetic trees are mathematical objects that are useful in modelling hierarchical data and evolutionary relationships with applications to many fields such as evolutionary biology and genetic epidemiology. Bayesian phylogenetic inference usually explores the posterior distribution of trees via Markov Chain Monte Carlo methods. However, assessing uncertainty and summarizing distributions remains challenging for these type of structures. While labelled phylogenetic trees have been extensively studied, relatively less literature exists for unlabelled trees which are increasingly useful, for example when % Similarly, in many instances, one seeks to summarize samples of trees obtained with different methods, or from different samples and environments, and wishes to assess the stability and generalizability of these summaries. In our paper, we exploit recently proposed distance metrics of unlabelled ranked binary trees and unlabelled ranked genealogies, or trees equipped with branch lengths, to define the Fréchet mean, variance, and interquartile sets as summaries of these tree distributions. We provide an efficient combinatorial optimization algorithm for computing the Fréchet mean of a sample or of distributions on unlabelled ranked tree shapes and unlabelled ranked genealogies. We show the applicability of our summary statistics for studying popular tree distributions and for comparing the SARS-CoV-2 evolutionary trees across different locations during the COVID-19 epidemic in 2020. Our current implementations are publicly available at github.com/RSamyak/fmatrix
- Published
- 2023
30. Benchmarking small area estimators
- Author
-
Malay Ghosh, William R. Bell, and Gauri Sankar Datta
- Subjects
Statistics and Probability ,Class (set theory) ,Applied Mathematics ,General Mathematics ,Orthographic projection ,Estimator ,Context (language use) ,Benchmarking ,Agricultural and Biological Sciences (miscellaneous) ,Small area estimation ,Econometrics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Mathematics - Abstract
This paper considers benchmarking issues in the context of small area estimation. We find optimal estimators within the class of benchmarked linear estimators under linear constraints. This extends existing results for external and internal benchmarking, and also links the two. Necessary and sufficient conditions for self-benchmarking are found for an augmented model. Most results of this paper are found using ideas of orthogonal projection Copyright 2013, Oxford University Press.
- Published
- 2012
31. Testing a linear time series model against its threshold extension
- Author
-
Guodong Li and Wai Keung Li
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Linear model ,Asymptotic distribution ,Agricultural and Biological Sciences (miscellaneous) ,Moving-average model ,Test statistic ,Null distribution ,Econometrics ,Applied mathematics ,Autoregressive–moving-average model ,Statistics, Probability and Uncertainty ,Threshold model ,General Agricultural and Biological Sciences ,Mathematics ,Statistical hypothesis testing - Abstract
This paper derives the asymptotic null distribution of a quasilikelihood ratio test statistic for an autoregressive moving average model against its threshold extension. The null hypothesis is that of no threshold, and the error term could be dependent. The asymptotic distribution is rather complicated, and all existing methods for approximating a distribution in the related literature fail to work. Hence, a novel bootstrap approximation based on stochastic permutation is proposed in this paper. Besides being robust to the assumptions on the error term, our method enjoys more flexibility and needs less computation when compared with methods currently used in the literature. Monte Carlo experiments give further support to the new approach, and an illustration is reported. Copyright 2011, Oxford University Press.
- Published
- 2011
32. Lasso-adjusted treatment effect estimation under covariate-adaptive randomization
- Author
-
Hanzhong Liu, Fuyi Tu, and Wei Ma
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
Summary We consider the problem of estimating and inferring treatment effects in randomized experiments. In practice, stratified randomization, or more generally, covariate-adaptive randomization, is routinely used in the design stage to balance treatment allocations with respect to a few variables that are most relevant to the outcomes. Then, regression is performed in the analysis stage to adjust the remaining imbalances to yield more efficient treatment effect estimators. Building upon and unifying recent results obtained for ordinary-least-squares adjusted estimators under covariate-adaptive randomization, this paper presents a general theory of regression adjustment that allows for model mis-specification and the presence of a large number of baseline covariates. We exemplify the theory on two lasso-adjusted treatment effect estimators, both of which are optimal in their respective classes. In addition, nonparametric consistent variance estimators are proposed to facilitate valid inferences, which work irrespective of the specific randomization methods used. The robustness and improved efficiency of the proposed estimators are demonstrated through numerical studies.
- Published
- 2022
33. Compound optimal allocation for individual and collective ethics in binary clinical trials
- Author
-
Alessandro Baldi Antognini, Alessandra Giovagnoli, A. Baldi Antognini, and A. Giovagnoli
- Subjects
Statistics and Probability ,Operations research ,Applied Mathematics ,General Mathematics ,Decision theory ,Survey sampling ,Binary number ,Agricultural and Biological Sciences (miscellaneous) ,Clinical trial ,Order (exchange) ,Optimal allocation ,Treatment effect ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Weighted arithmetic mean ,Mathematics - Abstract
In recent years, several authors have investigated response-adaptive allocation rules for comparative clinical trials, in order to favour, at each stage of the trial, the treatment that appears to be best. In this paper, we define admissible allocations, namely treatment assignments that cannot be simultaneously improved upon with respect to both a specific design criterion, reflecting the inferential properties of the experiment, and the proportion of patients assigned to the best treatment or treatments; we survey existing designs from this viewpoint. We also suggest combining information and ethical considerations by taking a suitable weighted mean of two corresponding standardized criteria, with weights that depend on the actual treatment effects. This compound criterion leads to a locally optimal allocation that can be targeted by some response-adaptive randomization rule. The paper mainly deals with the case of two treatments, but the suggested methodology is shown to extend to more than two. Copyright 2010, Oxford University Press.
- Published
- 2010
34. High-dimensional analysis of variance in multivariate linear regression
- Author
-
Zhipeng Lou, Xianyang Zhang, and Wei Biao Wu
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
SummaryIn this paper, we develop a systematic theory for high-dimensional analysis of variance in multivariate linear regression, where the dimension and the number of coefficients can both grow with the sample size. We propose a new U-type statistic to test linear hypotheses and establish a high-dimensional Gaussian approximation result under fairly mild moment assumptions. Our general framework and theory can be used to deal with the classical one-way multivariate analysis of variance, and the nonparametric one-way multivariate analysis of variance in high dimensions. To implement the test procedure, we introduce a sample-splitting-based estimator of the second moment of the error covariance and discuss its properties. A simulation study shows that our proposed test outperforms some existing tests in various settings.
- Published
- 2023
35. Efficient Bayesian inference for Gaussian copula regression models
- Author
-
Robert Kohn, David Chan, and Michael K. Pitt
- Subjects
Statistics and Probability ,Statistics::Theory ,Applied Mathematics ,General Mathematics ,Gaussian ,Bayesian probability ,Posterior probability ,Inference ,Regression analysis ,Bayesian inference ,Agricultural and Biological Sciences (miscellaneous) ,Statistics::Computation ,symbols.namesake ,Econometrics ,symbols ,Statistics::Methodology ,Graphical model ,Statistics, Probability and Uncertainty ,Marginal distribution ,General Agricultural and Biological Sciences ,Algorithm ,Mathematics - Abstract
A Gaussian copula regression model gives a tractable way of handling a multivariate regression when some of the marginal distributions are non-Gaussian. Our paper presents a general Bayesian approach for estimating a Gaussian copula model that can handle any combination of discrete and continuous marginals, and generalises Gaussian graphical models to the Gaussian copula framework. Posterior inference is carried out using a novel and efficient simulation method. The methods in the paper are applied to simulated and real data.
- Published
- 2006
36. On recovering a population covariance matrix in the presence of selection bias
- Author
-
Zhihong Cai and Manabu Kuroki
- Subjects
Statistics and Probability ,Selection bias ,education.field_of_study ,Covariance function ,Covariance matrix ,Applied Mathematics ,General Mathematics ,media_common.quotation_subject ,Population ,Agricultural and Biological Sciences (miscellaneous) ,Estimation of covariance matrices ,symbols.namesake ,Causal inference ,Statistics ,Econometrics ,symbols ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,education ,Fisher information ,media_common ,Mathematics ,Statistical hypothesis testing - Abstract
This paper considers the problem of using observational data in the presence of selection bias to identify causal effects in the framework of linear structural equation models. We propose a criterion for testing whether or not observed statistical dependencies among variables are generated by conditioning on a common response variable. When the answer is affirmative, we further provide formulations for recovering the covariance matrix of the whole population from that of the selected population. The results of this paper provide guidance for reliable causal inference, based on the recovered covariance matrix obtained from the statistical information with selection bias.
- Published
- 2006
37. Using the periodogram to estimate period in nonparametric regression
- Author
-
Peter Hall and Ming Li
- Subjects
Statistics and Probability ,Schedule ,Applied Mathematics ,General Mathematics ,Nonparametric statistics ,Estimator ,Context (language use) ,Variance (accounting) ,Agricultural and Biological Sciences (miscellaneous) ,Nonparametric regression ,Rate of convergence ,Statistics ,Econometrics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Mathematics ,Central limit theorem - Abstract
SUMMARY Properties of the periodogram are seldom studied in the setting of nonparametric regression, although that is the context in which the periodogram is widely applied in astronomy. There it is a competitor with more recent least-squares methods. The periodogram has the advantage of providing significant graphical insight into statistical and numerical aspects of the problem. However, as we show in the present paper, it also has drawbacks. The estimator that it produces has somewhat higher variance than its least-squares counterpart, and a periodogram-based approach is more prone to suffer difficulties caused by periodicity of the observation schedule. While the periodogram remains a very attractive tool, the information provided in this paper allows users to assess more readily the extent to which it can be relied upon in a nonparametric setting. This aspect of our contributions is discussed theoretically and illustrated by numerical studies involving a real dataset.
- Published
- 2006
38. An approximate randomization test for the high-dimensional two-sample Behrens–Fisher problem under arbitrary covariances
- Author
-
Rui Wang and Wangli Xu
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,FOS: Mathematics ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
This paper is concerned with the problem of comparing the population means of two groups of independent observations. An approximate randomization test procedure based on the test statistic of Chen and Qin (2010) is proposed. The asymptotic behavior of the test statistic as well as the randomized statistic is studied under weak conditions. In our theoretical framework, observations are not assumed to be identically distributed even within groups. No condition on the eigenstructure of the covariance matrices is imposed. And the sample sizes of the two groups are allowed to be unbalanced. Under general conditions, all possible asymptotic distributions of the test statistic are obtained. We derive the asymptotic level and local power of the approximate randomization test procedure. Our theoretical results show that the proposed test procedure can adapt to all possible asymptotic distributions of the test statistic and always has correct test level asymptotically. Also, the proposed test procedure has good power behavior. Our numerical experiments show that the proposed test procedure has favorable performance compared with several alternative test procedures., Comment: 52 pages. To appear in Biometrika
- Published
- 2022
39. Functional hybrid factor regression model for handling heterogeneity in imaging studies
- Author
-
C, Huang and H, Zhu
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
Summary This paper develops a functional hybrid factor regression modelling framework to handle the heterogeneity of many large-scale imaging studies, such as the Alzheimer’s disease neuroimaging initiative study. Despite the numerous successes of those imaging studies, such heterogeneity may be caused by the differences in study environment, population, design, protocols or other hidden factors, and it has posed major challenges in integrative analysis of imaging data collected from multicentres or multistudies. We propose both estimation and inference procedures for estimating unknown parameters and detecting unknown factors under our new model. The asymptotic properties of both estimation and inference procedures are systematically investigated. The finite-sample performance of our proposed procedures is assessed by using Monte Carlo simulations and a real data example on hippocampal surface data from the Alzheimer’s disease study.
- Published
- 2022
40. Regression of exchangeable relational arrays
- Author
-
F W Marrs, B K Fosdick, and T H Mccormick
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
Summary Relational arrays represent measures of association between pairs of actors, often in varied contexts or over time. Trade flows between countries, financial transactions between individuals, contact frequencies between school children in classrooms and dynamic protein-protein interactions are all examples of relational arrays. Elements of a relational array are often modelled as a linear function of observable covariates. Uncertainty estimates for regression coefficient estimators, and ideally the coefficient estimators themselves, must account for dependence between elements of the array, e.g., relations involving the same actor. Existing estimators of standard errors that recognize such relational dependence rely on estimating extremely complex, heterogeneous structure across actors. This paper develops a new class of parsimonious coefficient and standard error estimators for regressions of relational arrays. We leverage an exchangeability assumption to derive standard error estimators that pool information across actors, and are substantially more accurate than existing estimators in a variety of settings. This exchangeability assumption is pervasive in network and array models in the statistics literature, but not previously considered when adjusting for dependence in a regression setting with relational data. We demonstrate improvements in inference theoretically, via a simulation study, and by analysis of a dataset involving international trade.
- Published
- 2022
41. Adaptive sampling for Bayesian variable selection
- Author
-
David J. Nott and Robert Kohn
- Subjects
Statistics and Probability ,Mathematical optimization ,Adaptive sampling ,Applied Mathematics ,General Mathematics ,Rejection sampling ,Slice sampling ,Sampling (statistics) ,Agricultural and Biological Sciences (miscellaneous) ,Statistics::Computation ,symbols.namesake ,Metropolis–Hastings algorithm ,Sampling design ,symbols ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Importance sampling ,Mathematics ,Gibbs sampling - Abstract
Our paper proposes adaptive Monte Carlo sampling schemes for Bayesian variable selection in linear regression that improve on standard Markov chain methods. We do so by considering Metropolis-Hastings proposals that make use of accumulated information about the posterior distribution obtained during sampling. Adaptation needs to be done carefully to ensure that sampling is from the correct ergodic distribution. We give conditions for the validity of an adaptive sampling scheme in this problem, and for simulating from a distribution on a finite state space in general, and suggest a class of adaptive proposal densities which uses best linear prediction to approximate the Gibbs sampler. Our sampling scheme is computationally much faster per iteration than the Gibbs sampler, and when this is taken into account the efficiency gains when using our sampling scheme compared to alternative approaches are substantial in terms of precision of estimation of posterior quantities of interest for a given amount of computation time. We compare our method with other sampling schemes for examples involving both real and simulated data. The methodology developed in the paper can be extended to variable selection in more general problems.
- Published
- 2005
42. Diagnostic checking for time series models with conditional heteroscedasticity estimated by the least absolute deviation approach
- Author
-
Wai Keung Li and Guodong Li
- Subjects
Statistics and Probability ,Statistics::Theory ,Heteroscedasticity ,Applied Mathematics ,General Mathematics ,Autoregressive conditional heteroskedasticity ,Asymptotic distribution ,Conditional probability distribution ,Asymptotic theory (statistics) ,Agricultural and Biological Sciences (miscellaneous) ,Autoregressive model ,Statistics ,Least absolute deviations ,Median absolute deviation ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Mathematics - Abstract
The recent paper by Peng & Yao (2003) gave an interesting extension of least absolute deviation estimation to generalised autoregressive conditional heteroscedasticity, GARCH, time series models. The asymptotic distributions of absolute residual autocorrelations and squared residual autocorrelations from the GARCH model estimated by the least absolute deviation method are derived in this paper. These results lead to two useful diagnostic tools which can be used to check whether or not a GARCH model fitted by using the least absolute deviation method is adequate. Some simulation experiments give further support to the asymptotic theory and a real data example is also reported.
- Published
- 2005
43. A paradox concerning nuisance parameters and projected estimating functions
- Author
-
Masayuki Henmi and Shinto Eguchi
- Subjects
Statistics and Probability ,Nuisance variable ,Estimation theory ,Applied Mathematics ,General Mathematics ,Estimator ,Statistical model ,Context (language use) ,Agricultural and Biological Sciences (miscellaneous) ,Efficient estimator ,Econometrics ,Nuisance parameter ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Projection (set theory) ,Mathematics - Abstract
SUMMARY This paper is concerned with a paradox associated with parameter estimation in the presence of nuisance parameters. In a statistical model with unknown nuisance parameters, the efficiency of an estimator of a parameter usually increases when the nuisance para meters are known. However the opposite phenomenon can sometimes occur. In this paper, we elucidate the occurrence of this paradox by examining estimating functions. In particular, we focus on the projected estimating function, which is defined by the projection of the score function on to a given estimating function. A sufficient condition for the paradox to occur is the orthogonality of the two components of the projected estimating functions corresponding to parameters of interest and nuisance parameters. In addition, a numerical assessment is conducted in the context of a simple model to investigate the improvement of the asymptotic efficiency of estimators.
- Published
- 2004
44. Measures for designs in experiments with correlated errors
- Author
-
Werner G. Müller and Andrej Pázman
- Subjects
Statistics and Probability ,Optimal design ,Biometrics ,Iterative method ,Applied Mathematics ,General Mathematics ,Computation ,Sampling (statistics) ,Agricultural and Biological Sciences (miscellaneous) ,Statistics ,Linear regression ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Algorithm ,Smoothing ,Brownian motion ,Mathematics - Abstract
SUMMARY In this paper we consider optimal design of experiments in the case of correlated obser- vations. We use and further develop the concept of design measures introduced by Paizman & Mtiller (1998) for the construction of a simple, quick and elegant design algorithm. We support the construction of this algorithm for a general correlation structure by an interpretation in terms of norms. Examples demonstrate that our results are useful for generating exact designs by sampling from the obtained design measures. Most of the literature on experimental design operates under the assumption of uncorre- lated errors and by exploiting the concept of a design measure introduced by Kiefer (1959). In this setting an optimal design measure is usually computed by an iterative algorithm and then an exact design with independent replications approximately proportional to that measure is employed. In the correlated case when replications are not allowed the implementation of design measures is not so straightforward. Recently, however, by adding virtual design-dependent noise to the process, Paizman & Muiller (1998) have introduced a way of using this popular concept, albeit requiring a very different interpretation. Furthermore, Paizman & Muller (2000) showed that solving minimisation problems in terms of these measures corresponds to computation of exact designs. These measures serve as building blocks for the methods presented in this paper, which provides a general numerical tool for screening designs. By a different type of smoothing of some important nondifferentiable terms in the algorithm, and without imposing restric- tions on the number of support points, we obtain a design measure. The relative magnitude of this measure for each support point reflects the importance of its inclusion in any exact design. This algorithm can also be interpreted in terms of certain 'information' norms as pre
- Published
- 2003
45. Selective factor extraction in high dimensions
- Author
-
Yiyuan She
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Rank (linear algebra) ,General Mathematics ,Machine Learning (stat.ML) ,Feature selection ,02 engineering and technology ,01 natural sciences ,Methodology (stat.ME) ,010104 statistics & probability ,Statistics - Machine Learning ,0202 electrical engineering, electronic engineering, information engineering ,0101 mathematics ,Coefficient matrix ,Projection (set theory) ,Statistics - Methodology ,Selection (genetic algorithm) ,Mathematics ,business.industry ,Applied Mathematics ,Model selection ,020206 networking & telecommunications ,Pattern recognition ,Data structure ,Agricultural and Biological Sciences (miscellaneous) ,Unsupervised learning ,Artificial intelligence ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,business - Abstract
SUMMARY This paper studies simultaneous feature selection and extraction in supervised and unsupervised learning. We propose and investigate selective reduced rank regression for constructing optimal explanatory factors from a parsimonious subset of input features. The proposed estimators enjoy sharp oracle inequalities, and with a predictive information criterion for model selection, they adapt to unknown sparsity by controlling both rank and row support of the coefficient matrix. A class of algorithms is developed that can accommodate various convex and nonconvex sparsity-inducing penalties, and can be used for rank-constrained variable screening in high-dimensional multivariate data. The paper also showcases applications in macroeconomics and computer vision to demonstrate how low-dimensional data structures can be effectively captured by joint variable selection and projection.
- Published
- 2017
46. Joint latent space models for network data with high-dimensional node variables
- Author
-
Xuefei Zhang, Gongjun Xu, and Ji Zhu
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
Summary Network latent space models assume that each node is associated with an unobserved latent position in a Euclidean , and such latent variables determine the probability of two nodes connecting with each other. In many applications, nodes in the network are often observed along with high-dimensional node variables, and these node variables provide important information for understanding the network structure. However, classical network latent space models have several limitations in incorporating node variables. In this paper, we propose a joint latent space model where we assume that the latent variables not only explain the network structure, but are also informative for the multivariate node variables. We develop a projected gradient descent algorithm that estimates the latent positions using a criterion incorporating both network structure and node variables. We establish theoretical properties of the estimators and provide insights into how incorporating high-dimensional node variables could improve the estimation accuracy of the latent positions. We demonstrate the improvement in latent variable estimation and the improvements in associated downstream tasks, such as missing value imputation for node variables, by simulation studies and an application to a Facebook data example.
- Published
- 2021
47. Instrumental variable estimation of the marginal structural Cox model for time-varying treatments
- Author
-
Y Cui, H Michael, F Tanser, and E Tchetgen Tchetgen
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
Summary Robins (1998) introduced marginal structural models, a general class of counterfactual models for the joint effects of time-varying treatments in complex longitudinal studies subject to time-varying confounding. Robins (1998) established the identification of marginal structural model parameters under a sequential randomization assumption, which rules out unmeasured confounding of treatment assignment over time. The marginal structural Cox model is one of the most popular marginal structural models for evaluating the causal effect of time-varying treatments on a censored failure time outcome. In this paper, we establish sufficient conditions for identification of marginal structural Cox model parameters with the aid of a time-varying instrumental variable, in the case where sequential randomization fails to hold due to unmeasured confounding. Our instrumental variable identification condition rules out any interaction between an unmeasured confounder and the instrumental variable in its additive effects on the treatment process, the longitudinal generalization of the identifying condition of Wang & Tchetgen Tchetgen (2018). We describe a large class of weighted estimating equations that give rise to consistent and asymptotically normal estimators of the marginal structural Cox model, thereby extending the standard inverse probability of treatment weighted estimation of marginal structural models to the instrumental variable setting. Our approach is illustrated via extensive simulation studies and an application to estimating the effect of community antiretroviral therapy coverage on HIV incidence.
- Published
- 2021
48. Searching for robust associations with a multi-environment knockoff filter
- Author
-
Matteo Sesia, Yaniv Romano, Chiara Sabatti, Emmanuel J. Candès, and Shuangning Li
- Subjects
Statistics and Probability ,Filter (video) ,business.industry ,Applied Mathematics ,General Mathematics ,Computer vision ,Artificial intelligence ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,business ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics - Abstract
Summary In this article we develop a method based on model-X knockoffs to find conditional associations that are consistent across environments, while controlling the false discovery rate. The motivation for this problem is that large datasets may contain numerous associations that are statistically significant and yet misleading, as they are induced by confounders or sampling imperfections. However, associations replicated under different conditions may be more interesting. In fact, sometimes consistency provably leads to valid causal inferences even if conditional associations do not. Although the proposed method is widely applicable, in this paper we highlight its relevance to genome-wide association studies, in which robustness across populations with diverse ancestries mitigates confounding due to unmeasured variants. The effectiveness of this approach is demonstrated by simulations and applications to UK Biobank data.
- Published
- 2021
49. On the power of Chatterjee’s rank correlation
- Author
-
Hongjian Shi, Mathias Drton, and Fang Han
- Subjects
Statistics and Probability ,Power (social and political) ,Applied Mathematics ,General Mathematics ,Statistics ,Chatterjee ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) ,Mathematics ,Rank correlation - Abstract
Summary Chatterjee (2021) introduced a simple new rank correlation coefficient that has attracted much attention recently. The coefficient has the unusual appeal that it not only estimates a population quantity first proposed by Dette et al. (2013) that is zero if and only if the underlying pair of random variables is independent, but also is asymptotically normal under independence. This paper compares Chatterjee’s new correlation coefficient with three established rank correlations that also facilitate consistent tests of independence, namely Hoeffding’s $D$, Blum–Kiefer–Rosenblatt’s $R$, and Bergsma–Dassios–Yanagimoto’s $\tau^*$. We compare the computational efficiency of these rank correlation coefficients in light of recent advances, and investigate their power against local rotation and mixture alternatives. Our main results show that Chatterjee’s coefficient is unfortunately rate-suboptimal compared to $D$, $R$ and $\tau^*$. The situation is more subtle for a related earlier estimator of Dette et al. (2013). These results favour $D$, $R$ and $\tau^*$ over Chatterjee’s new correlation coefficient for the purpose of testing independence.
- Published
- 2021
50. Dependent censoring based on parametric copulas
- Author
-
C Czado and I Van Keilegom
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Agricultural and Biological Sciences (miscellaneous) - Abstract
Summary Consider a survival time $T$ that is subject to random right censoring, and suppose that $T$ is stochastically dependent on the censoring time $C$. We are interested in the marginal distribution of $T$. This situation is often encountered in practice. Consider, for example, the case where $T$ is a patient’s time to death from a certain disease. Then the censoring time $C$ could be the time until the patient leaves the study or the time until death from another cause. If the reason for leaving the study is related to the health condition of the patient, or if the patient dies from a disease that has similar risk factors to the disease of interest, then $T$ and $C$ are likely to be dependent. In this paper we propose a new model that takes such dependence into account. The model is based on a parametric copula for the relationship between $T$ and $C$, and on parametric marginal distributions for $T$ and $C$. Unlike most other authors, we do not assume that the parameter defining the copula is known. We give sufficient conditions on these parametric copulas and marginals under which the bivariate distribution of $(T,C)$ is identified. These sufficient conditions are then checked for a wide range of common copulas and marginals. We also study the estimation of the model, and carry out extensive simulations and analysis on a pancreatic cancer dataset to illustrate the proposed model and estimation procedure.
- Published
- 2022
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.