2,907 results on '"Gustafson, Paul"'
Search Results
2. Issues of parameterization and computation for posterior inference in partially identified models
- Author
-
Lee, Seren and Gustafson, Paul
- Subjects
Statistics - Computation - Abstract
A partially identified model, where the parameters can not be uniquely identified, often arises during statistical analysis. While researchers frequently use Bayesian inference to analyze the models, when Bayesian inference with an off-the-shelf MCMC sampling algorithm is applied to a partially identified model, the computational performance can be poor. It is found that using importance sampling with transparent reparameterization (TP) is one remedy. This method is preferable since the model is known to be rendered as identified with respect to the new parameterization, and at the same time, it may allow faster, i.i.d. Monte Carlo sampling by using conjugate convenience priors. In this paper, we explain the importance sampling method with the TP and a pseudo-TP. We introduce the pseudo-TP, an alternative to TP, since finding a TP is sometimes difficult. Then, we test the methods' performance in some scenarios and compare it to the performance of the off-the-shelf MCMC method - Gibbs sampling - applied in the original parameterization. While the importance sampling with TP (ISTP) shows generally better results than off-the-shelf MCMC methods, as seen in the compute time and trace plots, it is also seen that finding a TP which is necessary for the method may not be easy. On the other hand, the pseudo-TP method shows a mixed result and room for improvement since it relies on an approximation, which may not be adequate for a given model and dataset.
- Published
- 2024
3. Unmasking Bias: A Framework for Evaluating Treatment Benefit Predictors Using Observational Studies
- Author
-
Xia, Yuan, Sadatsafavi, Mohsen, and Gustafson, Paul
- Subjects
Statistics - Methodology - Abstract
Treatment benefit predictors (TBPs) map patient characteristics into an estimate of the treatment benefit tailored to individual patients, which can support optimizing treatment decisions. However, the assessment of their performance might be challenging with the non-random treatment assignment. This study conducts a conceptual analysis, which can be applied to finite-sample studies. We present a framework for evaluating TBPs using observational data from a target population of interest. We then explore the impact of confounding bias on TBP evaluation using measures of discrimination and calibration, which are the moderate calibration and the concentration of the benefit index ($C_b$), respectively. We illustrate that failure to control for confounding can lead to misleading values of performance metrics and establish how the confounding bias propagates to an evaluation bias to quantify the explicit bias for the performance metrics. These findings underscore the necessity of accounting for confounding factors when evaluating TBPs, ensuring more reliable and contextually appropriate treatment decisions., Comment: 31 pages, 5 figures
- Published
- 2024
4. A fully Bayesian approach for the imputation and analysis of derived outcome variables with missingness
- Author
-
Campbell, Harlan, Morris, Tim, and Gustafson, Paul
- Subjects
Statistics - Methodology - Abstract
Derived variables are variables that are constructed from one or more source variables through established mathematical operations or algorithms. For example, body mass index (BMI) is a derived variable constructed from two source variables: weight and height. When using a derived variable as the outcome in a statistical model, complications arise when some of the source variables have missing values. In this paper, we propose how one can define a single fully Bayesian model to simultaneously impute missing values and sample from the posterior. We compare our proposed method with alternative approaches that rely on multiple imputation, and, with a simulated dataset, consider how best to estimate the risk of microcephaly in newborns exposed to the ZIKA virus.
- Published
- 2024
5. Integrating representative and non-representative survey data for efficient inference
- Author
-
Dyrkton, Nathaniel, Gustafson, Paul, and Campbell, Harlan
- Subjects
Statistics - Methodology ,Statistics - Applications - Abstract
Non-representative surveys are commonly used and widely available but suffer from selection bias that generally cannot be entirely eliminated using weighting techniques. Instead, we propose a Bayesian method to synthesize longitudinal representative unbiased surveys with non-representative biased surveys by estimating the degree of selection bias over time. We show using a simulation study that synthesizing biased and unbiased surveys together out-performs using the unbiased surveys alone, even if the selection bias may evolve in a complex manner over time. Using COVID-19 vaccination data, we are able to synthesize two large sample biased surveys with an unbiased survey to reduce uncertainty in now-casting and inference estimates while simultaneously retaining the empirical credible interval coverage. Ultimately, we are able to conceptually obtain the properties of a large sample unbiased survey if the assumed unbiased survey, used to anchor the estimates, is unbiased for all time-points., Comment: New version includes fixed typos, Monte Carlo Standard error for the simulation (added in V2), and some clarifications
- Published
- 2024
6. The expected value of sample information calculations for external validation of risk prediction models
- Author
-
Sadatsafavi, Mohsen, Vickers, Andrew J, Lee, Tae Yoon, Gustafson, Paul, and Wynants, Laure
- Subjects
Statistics - Applications - Abstract
In designing external validation studies of clinical prediction models, contemporary sample size calculation methods are based on the frequentist inferential paradigm. One of the widely reported metrics of model performance is net benefit (NB), and the relevance of conventional inference around NB as a measure of clinical utility is doubtful. Value of Information methodology quantifies the consequences of uncertainty in terms of its impact on clinical utility of decisions. We introduce the expected value of sample information (EVSI) for validation as the expected gain in NB from conducting an external validation study of a given size. We propose algorithms for EVSI computation, and in a case study demonstrate how EVSI changes as a function of the amount of current information and future study's sample size. Value of Information methodology provides a decision-theoretic lens to the process of planning a validation study of a risk prediction model and can complement conventional methods when designing such studies., Comment: 14 pages, 4 figures, 0 tables
- Published
- 2024
7. Boolean TQFTs with accumulating defects, sofic systems, and automata for infinite words
- Author
-
Gustafson, Paul, Im, Mee Seong, and Khovanov, Mikhail
- Subjects
Mathematics - Category Theory ,Computer Science - Formal Languages and Automata Theory ,Mathematical Physics ,Mathematics - Dynamical Systems ,Mathematics - Quantum Algebra ,Primary: 57K16, 68Q45, 18M05, 37B10, Secondary: 06A12, 68Q70, 18B20 - Abstract
Any finite state automaton gives rise to a Boolean one-dimensional TQFT with defects and inner endpoints of cobordisms. This paper extends the correspondence to Boolean TQFTs where defects accumulate toward inner endpoints, relating such TQFTs and topological theories to sofic systems and $\omega$-automata., Comment: 31 pages, many figures
- Published
- 2023
8. Partial identification for discrete data with nonignorable missing outcomes
- Author
-
Daly-Grafstein, Daniel and Gustafson, Paul
- Subjects
Statistics - Methodology - Abstract
Nonignorable missing outcomes are common in real world datasets and often require strong parametric assumptions to achieve identification. These assumptions can be implausible or untestable, and so we may forgo them in favour of partially identified models that narrow the set of a priori possible values to an identification region. Here we propose a new nonparametric Bayes method that allows for the incorporation of multiple clinically relevant restrictions of the parameter space simultaneously. We focus on two common restrictions, instrumental variables and the direction of missing data bias, and investigate how these restrictions narrow the identification region for parameters of interest. Additionally, we propose a rejection sampling algorithm that allows us to quantify the evidence for these assumptions in the data. We compare our method to a standard Heckman selection model in both simulation studies and in an applied problem examining the effectiveness of cash-transfers for people experiencing homelessness., Comment: 41 pages, 4 figures, 4 tables, added section 2.3 and updated figure 4
- Published
- 2023
9. Application of causal inference methods in individual-participant data meta-analyses in medicine: addressing data handling and reporting gaps with new proposed reporting guidelines
- Author
-
Hufstedler, Heather, Mauer, Nicole, Yeboah, Edmund, Carr, Sinclair, Rahman, Sabahat, Danzer, Alexander M., Debray, Thomas P. A., de Jong, Valentijn M.T., Campbell, Harlan, Gustafson, Paul, Maxwell, Lauren, Jaenisch, Thomas, Matthay, Ellicott C., and Bärnighausen, Till
- Published
- 2024
- Full Text
- View/download PDF
10. Automata and one-dimensional TQFTs with defects
- Author
-
Gustafson, Paul, Im, Mee Seong, Kaldawy, Remy, Khovanov, Mikhail, and Lihn, Zachary
- Subjects
Mathematics - Quantum Algebra ,Computer Science - Formal Languages and Automata Theory ,Mathematics - Category Theory ,Primary: 57K16, 68Q45, 18M10, 18M30, Secondary: 06A12, 68Q70, 18B20 - Abstract
This paper explains how any nondeterministic automaton for a regular language $L$ gives rise to a one-dimensional oriented Topological Quantum Field Theory (TQFT) with inner endpoints and zero-dimensional defects labelled by letters of the alphabet for $L$. The TQFT is defined over the Boolean semiring $\mathbb{B}$. Different automata for a fixed language $L$ produce TQFTs that differ by their values on decorated circles, while the values on decorated intervals are described by the language $L$. The language $L$ and the TQFT associated to an automaton can be given a path integral interpretation. In this TQFT the state space of a one-point 0-manifold is a free module over $\mathbb{B}$ with the basis of states of the automaton. Replacing a free module by a finite projective $\mathbb{B}$-module $P$ allows to generalize automata and this type of TQFT to a structure where defects act on open subsets of a finite topological space. Intersection of open subsets induces a multiplication on $P$ allowing to extend the TQFT to a TQFT for one-dimensional foams (oriented graphs with defects modulo a suitable equivalence relation). A linear version of these constructions is also explained, with the Boolean semiring replaced by a commutative ring., Comment: Corollary 3.5 added. 36 pages, many figures
- Published
- 2023
11. Defining a credible interval is not always possible with 'point-null' priors: A lesser-known correlate of the Jeffreys-Lindley paradox
- Author
-
Campbell, Harlan and Gustafson, Paul
- Subjects
Mathematics - Statistics Theory - Abstract
In many common situations, a Bayesian credible interval will be, given the same data, very similar to a frequentist confidence interval, and researchers will interpret these intervals in a similar fashion. However, no predictable similarity exists when credible intervals are based on model-averaged posteriors whenever one of the two nested models under consideration is a so called ''point-null''. Not only can this model-averaged credible interval be quite different than the frequentist confidence interval, in some cases it may be undefined. This is a lesser-known correlate of the Jeffreys-Lindley paradox and is of particular interest given the popularity of the Bayes factor for testing point-null hypotheses.
- Published
- 2022
12. Issues in Implementing Regression Calibration Analyses
- Author
-
Boe, Lillian, Shaw, Pamela A., Midthune, Douglas, Gustafson, Paul, Kipnis, Victor, Park, Eunyoung, Sotres-Alvarez, Daniela, and Freedman, Laurence
- Subjects
Statistics - Methodology - Abstract
Regression calibration is a popular approach for correcting biases in estimated regression parameters when exposure variables are measured with error. This approach involves building a calibration equation to estimate the value of the unknown true exposure given the error-prone measurement and other confounding covariates. The estimated, or calibrated, exposure is then substituted for the true exposure in the health outcome regression model. When used properly, regression calibration can greatly reduce the bias induced by exposure measurement error. Here, we first provide an overview of the statistical framework for regression calibration, specifically discussing how a special type of error, called Berkson error, arises in the estimated exposure. We then present practical issues to consider when applying regression calibration, including: (1) how to develop the calibration equation and which covariates to include; (2) valid ways to calculate standard errors (SE) of estimated regression coefficients; and (3) problems arising if one of the covariates in the calibration model is a mediator of the relationship between the exposure and outcome. Throughout the paper, we provide illustrative examples using data from the Hispanic Community Health Study/Study of Latinos (HCHS/SOL) and simulations. We conclude with recommendations for how to perform regression calibration.
- Published
- 2022
13. Methodological concerns about 'concordance-statistic for benefit' as a measure of discrimination in treatment benefit prediction
- Author
-
Xia, Yuan, Gustafson, Paul, and Sadatsafavi, Mohsen
- Subjects
Statistics - Methodology ,Statistics - Applications - Abstract
Prediction algorithms that quantify the expected benefit of a given treatment conditional on patient characteristics can critically inform medical decisions. Quantifying the performance of treatment benefit prediction algorithms is an active area of research. A recently proposed metric, the concordance statistic for benefit (cfb), evaluates the discriminative ability of a treatment benefit predictor by directly extending the concept of the concordance statistic from a risk model with a binary outcome to a model for treatment benefit. In this work, we scrutinize $cfb$ on multiple fronts. Through numerical examples and theoretical developments, we show that cfb is not a proper scoring rule. We also show that it is sensitive to the unestimable correlation between counterfactual outcomes and to the definition of matched pairs. We argue that measures of statistical dispersion applied to predicted benefits do not suffer from these issues and can be an alternative metric for the discriminatory performance of treatment benefit predictors., Comment: 12 pages, 6 figures
- Published
- 2022
- Full Text
- View/download PDF
14. Value of Information Analysis for External Validation of Risk Prediction Models
- Author
-
Sadatsafavi, Mohsen, Lee, Tae Yoon, Wynants, Laure, Vickers, Andrew, and Gustafson, Paul
- Subjects
Statistics - Applications - Abstract
Background: Before being used to inform patient care, a risk prediction model needs to be validated in a representative sample from the target population. The finite size of the validation sample entails that there is uncertainty with respect to estimates of model performance. We apply value-of-information methodology as a framework to quantify the consequence of such uncertainty in terms of NB. Methods: We define the Expected Value of Perfect Information (EVPI) for model validation as the expected loss in NB due to not confidently knowing which of the alternative decisions confers the highest NB at a given risk threshold. We propose methods for EVPI calculations based on Bayesian or ordinary bootstrapping of NBs, as well as an asymptotic approach supported by the central limit theorem. We conducted brief simulation studies to compare the performance of these methods, and used subsets of data from an international clinical trial for predicting mortality after myocardial infarction as a case study. Results: The three computation methods generated similar EVPI values in simulation studies. In the case study, at the pre-specified threshold of 0.02, the best decision with current information would be to use the model, with an expected incremental NB of 0.0020 over treating all. At this threshold, EVPI was 0.0005 (a relative EVPI of 25%). When scaled to the annual number of heart attacks in the US, this corresponds to a loss of 400 true positives, or extra 19,600 false positives (unnecessary treatments) per year, indicating the value of further model validation. As expected, the validation EVPI generally declined with larger samples. Conclusion: Value-of-information methods can be applied to the NB calculated during external validation of clinical prediction models to provide a decision-theoretic perspective to the consequences of uncertainty., Comment: 24 pages, 4,484 words, 1 table, 2 boxes, 5 figures
- Published
- 2022
- Full Text
- View/download PDF
15. Closed-Form Solution of the Unit Normal Loss Integral in Two-Dimensions
- Author
-
Lee, Tae Yoon, Gustafson, Paul, and Sadatsafavi, Mohsen
- Subjects
Statistics - Computation ,Mathematics - Statistics Theory - Abstract
In Value of Information (VoI) analysis, the unit normal loss integral (UNLI) frequently emerges as a solution for the computation of various VoI metrics. However, one limitation of the UNLI has been that its closed-form solution is available for only one dimension, and thus can be used for comparisons involving only two strategies (where it is applied to the scalar incremental net benefit). We derived a closed-form solution for the two-dimensional UNLI, enabling closed-form VoI calculations for three strategies. We verified the accuracy of this method via simulation studies. A case study based on a three-arm clinical trial was used as an example. VoI methods based on the closed-form solutions for the UNLI can now be extended to three-decision comparisons, taking a fraction of a second to compute and not being subject to Monte Carlo error. An R implementation of this method is provided as part of the predtools package (https://github.com/resplab/predtools/)., Comment: 1 table, 1 figure, will be submitted to MDM - technical note
- Published
- 2022
- Full Text
- View/download PDF
16. Bayes factors and posterior estimation: Two sides of the very same coin
- Author
-
Campbell, Harlan and Gustafson, Paul
- Subjects
Mathematics - Statistics Theory - Abstract
Recently, several researchers have claimed that conclusions obtained from a Bayes factor (or the posterior odds) may contradict those obtained from Bayesian posterior estimation. In this short paper, we wish to point out that no such "incompatibility" exists if one is willing to consistently define one's priors and posteriors. The key for compatibility is that the (implied) prior model odds used for testing are the same as those used for estimation. Our recommendation is simple: If one reports a Bayes factor comparing two models, then one should also report posterior estimates which appropriately acknowledge the uncertainty with regards to which of the two models is correct., Comment: 12 pages
- Published
- 2022
17. Risk Prediction in Sexual Health Contexts: Protocol
- Author
-
Falasinnu, Titilola, Gustafson, Paul, Gilbert, Mark, and Shoveller, Jean
- Subjects
Medicine ,Computer applications to medicine. Medical informatics ,R858-859.7 - Abstract
BackgroundIn British Columbia (BC), we are developing Get Checked Online (GCO), an Internet-based testing program that provides Web-based access to sexually transmitted infections (STI) testing. Much is still unknown about how to implement risk assessment and recommend tests in Web-based settings. Prediction tools have been shown to successfully increase efficiency and cost-effectiveness of STI case finding in the following settings. ObjectiveThis project was designed with three main objectives: (1) to derive a risk prediction rule for screening chlamydia and gonorrhea among clients attending two public sexual health clinics between 2000 and 2006 in Vancouver, BC, (2) to assess the temporal generalizability of the prediction rule among more recent visits in the Vancouver clinics (2007-2012), and (3) to assess the geographical generalizability of the rule in seven additional clinics in BC. MethodsThis study is a population-based, cross-sectional analysis of electronic records of visits collected at nine publicly funded STI clinics in BC between 2000 and 2012. We will derive a risk score from the multivariate logistic regression of clinic visit data between 2000 and 2006 at two clinics in Vancouver using newly diagnosed chlamydia and gonorrhea infections as the outcome. The area under the receiver operating characteristic curve (AUC) and the Hosmer-Lemeshow statistic will examine the model’s discrimination and calibration, respectively. We will also examine the sensitivity and proportion of patients that would need to be screened at different cutoffs of the risk score. Temporal and geographical validation will be assessed using patient visit data from more recent visits (2007-2012) at the Vancouver clinics and at clinics in the rest of BC, respectively. Statistical analyses will be performed using SAS, version 9.3. ResultsThis is an ongoing research project with initial results expected in 2014. ConclusionsThe results from this research will have important implications for scaling up of Internet-based testing in BC. If a prediction rule with good calibration, discrimination, and high sensitivity to detect infection is found during this project, the prediction rule could be programmed into GCO so that the program offers individualized testing recommendations to clients. Further, the prediction rule could be adapted into educational materials to inform other Web-based content by creating awareness about STI risk factors, which may stimulate health care seeking behavior among individuals accessing the website.
- Published
- 2013
- Full Text
- View/download PDF
18. Adjusting for Misclassification of an Exposure in an Individual Participant Data Meta-Analysis
- Author
-
de Jong, Valentijn M. T., Campbell, Harlan, Maxwell, Lauren, Jaenisch, Thomas, Gustafson, Paul, and Debray, Thomas P. A.
- Abstract
A common problem in the analysis of multiple data sources, including individual participant data meta-analysis (IPD-MA), is the misclassification of binary variables. Misclassification may lead to biased estimators of model parameters, even when the misclassification is entirely random. We aimed to develop statistical methods that facilitate unbiased estimation of adjusted and unadjusted exposure-outcome associations and between-study heterogeneity in IPD-MA, where the extent and nature of exposure misclassification may vary across studies. We present Bayesian methods that allow misclassification of binary exposure variables to depend on study- and participant-level characteristics. In an example of the differential diagnosis of dengue using two variables, where the gold standard measurement for the exposure variable was unavailable for some studies which only measured a surrogate prone to misclassification, our methods yielded more accurate estimates than analyses naive with regard to misclassification or based on gold standard measurements alone. In a simulation study, the evaluated misclassification model yielded valid estimates of the exposure-outcome association, and was more accurate than analyses restricted to gold standard measurements. Our proposed framework can appropriately account for the presence of binary exposure misclassification in IPD-MA. It requires that some studies supply IPD for the surrogate and gold standard exposure, and allows misclassification to follow a random effects distribution across studies conditional on observed covariates (and outcome). The proposed methods are most beneficial when few large studies that measured the gold standard are available, and when misclassification is frequent.
- Published
- 2023
- Full Text
- View/download PDF
19. Adjusting for misclassification of an exposure in an individual participant data meta-analysis
- Author
-
de Jong, Valentijn M. T., Campbell, Harlan, Maxwell, Lauren, Jaenisch, Thomas, Gustafson, Paul, and Debray, Thomas P. A.
- Subjects
Statistics - Methodology - Abstract
A common problem in the analysis of multiple data sources, including individual participant data meta-analysis (IPD-MA), is the misclassification of binary variables. Misclassification may lead to biased estimates of model parameters, even when the misclassification is entirely random. We aimed to develop statistical methods that facilitate unbiased estimation of adjusted and unadjusted exposure-outcome associations and between-study heterogeneity in IPD-MA, where the extent and nature of exposure misclassification may vary across studies. We present Bayesian methods that allow misclassification of binary exposure variables to depend on study- and participant-level characteristics. In an example of the differential diagnosis of dengue using two variables, where the gold standard measurement for the exposure variable was unavailable for some studies which only measured a surrogate prone to misclassification, our methods yielded more accurate estimates than analyses naive with regard to misclassification or based on gold standard measurements alone. In a simulation study, the evaluated misclassification model yielded valid estimates of the exposure-outcome association, and was more accurate than analyses restricted to gold standard measurements. Our proposed framework can appropriately account for the presence of binary exposure misclassification in IPD-MA. It requires that some studies supply IPD for the surrogate and gold standard exposure and misclassification is exchangeable across studies conditional on observed covariates (and outcome). The proposed methods are most beneficial when few large studies that measured the gold standard are available, and when misclassification is frequent.
- Published
- 2021
20. Combining Parametric and Nonparametric Models to Estimate Treatment Effects in Observational Studies
- Author
-
Daly-Grafstein, Daniel and Gustafson, Paul
- Subjects
Statistics - Methodology ,Statistics - Computation - Abstract
Performing causal inference in observational studies requires we assume confounding variables are correctly adjusted for. G-computation methods are often used in these scenarios, with several recent proposals using Bayesian versions of g-computation. In settings with few confounders, standard models can be employed, however as the number of confounders increase these models become less feasible as there are fewer observations available for each unique combination of confounding variables. In this paper we propose a new model for estimating treatment effects in observational studies that incorporates both parametric and nonparametric outcome models. By conceptually splitting the data, we can combine these models while maintaining a conjugate framework, allowing us to avoid the use of MCMC methods. Approximations using the central limit theorem and random sampling allows our method to be scaled to high dimensional confounders while maintaining computational efficiency. We illustrate the model using carefully constructed simulation studies, as well as compare the computational costs to other benchmark models., Comment: 34 pages, 5 figures
- Published
- 2021
21. A pasting lemma for Lipschitz functions
- Author
-
Kvalheim, Matthew D., Gustafson, Paul, and Burden, Samuel A.
- Subjects
Mathematics - Classical Analysis and ODEs ,Mathematics - General Topology ,54E35, 54E45, 51F30 - Abstract
We give a necessary and sufficient condition ensuring that any function which is separately Lipschitz on two fixed compact sets is Lipschitz on their union., Comment: 4 pages
- Published
- 2021
22. Hybrid dynamical type theories for navigation
- Author
-
Gustafson, Paul, Culbertson, Jared, and Koditschek, Daniel E.
- Subjects
Computer Science - Logic in Computer Science ,Computer Science - Programming Languages ,Computer Science - Robotics ,Mathematics - Category Theory ,18C50 ,F.4.1 ,I.2.9 - Abstract
We present a hybrid dynamical type theory equipped with useful primitives for organizing and proving safety of navigational control algorithms. This type theory combines the framework of Fu--Kishida--Selinger for constructing linear dependent type theories from state-parameter fibrations with previous work on categories of hybrid systems under sequential composition. We also define a conjectural embedding of a fragment of linear-time temporal logic within our type theory, with the goal of obtaining interoperability with existing state-of-the-art tools for automatic controller synthesis from formal task specifications. As a case study, we use the type theory to organize and prove safety properties for an obstacle-avoiding navigation algorithm of Arslan--Koditschek as implemented by Vasilopoulos. Finally, we speculate on extensions of the type theory to deal with conjugacies between model and physical spaces, as well as hierarchical template-anchor relationships., Comment: 6 pages, 6 figures
- Published
- 2021
23. Methodological concerns about “concordance-statistic for benefit” as a measure of discrimination in predicting treatment benefit
- Author
-
Xia, Yuan, Gustafson, Paul, and Sadatsafavi, Mohsen
- Published
- 2023
- Full Text
- View/download PDF
24. Uncertainty and Value of Information in Risk Prediction Modeling
- Author
-
Sadatsafavi, Mohsen, Lee, Tae Yoon, and Gustafson, Paul
- Subjects
Statistics - Applications - Abstract
Background: Due to the finite size of the development sample, predicted probabilities from a risk prediction model are inevitably uncertain. We apply Value of Information methodology to evaluate the decision-theoretic implications of prediction uncertainty. Methods: Adopting a Bayesian perspective, we extend the definition of the Expected Value of Perfect Information (EVPI) from decision analysis to net benefit calculations in risk prediction. In the context of model development, EVPI is the expected gain in net benefit by using the correct predictions as opposed to predictions from a proposed model. We suggest bootstrap methods for sampling from the posterior distribution of predictions for EVPI calculation using Monte Carlo simulations. In a case study, we used subsets of data of various sizes from a clinical trial for predicting mortality after myocardial infarction to show how EVPI changes with sample size. Results: With a sample size of 1,000 and at the pre-specified threshold of 2% on predicted risks, the gain in net benefit by using the proposed and the correct models were 0.0006 and 0.0011, respectively, resulting in an EVPI of 0.0005 and a relative EVPI of 87%. EVPI was zero only at unrealistically high thresholds (>85%). As expected, EVPI declined with larger samples. We summarize an algorithm for incorporating EVPI calculations into the commonly used bootstrap method for optimism correction. Conclusion: Value of Information methods can be applied to explore decision-theoretic consequences of uncertainty in risk prediction and can complement inferential methods when developing risk prediction models. R code for implementing this method is provided., Comment: 24 pages, 1 table, 3 figures
- Published
- 2021
- Full Text
- View/download PDF
25. From Torus Bundles to Particle-Hole Equivariantization
- Author
-
Cui, Shawn X., Gustafson, Paul, Qiu, Yang, and Zhang, Qing
- Subjects
Mathematics - Quantum Algebra ,Mathematical Physics ,Mathematics - Geometric Topology ,Quantum Physics - Abstract
We continue the program of constructing (pre)modular tensor categories from 3-manifolds first initiated by Cho-Gang-Kim using $M$ theory in physics and then mathematically studied by Cui-Qiu-Wang. An important structure involved is a collection of certain $\text{SL}(2, \mathbb{C})$ characters on a given manifold which serve as the simple object types in the corresponding category. Chern-Simons invariants and adjoint Reidemeister torsions play a key role in the construction, and they are related to topological twists and quantum dimensions, respectively, of simple objects. The modular $S$-matrix is computed from local operators and follows a trial-and-error procedure. It is currently unknown how to produce data beyond the modular $S$- and $T$-matrices. There are also a number of subtleties in the construction which remain to be solved. In this paper, we consider an infinite family of 3-manifolds, that is, torus bundles over the circle. We show that the modular data produced by such manifolds are realized by the $\mathbb{Z}_2$-equivariantization of certain pointed premodular categories. Here the equivariantization is performed for the $\mathbb{Z}_2$-action sending a simple (invertible) object to its inverse, also called the particle-hole symmetry. It is our hope that this extensive class of examples will shed light on how to improve the program to recover the full data of a premodular category., Comment: 16 pages, minor changes, to appear in Lett Math Phys
- Published
- 2021
- Full Text
- View/download PDF
26. re:Linde et al. (2021): The Bayes factor, HDI-ROPE and frequentist equivalence tests can all be reverse engineered -- almost exactly -- from one another
- Author
-
Campbell, Harlan and Gustafson, Paul
- Subjects
Statistics - Methodology - Abstract
Following an extensive simulation study comparing the operating characteristics of three different procedures used for establishing equivalence (the frequentist "TOST", the Bayesian "HDI-ROPE", and the Bayes factor interval null procedure), Linde et al. (2021) conclude with the recommendation that "researchers rely more on the Bayes factor interval null approach for quantifying evidence for equivalence." We redo the simulation study of Linde et al. (2021) in its entirety but with the different procedures calibrated to have the same predetermined maximum type 1 error rate. Our results suggest that, when calibrated in this way, the Bayes Factor, HDI-ROPE, and frequentist equivalence tests all have similar -- almost exactly -- type 2 error rates. In general any advocating for frequentist testing as better or worse than Bayesian testing in terms of empirical findings seems dubious at best. If one decides on which underlying principle to subscribe to in tackling a given problem, then the method follows naturally. Bearing in mind that each procedure can be reverse-engineered from the others (at least approximately), trying to use empirical performance to argue for one approach over another seems like tilting at windmills., Comment: 16 pages, 5 figures
- Published
- 2021
27. Measurement Error in Meta-Analysis (MEMA) -- a Bayesian framework for continuous outcome data
- Author
-
Campbell, Harlan, de Jong, Valentijn M. T., Maxwell, Lauren, Debray, Thomas P. A., Jaenisch, Thomas, and Gustafson, Paul
- Subjects
Statistics - Methodology - Abstract
Ideally, a meta-analysis will summarize data from several unbiased studies. Here we consider the less than ideal situation in which contributing studies may be compromised by measurement error. Measurement error affects every study design, from randomized controlled trials to retrospective observational studies. We outline a flexible Bayesian framework for continuous outcome data which allows one to obtain appropriate point and interval estimates with varying degrees of prior knowledge about the magnitude of the measurement error. We also demonstrate how, if individual-participant data (IPD) are available, the Bayesian meta-analysis model can adjust for multiple participant-level covariates, measured with or without measurement error.
- Published
- 2020
- Full Text
- View/download PDF
28. Parameter Restrictions for the Sake of Identification: Is there Utility in Asserting that Perhaps a Restriction Holds?
- Author
-
Gustafson, Paul
- Subjects
Statistics - Methodology - Abstract
Statistical modeling can involve a tension between assumptions and statistical identification. The law of the observable data may not uniquely determine the value of a target parameter without invoking a key assumption, and, while plausible, this assumption may not be obviously true in the scientific context at hand. Moreover, there are many instances of key assumptions which are untestable, hence we cannot rely on the data to resolve the question of whether the target is legitimately identified. Working in the Bayesian paradigm, we consider the grey zone of situations where a key assumption, in the form of a parameter space restriction, is scientifically reasonable but not incontrovertible for the problem being tackled. Specifically, we investigate statistical properties that ensue if we structure a prior distribution to assert that `maybe' or `perhaps' the assumption holds. Technically this simply devolves to using a mixture prior distribution putting just some prior weight on the assumption, or one of several assumptions, holding. However, while the construct is straightforward, there is very little literature discussing situations where Bayesian model averaging is employed across a mix of fully identified and partially identified models.
- Published
- 2020
29. Bayesian adjustment for preferential testing in estimating the COVID-19 infection fatality rate
- Author
-
Campbell, Harlan, de Valpine, Perry, Maxwell, Lauren, de Jong, Valentijn MT, Debray, Thomas, Jänisch, Thomas, and Gustafson, Paul
- Subjects
Statistics - Methodology ,Statistics - Applications - Abstract
A key challenge in estimating the infection fatality rate (IFR) -- and its relation with various factors of interest -- is determining the total number of cases. The total number of cases is not known because not everyone is tested, but also, more importantly, because tested individuals are not representative of the population at large. We refer to the phenomenon whereby infected individuals are more likely to be tested than non-infected individuals, as "preferential testing." An open question is whether or not it is possible to reliably estimate the IFR without any specific knowledge about the degree to which the data are biased by preferential testing. In this paper we take a partial identifiability approach, formulating clearly where deliberate prior assumptions can be made and presenting a Bayesian model which pools information from different samples. When the model is fit to European data obtained from seroprevalence studies and national official COVID-19 statistics, we estimate the overall COVID-19 IFR for Europe to be 0.53%, 95% C.I. = [0.39%, 0.69%]., Comment: 53 pages, 14 figures
- Published
- 2020
30. Conley's fundamental theorem for a class of hybrid systems
- Author
-
Kvalheim, Matthew D., Gustafson, Paul, and Koditschek, Daniel E.
- Subjects
Mathematics - Dynamical Systems ,Computer Science - Robotics ,Mathematics - General Topology ,34A38 (Primary) 37B20, 37B25, 37C70, 68T40, 93C30 (Secondary) - Abstract
We establish versions of Conley's (i) fundamental theorem and (ii) decomposition theorem for a broad class of hybrid dynamical systems. The hybrid version of (i) asserts that a globally-defined "hybrid complete Lyapunov function" exists for every hybrid system in this class. Motivated by mechanics and control settings where physical or engineered events cause abrupt changes in a system's governing dynamics, our results apply to a large class of Lagrangian hybrid systems (with impacts) studied extensively in the robotics literature. Viewed formally, these results generalize those of Conley and Franks for continuous-time and discrete-time dynamical systems, respectively, on metric spaces. However, we furnish specific examples illustrating how our statement of sufficient conditions represents merely an early step in the longer project of establishing what formal assumptions can and cannot endow hybrid systems models with the topologically well characterized partitions of limit behavior that make Conley's theory so valuable in those classical settings., Comment: Simplified exposition in Sec. 4; minor fixes
- Published
- 2020
31. Concentration of Benefit index: A threshold-free summary metric for quantifying the capacity of covariates to yield efficient treatment rules
- Author
-
Sadatsafavi, Mohsen, Mansournia, Mohammad Ali, and Gustafson, Paul
- Subjects
Statistics - Applications ,Statistics - Methodology - Abstract
When data on treatment assignment, outcomes, and covariates from a randomized trial are available, a question of interest is to what extent covariates can be used to optimize treatment decisions. Statistical hypothesis testing of covariate-by-treatment interaction is ill-suited for this purpose. The application of decision theory results in treatment rules that compare the expected benefit of treatment given the patient's covariates against a treatment threshold. However, determining treatment threshold is often context-specific, and any given threshold might seem arbitrary when the overall capacity towards predicting treatment benefit is of concern. We propose the Concentration of Benefit index (Cb), a threshold-free metric that quantifies the combined performance of covariates towards finding individuals who will benefit the most from treatment. The construct of the proposed index is comparing expected treatment outcomes with and without knowledge of covariates when one of a two randomly selected patients are to be treated. We show that the resulting index can also be expressed in terms of the integrated efficiency of individualized treatment decision over the entire range of treatment thresholds. We propose parametric and semi-parametric estimators, the latter being suitable for out-of-sample validation and correction for optimism. We used data from a clinical trial to demonstrate the calculations in a step-by-step fashion, and have provided the R code for implementation (https://github.com/msadatsafavi/txBenefit). The proposed index has intuitive and theoretically sound interpretation and can be estimated with relative ease for a wide class of regression models. Beyond the conceptual developments, various aspects of estimation and inference for such a metric need to be pursued in future research., Comment: This submission was intended to be an update of the previous work (arXiv:1901.05124) and not a new record. As such, the authors decided to withdraw this record and will update arXiv:1901.05124 with an identical copy
- Published
- 2020
- Full Text
- View/download PDF
32. Formal composition of hybrid systems
- Author
-
Culbertson, Jared, Gustafson, Paul, Koditschek, Daniel E., and Stiller, Peter F.
- Subjects
Mathematics - Category Theory ,Computer Science - Robotics ,Mathematics - Dynamical Systems ,18D05, 93C30, 37C70, 68T40 - Abstract
We develop a compositional framework for formal synthesis of hybrid systems using the language of category theory. More specifically, we provide mutually compatible tools for hierarchical, sequential, and independent parallel composition. In our framework, hierarchies of hybrid systems correspond to template-anchor pairs, which we model as spans of subdividing and embedding semiconjugacies. Hierarchical composition of template-anchor pairs corresponds to the composition of spans via pullback. To model sequential composition, we introduce "directed hybrid systems," each of which flows from an initial subsystem to a final subsystem in a Conley-theoretic sense. Sequential composition of directed systems is given by a pushout of graph embeddings, rewriting the continuous dynamics of the overlapping subsystem to prioritize the second directed system. Independent parallel composition corresponds to a categorical product with respect to semiconjugacy. To formalize the compatibility of these three types of composition, we construct a vertically cartesian double category of hybrid systems where the vertical morphisms are semiconjugacies, and the horizontal morphisms are directed hybrid systems., Comment: 49 pages; v2: added figures to accompany examples, minor corrections, more detail to various proofs
- Published
- 2019
33. Measurement Error in Meta-Analysis (MEMA)--A Bayesian Framework for Continuous Outcome Data Subject to Non-Differential Measurement Error
- Author
-
Campbell, Harlan, de Jong, Valentijn M. T., Maxwell, Lauren, Jaenisch, Thomas, Debray, Thomas P. A., and Gustafson, Paul
- Abstract
Ideally, a meta-analysis will summarize data from several unbiased studies. Here we look into the less than ideal situation in which contributing studies may be compromised by non-differential measurement error in the exposure variable. Specifically, we consider a meta-analysis for the association between a continuous outcome variable and one or more continuous exposure variables, where the associations may be quantified as regression coefficients of a linear regression model. A flexible Bayesian framework is developed which allows one to obtain appropriate point and interval estimates with varying degrees of prior knowledge about the magnitude of the measurement error. We also demonstrate how, if individual-participant data (IPD) are available, the Bayesian meta-analysis model can adjust for multiple participant-level covariates, these being measured with or without measurement error.
- Published
- 2021
- Full Text
- View/download PDF
34. Braid group representations from twisted tensor products of algebras
- Author
-
Gustafson, Paul, Kimball, Andrew, Rowell, Eric C., and Zhang, Qing
- Subjects
Mathematics - Quantum Algebra ,Mathematics - Representation Theory - Abstract
We unify and generalize several approaches to constructing braid group representations from finite groups, using iterated twisted tensor products. Our results hint at a relationship between the braidings on the $G$-gaugings of a pointed modular category $\mathcal{C}(A,Q)$ and that of $\mathcal{C}(A,Q)$ itself.
- Published
- 2019
35. Logistic Box-Cox Regression to Assess the Shape and Median Effect under Uncertainty about Model Specification
- Author
-
Xing, Li, Zhang, Xuekui, Burstyn, Igor, and Gustafson, Paul
- Subjects
Statistics - Methodology - Abstract
The shape of the relationship between a continuous exposure variable and a binary disease variable is often central to epidemiologic investigations. This paper investigates a number of issues surrounding inference and the shape of the relationship. Presuming that the relationship can be expressed in terms of regression coefficients and a shape parameter, we investigate how well the shape can be inferred in settings which might typify epidemiologic investigations and risk assessment. We also consider a suitable definition of the median effect of exposure, and investigate how precisely this can be inferred. This is done both in the case of using a model acknowledging uncertainty about the shape parameter and in the case of ignoring this uncertainty and using a two-step method, where in step one we transform the predictor and in step two we fit a simple linear model with transformed predictor. All these investigations require a family of exposure-disease relationships indexed by a shape parameter. For this purpose, we employ a family based on the Box-Cox transformation., Comment: 18 pages, 14 figures
- Published
- 2019
36. A threshold-free summary index for quantifying the capacity of covariates to yield efficient treatment rules
- Author
-
Sadatsafavi, Mohsen, Mansournia, Mohammad, and Gustafson, Paul
- Subjects
Statistics - Applications - Abstract
The focus of this paper is on quantifying the capacity of covariates in devising efficient treatment rules when data from a randomized trial are available. Conventional one-variable-at-a-time subgroup analysis based on statistical hypothesis testing of covariate-by-treatment interaction is ill-suited for this purpose. The application of decision theory results in treatment rules that compare the expected benefit of treatment given the patient's covariates against a treatment threshold. However, determining treatment threshold is often context-specific, and any given threshold might seem arbitrary at the reporting stages of a clinical trial. We propose a threshold-free metric that quantifies the capacity of a set of covariates towards finding individuals who will benefit the most from treatment. The construct of the proposed metric is comparing the expected outcomes with and without knowledge of covariates when one of a two randomly selected patients are to be treated. We show that the resulting index can also be expressed in terms of integrated treatment benefit as a function of covariates over the entire range of treatment thresholds. We also propose a semi-parametric estimation method suitable for out-of-sample validation and adjustment for optimism. We use data from a clinical trial of preventive antibiotic therapy for reducing exacerbation rate in Chronic Obstructive Pulmonary Disease to demonstrate the calculations in a step-by-step fashion. The proposed index has intuitive and theoretically sound interpretation and can be estimated with relative ease for a wide class of regression models. Beyond the conceptual developments presented in this work, various aspects of estimation and inference for such metrics need to be pursued in future research., Comment: 27 pages, 2 figures, 2 tables
- Published
- 2019
- Full Text
- View/download PDF
37. Integral Metaplectic Modular Categories
- Author
-
Deaton, Adam, Gustafson, Paul, Mavrakis, Leslie, Rowell, Eric C., Poltoratski, Sasha, Timmerman, Sydney, Warren, Benjamin, and Zhang, Qing
- Subjects
Mathematics - Quantum Algebra ,18D10 (Primary) 20F36, 57M27 (Secondary) - Abstract
A braided fusion category is said to have Property $\textbf{F}$ if the associated braid group representations factor over a finite group. We verify integral metaplectic modular categories have property $\textbf{F}$ by showing these categories are group theoretical. For the special case of integral categories $\mathcal{C}$ with the fusion rules of $SO(8)_2$ we determine the finite group $G$ for which $Rep(D^{\omega}G)$ is braided equivalent to $\mathcal{Z}(\mathcal{C})$. In addition, we determine the associated classical link invariant, an evaluation of the 2-variable Kauffman polynomial at a point., Comment: 10 pages
- Published
- 2019
38. Generalizability of Risk Stratification Algorithms for Exacerbations in COPD
- Author
-
Bansback, Nick, Bottorff, Joan L., Bryan, Stirling, Burns, Paloma, Carlsten, Chris, Conklin, Annalijn I., De Vera, Mary, Gershon, Andrea, Gupta, Samir, Gustafson, Paul, Harvard, Stephanie, Hoens, Alison M., Mokhtaran, Mehrshad, Johnson, Jim, Joshi, Phalgun, Leung, Janice, Lynd, Larry D., Metcalfe, Rebecca K., Michaux, Kristina D., Sadatsafavi, Mohsen, Simmers, Brian, Sin, Don D., Smith, Daniel, Struik, Laura, Vinay, Dhingra, Ho, Joseph Khoa, Safari, Abdollah, Adibi, Amin, and Johnson, Kate
- Published
- 2023
- Full Text
- View/download PDF
39. What is the ideal time to begin tapering opioid agonist treatment? A protocol for a retrospective population-based comparative effectiveness study in British Columbia, Canada
- Author
-
Yan, Ruyu, primary, Kurz, Megan, additional, Guerra-Alejos, B Carolina, additional, Min, Jeong Eun, additional, Bach, Paxton, additional, Greenland, Sander, additional, Gustafson, Paul, additional, Karim, Ehsan, additional, Korthuis, P Todd, additional, Loughin, Tom, additional, McCandless, Lawrence, additional, Platt, Robert W, additional, Schnepel, Kevin, additional, Seaman, Shaun, additional, Socías, M Eugenia, additional, Wood, Evan, additional, Xie, Hui, additional, and Nosyk, Bohdan, additional
- Published
- 2024
- Full Text
- View/download PDF
40. Investigating Causal DIF via Propensity Score Methods
- Author
-
Liu, Yan, Zumbo, Bruno D., Gustafson, Paul, Huang, Yi, Kroc, Edward, and Wu, Amery D.
- Abstract
A variety of differential item functioning (DIF) methods have been proposed and used for ensuring that a test is fair to all test takers in a target population in the situations of, for example, a test being translated to other languages. However, once a method flags an item as DIF, it is difficult to conclude that the grouping variable (e.g., test language) is responsible for the DIF result because there may exist many confounding variables that lead to the DIF result. The present study aims to (i) demonstrate the application of propensity score methods in psychometric research on DIF for day-to-day researchers, and (ii) describe conditional logistic regression for matched data in a DIF context. Propensity score methods can help to achieve the comparability between different populations or groups with respect to participants' pre-test differences, which can assist in examining the validity of making a causal claim with regard to DIF.
- Published
- 2016
41. Metaplectic Categories, Gauging and Property F
- Author
-
Gustafson, Paul, Rowell, Eric, and Ruan, Yuze
- Subjects
Mathematics - Quantum Algebra - Abstract
$N$-Metaplectic categories, unitary modular categories with the same fusion rules as $SO(N)_2$, are prototypical examples of weakly integral modular categories. As such, a conjecture of the second author would imply that images of the braid group representations associated with metaplectic categories are finite groups, i.e. have property $F$. While it was recently shown that $SO(N)_2$ itself has property $F$, proving property $F$ for the more general class of metaplectic modular categories is an open problem. We verify this conjecture for $N$-metaplectic modular categories when $N$ is odd, exploiting their classification and enumeration to relate them to $SO(N)_2$. In another direction, we prove that when $N$ is divisible by $8$ the $N$-metaplectic categories have $3$ non-trivial bosons, and the boson condensation procedure applied to 2 of these bosons yields $\frac{N}{4}$-metaplectic categories. Otherwise stated: any $8k$-metaplectic category is a $\mathbb{Z}_2$-gauging of a $2k$-metaplectic category, so that the $N$ even metaplectic categories lie towers of $\mathbb{Z}_2$-gaugings commencing with $2k$- or $4k$-metaplectic categories with $k$ odd., Comment: version 3: condensed proofs
- Published
- 2018
42. What to make of non-inferiority and equivalence testing with a post-specified margin?
- Author
-
Campbell, Harlan and Gustafson, Paul
- Subjects
Statistics - Methodology - Abstract
In order to determine whether or not an effect is absent based on a statistical test, the recommended frequentist tool is the equivalence test. Typically, it is expected that an appropriate equivalence margin has been specified before any data are observed. Unfortunately, this can be a difficult task. If the margin is too small, then the test's power will be substantially reduced. If the margin is too large, any claims of equivalence will be meaningless. Moreover, it remains unclear how defining the margin afterwards will bias one's results. In this short article, we consider a series of hypothetical scenarios in which the margin is defined post-hoc or is otherwise considered controversial. We also review a number of relevant, potentially problematic actual studies from clinical trials research, with the aim of motivating a critical discussion as to what is acceptable and desirable in the reporting and interpretation of equivalence tests.
- Published
- 2018
43. The world of research has gone berserk: modeling the consequences of requiring 'greater statistical stringency' for scientific publication
- Author
-
Campbell, Harlan and Gustafson, Paul
- Subjects
Statistics - Methodology ,Statistics - Applications - Abstract
In response to growing concern about the reliability and reproducibility of published science, researchers have proposed adopting measures of greater statistical stringency, including suggestions to require larger sample sizes and to lower the highly criticized p<0.05 significance threshold. While pros and cons are vigorously debated, there has been little to no modeling of how adopting these measures might affect what type of science is published. In this paper, we develop a novel optimality model that, given current incentives to publish, predicts a researcher's most rational use of resources in terms of the number of studies to undertake, the statistical power to devote to each study, and the desirable pre-study odds to pursue. We then develop a methodology that allows one to estimate the reliability of published research by considering a distribution of preferred research strategies. Using this approach, we investigate the merits of adopting measures of `greater statistical stringency' with the goal of informing the ongoing debate.
- Published
- 2018
44. Dimension as a quantum statistic and the classification of metaplectic categories
- Author
-
Bruillard, Paul, Gustafson, Paul, Plavnik, Julia Yael, and Rowell, Eric Carson
- Subjects
Mathematics - Quantum Algebra - Abstract
We discuss several useful interpretations of the categorical dimension of objects in a braided fusion category, as well as some conjectures demonstrating the value of quantum dimension as a quantum statistic for detecting certain behaviors of anyons in topological phases of matter. From this discussion we find that objects in braided fusion categories with integral squared dimension have distinctive properties. A large and interesting class of non-integral modular categories such that every simple object has integral squared-dimensions are the metaplectic categories that have the same fusion rules as $SO(N)_2$ for some $N$. We describe and complete their classification and enumeration, by recognizing them as $\mathbb{Z}_2$-gaugings of cyclic modular categories (i.e. metric groups). We prove that any modular category of dimension $2^km$ with $m$ square-free and $k\leq 4$, satisfying some additional assumptions, is a metaplectic category. This illustrates anew that dimension can, in some circumstances, determine a surprising amount of the category's structure., Comment: v2: Title and sections restructured to clarify the relationship between the paper's two major topics. Some previously sketched results are now structured as propositions and proofs. Many previously referenced-away definitions have been made explicit
- Published
- 2017
45. Conditional Equivalence Testing: an alternative remedy for publication bias
- Author
-
Campbell, Harlan and Gustafson, Paul
- Subjects
Statistics - Methodology - Abstract
We introduce a publication policy that incorporates conditional equivalence testing (CET), a two-stage testing scheme in which standard NHST is followed conditionally by testing for equivalence. The idea of CET is carefully considered as it has the potential to address recent concerns about reproducibility and the limited publication of null results. In this paper we detail the implementation of CET, investigate similarities with a Bayesian testing scheme, and outline the basis for how a scientific journal could proceed to reduce publication bias while remaining relevant., Comment: 37 pages, 9 figures
- Published
- 2017
- Full Text
- View/download PDF
46. Systematic Review Reveals Lack of Causal Methodology Applied to Pooled Longitudinal Observational Infectious Disease Studies
- Author
-
Hufstedler, Heather, Rahman, Sabahat, Danzer, Alexander M., Goymann, Hannah, de Jong, Valentijn M.T., Campbell, Harlan, Gustafson, Paul, Debray, Thomas P.A., Jaenisch, Thomas, Maxwell, Lauren, Matthay, Ellicott C., and Bärnighausen, Till
- Published
- 2022
- Full Text
- View/download PDF
47. Is the cholesterol-perfluoroalkyl substance association confounded by dietary fiber intake?: a Bayesian analysis of NHANES data with adjustment for measurement error in fiber intake
- Author
-
Linakis, Matthew W, Gustafson, Paul, Allen, Bruce C, Bachand, Annette M, Van Landingham, Cynthia, Keast, Debra R, and Longnecker, Matthew P
- Published
- 2022
- Full Text
- View/download PDF
48. rtestim: Time-varying reproduction number estimation with trend filtering.
- Author
-
Liu, Jiaping, Cai, Zhenglun, Gustafson, Paul, and McDonald, Daniel J.
- Subjects
NEWTON-Raphson method ,INFECTIOUS disease transmission ,COMMUNICABLE diseases ,ACQUISITION of data ,EPIDEMIOLOGISTS - Abstract
To understand the transmissibility and spread of infectious diseases, epidemiologists turn to estimates of the instantaneous reproduction number. While many estimation approaches exist, their utility may be limited. Challenges of surveillance data collection, model assumptions that are unverifiable with data alone, and computationally inefficient frameworks are critical limitations for many existing approaches. We propose a discrete spline-based approach that solves a convex optimization problem—Poisson trend filtering—using the proximal Newton method. It produces a locally adaptive estimator for instantaneous reproduction number estimation with heterogeneous smoothness. Our methodology remains accurate even under some process misspecifications and is computationally efficient, even for large-scale data. The implementation is easily accessible in a lightweight R package rtestim. Author summary: Instantaneous reproduction number estimation presents many challenges due to data collection, modelling assumptions, and computational burden. Our motivation is to develop a model that produces accurate estimates, is robust to model misspecification, is straightforward to use, and is computationally efficient, even for large counts and long time periods. We propose a convex optimization model with an ℓ
1 trend filtering penalty. It couples accurate estimation of the instantaneous reproduction number with desired smoothness. We solve the optimization using the proximal Newton method, which converges rapidly and is numerically stable. Our software, conveniently available in the R package rtestim, can produce estimates in seconds for incidence sequences with hundreds of observations. These estimates are produced for a sequence of tuning parameters and can be selected using a built-in cross validation procedure. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
49. Finiteness of Mapping Class Group Representations from Twisted Dijkgraaf-Witten Theory
- Author
-
Gustafson, Paul
- Subjects
Mathematics - Quantum Algebra ,Mathematics - Geometric Topology ,18D10, 20F36, 57M99 - Abstract
We show that any twisted Dijkgraaf-Witten representation of a mapping class group of an orientable, compact surface with boundary has finite image. This generalizes work of Etingof, Rowell and Witherspoon showing that the braid group images are finite. In particular, our result answers their question regarding finiteness of images of arbitrary mapping class group representations in the affirmative. Our approach is to translate the problem into manipulation of colored graphs embedded in the given surface. To do this translation, we use the fact that any twisted Dijkgraaf-Witten representation associated to a finite group $G$ and 3-cocycle $\omega$ is isomorphic to a Turaev-Viro-Barrett-Westbury (TVBW) representation associated to the spherical fusion category $\text{Vec}_G^\omega$ of twisted $G$-graded vector spaces. As shown by Kirillov, the representation space for this TVBW representation is canonically isomorphic to a vector space spanned by $\text{Vec}_G^\omega$-colored graphs embedded in the surface. By analyzing the action of the Birman generators on a finite spanning set of colored graphs, we find that the mapping class group acts by permutations on a slightly larger finite spanning set. This implies that the representation has finite image., Comment: Added more detail to the proof of the main theorem and to various definitions. Fixed definition of composition map. Clarified unstrictified colored graph construction
- Published
- 2016
50. A Bayesian mixture of experts approach to covariate misclassification
- Author
-
XIA, Michelle, HAHN, P. Richard, and GUSTAFSON, Paul
- Published
- 2020
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.