41 results on '"Holbrook, Andrew J."'
Search Results
2. Scaling Hawkes processes to one million COVID-19 cases
- Author
-
Ko, Seyoon, Suchard, Marc A., and Holbrook, Andrew J.
- Subjects
Statistics - Computation ,Statistics - Applications - Abstract
Hawkes stochastic point process models have emerged as valuable statistical tools for analyzing viral contagion. The spatiotemporal Hawkes process characterizes the speeds at which viruses spread within human populations. Unfortunately, likelihood-based inference using these models requires $O(N^2)$ floating-point operations, for $N$ the number of observed cases. Recent work responds to the Hawkes likelihood's computational burden by developing efficient graphics processing unit (GPU)-based routines that enable Bayesian analysis of tens-of-thousands of observations. We build on this work and develop a high-performance computing (HPC) strategy that divides 30 Markov chains between 4 GPU nodes, each of which uses multiple GPUs to accelerate its chain's likelihood computations. We use this framework to apply two spatiotemporal Hawkes models to the analysis of one million COVID-19 cases in the United States between March 2020 and June 2023. In addition to brute-force HPC, we advocate for two simple strategies as scalable alternatives to successful approaches proposed for small data settings. First, we use known county-specific population densities to build a spatially varying triggering kernel in a manner that avoids computationally costly nearest neighbors search. Second, we use a cut-posterior inference routine that accounts for infections' spatial location uncertainty by iteratively sampling latent locations uniformly within their respective counties of occurrence, thereby avoiding full-blown latent variable inference for 1,000,000 infection locations.
- Published
- 2024
3. Random-Effects Substitution Models for Phylogenetics via Scalable Gradient Approximations
- Author
-
Magee, Andrew F, Holbrook, Andrew J, Pekar, Jonathan E, Caviedes-Solis, Itzue W, Matsen Iv, Fredrick A, Baele, Guy, Wertheim, Joel O, Ji, Xiang, Lemey, Philippe, and Suchard, Marc A
- Subjects
Biological Sciences ,Ecology ,Evolutionary Biology ,Genetics ,Emerging Infectious Diseases ,Infectious Diseases ,Generic health relevance ,Phylogeny ,Classification ,SARS-CoV-2 ,Influenza A Virus ,H3N2 Subtype ,Models ,Genetic ,Markov Chains ,Bayes Theorem ,Bayesian inference ,Hamiltonian Monte Carlo ,phylogeography ,Evolutionary biology - Abstract
Phylogenetic and discrete-trait evolutionary inference depend heavily on an appropriate characterization of the underlying character substitution process. In this paper, we present random-effects substitution models that extend common continuous-time Markov chain models into a richer class of processes capable of capturing a wider variety of substitution dynamics. As these random-effects substitution models often require many more parameters than their usual counterparts, inference can be both statistically and computationally challenging. Thus, we also propose an efficient approach to compute an approximation to the gradient of the data likelihood with respect to all unknown substitution model parameters. We demonstrate that this approximate gradient enables scaling of sampling-based inference, namely Bayesian inference via Hamiltonian Monte Carlo, under random-effects substitution models across large trees and state-spaces. Applied to a dataset of 583 SARS-CoV-2 sequences, an HKY model with random-effects shows strong signals of nonreversibility in the substitution process, and posterior predictive model checks clearly show that it is a more adequate model than a reversible model. When analyzing the pattern of phylogeographic spread of 1441 influenza A virus (H3N2) sequences between 14 regions, a random-effects phylogeographic substitution model infers that air travel volume adequately predicts almost all dispersal rates. A random-effects state-dependent substitution model reveals no evidence for an effect of arboreality on the swimming mode in the tree frog subfamily Hylinae. Simulations reveal that random-effects substitution models can accommodate both negligible and radical departures from the underlying base substitution model. We show that our gradient-based inference approach is over an order of magnitude more time efficient than conventional approaches.
- Published
- 2024
4. Sparse Bayesian multidimensional scaling(s)
- Author
-
Sheth, Ami, Smith, Aaron, and Holbrook, Andrew J.
- Subjects
Statistics - Methodology ,Statistics - Computation - Abstract
Bayesian multidimensional scaling (BMDS) is a probabilistic dimension reduction tool that allows one to model and visualize data consisting of dissimilarities between pairs of objects. Although BMDS has proven useful within, e.g., Bayesian phylogenetic inference, its likelihood and gradient calculations require a burdensome order of $N^2$ floating-point operations, where $N$ is the number of data points. Thus, BMDS becomes impractical as $N$ grows large. We propose and compare two sparse versions of BMDS (sBMDS) that apply log-likelihood and gradient computations to subsets of the observed dissimilarity matrix data. Landmark sBMDS (L-sBMDS) extracts columns, while banded sBMDS (B-sBMDS) extracts diagonals of the data. These sparse variants let one specify a time complexity between $N^2$ and $N$. Under simplified settings, we prove posterior consistency for subsampled distance matrices. Through simulations, we examine the accuracy and computational efficiency across all models using both the Metropolis-Hastings and Hamiltonian Monte Carlo algorithms. We observe approximately 3-fold, 10-fold and 40-fold speedups with negligible loss of accuracy, when applying the sBMDS likelihoods and gradients to 500, 1,000 and 5,000 data points with 50 bands (landmarks); these speedups only increase with the size of data considered. Finally, we apply the sBMDS variants to the phylogeographic modeling of multiple influenza subtypes to better understand how these strains spread through global air transportation networks.
- Published
- 2024
5. More Quantum Speedups for Multiproposal MCMC
- Author
-
Lin, Chin-Yi, Chen, Kuo-Chin, Lemey, Philippe, Suchard, Marc A., Holbrook, Andrew J., and Hsieh, Min-Hsiu
- Subjects
Quantum Physics - Abstract
Multiproposal Markov chain Monte Carlo (MCMC) algorithms choose from multiple proposals at each iteration in order to sample from challenging target distributions more efficiently. Recent work demonstrates the possibility of quadratic quantum speedups for one such multiproposal MCMC algorithm. Using $P$ proposals, this quantum parallel MCMC QPMCMC algorithm requires only $\mathcal{O}(\sqrt{P})$ target evaluations at each step. Here, we present a fast new quantum multiproposal MCMC strategy, QPMCMC2, that only requires $\mathcal{O}(1)$ target evaluations and $\mathcal{O}(\log P)$ qubits. Unlike its slower predecessor, the QPMCMC2 Markov kernel (1) maintains detailed balance exactly and (2) is fully explicit for a large class of graphical models. We demonstrate this flexibility by applying QPMCMC2 to novel Ising-type models built on bacterial evolutionary networks and obtain significant speedups for Bayesian ancestral trait reconstruction for 248 observed salmonella bacteria.
- Published
- 2023
6. ANTsX neuroimaging-derived structural phenotypes of UK Biobank
- Author
-
Tustison, Nicholas J., Yassa, Michael A., Rizvi, Batool, Cook, Philip A., Holbrook, Andrew J., Sathishkumar, Mithra T., Tustison, Mia G., Gee, James C., Stone, James R., and Avants, Brian B.
- Published
- 2024
- Full Text
- View/download PDF
7. On the surprising effectiveness of a simple matrix exponential derivative approximation, with application to global SARS-CoV-2
- Author
-
Didier, Gustavo, Glatt-Holtz, Nathan E., Holbrook, Andrew J., Magee, Andrew F., and Suchard, Marc A.
- Subjects
Statistics - Computation ,Mathematics - Probability - Abstract
The continuous-time Markov chain (CTMC) is the mathematical workhorse of evolutionary biology. Learning CTMC model parameters using modern, gradient-based methods requires the derivative of the matrix exponential evaluated at the CTMC's infinitesimal generator (rate) matrix. Motivated by the derivative's extreme computational complexity as a function of state space cardinality, recent work demonstrates the surprising effectiveness of a naive, first-order approximation for a host of problems in computational biology. In response to this empirical success, we obtain rigorous deterministic and probabilistic bounds for the error accrued by the naive approximation and establish a "blessing of dimensionality" result that is universal for a large class of rate matrices with random entries. Finally, we apply the first-order approximation within surrogate-trajectory Hamiltonian Monte Carlo for the analysis of the early spread of SARS-CoV-2 across 44 geographic regions that comprise a state space of unprecedented dimensionality for unstructured (flexible) CTMC models within evolutionary biology., Comment: To appear in the Proceedings of the National Academy of Sciences
- Published
- 2023
8. Random-effects substitution models for phylogenetics via scalable gradient approximations
- Author
-
Magee, Andrew F., Holbrook, Andrew J., Pekar, Jonathan E., Caviedes-Solis, Itzue W., Matsen IV, Fredrick A., Baele, Guy, Wertheim, Joel O., Ji, Xiang, Lemey, Philippe, and Suchard, Marc A.
- Subjects
Quantitative Biology - Populations and Evolution ,Statistics - Computation - Abstract
Phylogenetic and discrete-trait evolutionary inference depend heavily on an appropriate characterization of the underlying character substitution process. In this paper, we present random-effects substitution models that extend common continuous-time Markov chain models into a richer class of processes capable of capturing a wider variety of substitution dynamics. As these random-effects substitution models often require many more parameters than their usual counterparts, inference can be both statistically and computationally challenging. Thus, we also propose an efficient approach to compute an approximation to the gradient of the data likelihood with respect to all unknown substitution model parameters. We demonstrate that this approximate gradient enables scaling of sampling-based inference, namely Bayesian inference via Hamiltonian Monte Carlo, under random-effects substitution models across large trees and state-spaces. Applied to a dataset of 583 SARS-CoV-2 sequences, an HKY model with random-effects shows strong signals of nonreversibility in the substitution process, and posterior predictive model checks clearly show that it is a more adequate model than a reversible model. When analyzing the pattern of phylogeographic spread of 1441 influenza A virus (H3N2) sequences between 14 regions, a random-effects phylogeographic substitution model infers that air travel volume adequately predicts almost all dispersal rates. A random-effects state-dependent substitution model reveals no evidence for an effect of arboreality on the swimming mode in the tree frog subfamily Hylinae. Simulations reveal that random-effects substitution models can accommodate both negligible and radical departures from the underlying base substitution model. We show that our gradient-based inference approach is over an order of magnitude more time efficient than conventional approaches.
- Published
- 2023
9. A Spatially Varying Hierarchical Random Effects Model for Longitudinal Macular Structural Data in Glaucoma Patients
- Author
-
Su, Erica, Weiss, Robert E., Nouri-Mahdavi, Kouros, and Holbrook, Andrew J.
- Subjects
Statistics - Applications - Abstract
We model longitudinal macular thickness measurements to monitor the course of glaucoma and prevent vision loss due to disease progression. The macular thickness varies over a 6$\times$6 grid of locations on the retina with additional variability arising from the imaging process at each visit. Currently, ophthalmologists estimate slopes using repeated simple linear regression for each subject and location. To estimate slopes more precisely, we develop a novel Bayesian hierarchical model for multiple subjects with spatially varying population-level and subject-level coefficients, borrowing information over subjects and measurement locations. We augment the model with visit effects to account for observed spatially correlated visit-specific errors. We model spatially varying (a) intercepts, (b) slopes, and (c) log residual standard deviations (SD) with multivariate Gaussian process priors with Mat\'ern cross-covariance functions. Each marginal process assumes an exponential kernel with its own SD and spatial correlation matrix. We develop our models for and apply them to data from the Advanced Glaucoma Progression Study. We show that including visit effects in the model reduces error in predicting future thickness measurements and greatly improves model fit.
- Published
- 2023
10. Parallel MCMC Algorithms: Theoretical Foundations, Algorithm Design, Case Studies
- Author
-
Glatt-Holtz, Nathan E., Holbrook, Andrew J., Krometis, Justin A., and Mondaini, Cecilia F.
- Subjects
Statistics - Computation ,Mathematics - Probability ,Mathematics - Statistics Theory ,62D05, 60J22, 65C05, 65Y05 - Abstract
Parallel Markov Chain Monte Carlo (pMCMC) algorithms generate clouds of proposals at each step to efficiently resolve a target probability distribution. We build a rigorous foundational framework for pMCMC algorithms that situates these methods within a unified 'extended phase space' measure-theoretic formalism. Drawing on our recent work that provides a comprehensive theory for reversible single proposal methods, we herein derive general criteria for multiproposal acceptance mechanisms which yield ergodic chains on general state spaces. Our formulation encompasses a variety of methodologies, including proposal cloud resampling and Hamiltonian methods, while providing a basis for the derivation of novel algorithms. In particular, we obtain a top-down picture for a class of methods arising from 'conditionally independent' proposal structures. As an immediate application, we identify several new algorithms including a multiproposal version of the popular preconditioned Crank-Nicolson (pCN) sampler suitable for high- and infinite-dimensional target measures which are absolutely continuous with respect to a Gaussian base measure. To supplement our theoretical results, we carry out a selection of numerical case studies that evaluate the efficacy of these novel algorithms. First, noting that the true potential of pMCMC algorithms arises from their natural parallelizability, we provide a limited parallelization study using TensorFlow and a graphics processing unit to scale pMCMC algorithms that leverage as many as 100k proposals at each step. Second, we use our multiproposal pCN algorithm (mpCN) to resolve a selection of problems in Bayesian statistical inversion for partial differential equations motivated by fluid measurement. These examples provide preliminary evidence of the efficacy of mpCN for high-dimensional target distributions featuring complex geometries and multimodal structures., Comment: Minor revisions from previous version
- Published
- 2022
11. Computational Statistics and Data Science in the Twenty-first Century
- Author
-
Holbrook, Andrew J., Nishimura, Akihiko, Ji, Xiang, and Suchard, Marc A.
- Subjects
Statistics - Other Statistics - Abstract
Data science has arrived, and computational statistics is its engine. As the scale and complexity of scientific and industrial data grow, the discipline of computational statistics assumes an increasingly central role among the statistical sciences. An explosion in the range of real-world applications means the development of more and more specialized computational methods, but five Core Challenges remain. We provide a high-level introduction to computational statistics by focusing on its central challenges, present recent model-specific advances and preach the ever-increasing role of non-sequential computational paradigms such as multi-core, many-core and quantum computing. Data science is bringing major changes to computational statistics, and these changes will shape the trajectory of the discipline in the 21st century.
- Published
- 2022
12. Synthesizing longitudinal cortical thickness estimates with a flexible and hierarchical multivariate measurement-error model
- Author
-
Birchfield, Jesse W., Tustison, Nicholas J., and Holbrook, Andrew J.
- Subjects
Statistics - Applications - Abstract
MRI-based entorhinal cortical thickness (eCT) measurements predict cognitive decline in Alzheimer's disease (AD) with low cost and minimal invasiveness. Two prominent imaging paradigms, FreeSurfer (FS) and Advanced Normalization Tools (ANTs), feature multiple pipelines for extracting region-specific eCT measurements from raw MRI, but the sheer complexity of these pipelines makes it difficult to choose between pipelines, compare results between pipelines, and characterize uncertainty in pipeline estimates. Worse yet, the EC is particularly difficult to image, leading to variations in thickness estimates between pipelines that overwhelm physiologicl variations predictive of AD. We examine the eCT outputs of seven different pipelines on MRIs from the Alzheimer's Disease Neuroimaging Initiative. Because of both theoretical and practical limitations, we have no gold standard by which to evaluate them. Instead, we use a Bayesian hierarchical model to combine the estimates. The resulting posterior distribution yields high-probability idealized eCT values that account for inherent uncertainty through a flexible multivariate error model that supports different constant offsets, standard deviations, tailedness, and correlation structures between pipelines. Our hierarchical model directly relates idealized eCTs to clinical outcomes in a way that propagates eCT estimation uncertainty to clinical estimates while accounting for longitudinal structure in the data. Surprisingly, even though it incorporates greater uncertainty in the predictor and regularization provided by the prior, the combined model reveals a stronger association between eCT and cognitive capacity than do nonhierarchical models based on data from single pipelines alone., Comment: 23 pages, 11 figures
- Published
- 2022
13. Accelerating Bayesian inference of dependency between complex biological traits
- Author
-
Zhang, Zhenyu, Nishimura, Akihiko, Trovão, Nídia S., Cherry, Joshua L., Holbrook, Andrew J., Ji, Xiang, Lemey, Philippe, and Suchard, Marc A.
- Subjects
Statistics - Methodology ,Quantitative Biology - Populations and Evolution ,Statistics - Computation - Abstract
Inferring dependencies between complex biological traits while accounting for evolutionary relationships between specimens is of great scientific interest yet remains infeasible when trait and specimen counts grow large. The state-of-the-art approach uses a phylogenetic multivariate probit model to accommodate binary and continuous traits via a latent variable framework, and utilizes an efficient bouncy particle sampler (BPS) to tackle the computational bottleneck -- integrating many latent variables from a high-dimensional truncated normal distribution. This approach breaks down as the number of specimens grows and fails to reliably characterize conditional dependencies between traits. Here, we propose an inference pipeline for phylogenetic probit models that greatly outperforms BPS. The novelty lies in 1) a combination of the recent Zigzag Hamiltonian Monte Carlo (Zigzag-HMC) with linear-time gradient evaluations and 2) a joint sampling scheme for highly correlated latent variables and correlation matrix elements. In an application exploring HIV-1 evolution from 535 viruses, the inference requires joint sampling from an 11,235-dimensional truncated normal and a 24-dimensional covariance matrix. Our method yields a 5-fold speedup compared to BPS and makes it possible to learn partial correlations between candidate viral mutations and virulence. Computational speedup now enables us to tackle even larger problems: we study the evolution of influenza H1N1 glycosylations on around 900 viruses. For broader applicability, we extend the phylogenetic probit model to incorporate categorical traits, and demonstrate its use to study Aquilegia flower and pollinator co-evolution., Comment: 39 pages, 5 figures, 3 tables
- Published
- 2022
14. A quantum parallel Markov chain Monte Carlo
- Author
-
Holbrook, Andrew J.
- Subjects
Quantum Physics ,Statistics - Computation - Abstract
We propose a novel hybrid quantum computing strategy for parallel MCMC algorithms that generate multiple proposals at each step. This strategy makes the rate-limiting step within parallel MCMC amenable to quantum parallelization by using the Gumbel-max trick to turn the generalized accept-reject step into a discrete optimization problem. When combined with new insights from the parallel MCMC literature, such an approach allows us to embed target density evaluations within a well-known extension of Grover's quantum search algorithm. Letting $P$ denote the number of proposals in a single MCMC iteration, the combined strategy reduces the number of target evaluations required from $\mathcal{O}(P)$ to $\mathcal{O}(P^{1/2})$. In the following, we review the rudiments of quantum computing, quantum search and the Gumbel-max trick in order to elucidate their combination for as wide a readership as possible., Comment: To appear in JCGS
- Published
- 2021
15. Generating MCMC proposals by randomly rotating the regular simplex
- Author
-
Holbrook, Andrew J.
- Subjects
Statistics - Computation ,Statistics - Other Statistics - Abstract
We present the simplicial sampler, a class of parallel MCMC methods that generate and choose from multiple proposals at each iteration. The algorithm's multiproposal randomly rotates a simplex connected to the current Markov chain state in a way that inherently preserves symmetry between proposals. As a result, the simplicial sampler leads to a simplified acceptance step: it simply chooses from among the simplex nodes with probability proportional to their target density values. We also investigate a multivariate Gaussian-based symmetric multiproposal mechanism and prove that it also enjoys the same simplified acceptance step. This insight leads to significant theoretical and practical speedups. While both algorithms enjoy natural parallelizability, we show that conventional implementations are sufficient to confer efficiency gains across an array of dimensions and a number of target distributions., Comment: To appear in Journal of Multivariate Analysis. Code here: https://github.com/andrewjholbrook/simplicialSampler
- Published
- 2021
16. Principled, practical, flexible, fast: a new approach to phylogenetic factor analysis
- Author
-
Hassler, Gabriel W., Gallone, Brigida, Aristide, Leandro, Allen, William L., Tolkoff, Max R., Holbrook, Andrew J., Baele, Guy, Lemey, Philippe, and Suchard, Marc A.
- Subjects
Quantitative Biology - Populations and Evolution ,Statistics - Applications ,Statistics - Methodology - Abstract
Biological phenotypes are products of complex evolutionary processes in which selective forces influence multiple biological trait measurements in unknown ways. Phylogenetic factor analysis disentangles these relationships across the evolutionary history of a group of organisms. Scientists seeking to employ this modeling framework confront numerous modeling and implementation decisions, the details of which pose computational and replicability challenges. General and impactful community employment requires a data scientific analysis plan that balances flexibility, speed and ease of use, while minimizing model and algorithm tuning. Even in the presence of non-trivial phylogenetic model constraints, we show that one may analytically address latent factor uncertainty in a way that (a) aids model flexibility, (b) accelerates computation (by as much as 500-fold) and (c) decreases required tuning. We further present practical guidance on inference and modeling decisions as well as diagnosing and solving common problems in these analyses. We codify this analysis plan in an automated pipeline that distills the potentially overwhelming array of modeling decisions into a small handful of (typically binary) choices. We demonstrate the utility of these methods and analysis plan in four real-world problems of varying scales., Comment: 27 pages, 7 figures, 1 table
- Published
- 2021
17. From viral evolution to spatial contagion: a biologically modulated Hawkes model.
- Author
-
Holbrook, Andrew J, Ji, Xiang, and Suchard, Marc A
- Subjects
Biodefense ,Prevention ,Bioengineering ,Infectious Diseases ,Emerging Infectious Diseases ,Vaccine Related ,Genetics ,2.5 Research design and methodologies (aetiology) ,Aetiology ,Infection ,Good Health and Well Being ,Humans ,Bayes Theorem ,Phylogeny ,Hemorrhagic Fever ,Ebola ,Disease Outbreaks ,Genome ,Viral ,Mathematical Sciences ,Biological Sciences ,Information and Computing Sciences ,Bioinformatics - Abstract
SummaryMutations sometimes increase contagiousness for evolving pathogens. During an epidemic, scientists use viral genome data to infer a shared evolutionary history and connect this history to geographic spread. We propose a model that directly relates a pathogen's evolution to its spatial contagion dynamics-effectively combining the two epidemiological paradigms of phylogenetic inference and self-exciting process modeling-and apply this phylogenetic Hawkes process to a Bayesian analysis of 23 421 viral cases from the 2014 to 2016 Ebola outbreak in West Africa. The proposed model is able to detect individual viruses with significantly elevated rates of spatiotemporal propagation for a subset of 1610 samples that provide genome data. Finally, to facilitate model application in big data settings, we develop massively parallel implementations for the gradient and Hessian of the log-likelihood and apply our high-performance computing framework within an adaptively pre-conditioned Hamiltonian Monte Carlo routine.Supplementary informationSupplementary data are available at Bioinformatics online.
- Published
- 2022
18. Embracing Uncertainty in 'Small Data' Problems: Estimating Earthquakes from Historical Anecdotes
- Author
-
Krometis, Justin A., Ringer, Hayden, Whitehead, Jared P., Glatt-Holtz, Nathan E., Harris, Ronald A., and Holbrook, Andrew J.
- Subjects
Statistics - Applications ,Physics - Geophysics ,62P35, 86A22 - Abstract
Improving understanding of current seismic risk is often dependent on developing a more complete characterization of earthquakes that have occurred in the past, and in particular before the development of modern sensing equipment in the middle of the twentieth century. However, accounts of such events are typically anecdotal in nature, limiting efforts to model them to more heuristic approaches. To address this shortfall, we develop a framework based in Bayesian inference to provide a more rigorous methodology for estimating pre-instrumental earthquakes. By directly modeling accounts of resultant tsunamis via probability distributions, the framework allows practitioners to make principled estimates of key characteristics (e.g., magnitude and location) of historical earthquakes. To illustrate this idea, we apply the methodology to the estimation of an earthquake in Eastern Indonesia in the mid 19th century, the source of which is currently the subject of considerable debate in the geological community. The approach taken here gives evidence that even "small data" that is limited in scope and extremely uncertain can still be used to yield information on past seismic events. Moreover, sensitivity bounds indicate that the results obtained here are robust despite the inherent uncertainty in the observations.
- Published
- 2021
19. From viral evolution to spatial contagion: a biologically modulated Hawkes model
- Author
-
Holbrook, Andrew J., Ji, Xiang, and Suchard, Marc A.
- Subjects
Quantitative Biology - Populations and Evolution ,Statistics - Applications - Abstract
Mutations sometimes increase contagiousness for evolving pathogens. During an epidemic, scientists use viral genome data to infer a shared evolutionary history and connect this history to geographic spread. We propose a model that directly relates a pathogen's evolution to its spatial contagion dynamics -- effectively combining the two epidemiological paradigms of phylogenetic inference and self-exciting process modeling -- and apply this \emph{phylogenetic Hawkes process} to a Bayesian analysis of 23,422 viral cases from the 2014-2016 Ebola outbreak in West Africa. The proposed model is able to detect individual viruses with significantly elevated rates of spatiotemporal propagation for a subset of 1,610 samples that provide genome data. Finally, to facilitate model application in big data settings, we develop massively parallel implementations for the gradient and Hessian of the log-likelihood and apply our high performance computing framework within an adaptively preconditioned Hamiltonian Monte Carlo routine.
- Published
- 2021
20. Bayesian mitigation of spatial coarsening for a Hawkes model applied to gunfire, wildfire and viral contagion
- Author
-
Holbrook, Andrew J., Ji, Xiang, and Suchard, Marc A.
- Subjects
Statistics - Methodology - Abstract
Self-exciting spatiotemporal Hawkes processes have found increasing use in the study of large-scale public health threats ranging from gun violence and earthquakes to wildfires and viral contagion. Whereas many such applications feature locational uncertainty, i.e., the exact spatial positions of individual events are unknown, most Hawkes model analyses to date have ignored spatial coarsening present in the data. Three particular 21st century public health crises -- urban gun violence, rural wildfires and global viral spread -- present qualitatively and quantitatively varying uncertainty regimes that exhibit (a) different collective magnitudes of spatial coarsening, (b) uniform and mixed magnitude coarsening, (c) differently shaped uncertainty regions and -- less orthodox -- (d) locational data distributed within the `wrong' effective space. We explicitly model such uncertainties in a Bayesian manner and jointly infer unknown locations together with all parameters of a reasonably flexible Hawkes model, obtaining results that are practically and statistically distinct from those obtained while ignoring spatial coarsening. This work also features two different secondary contributions: first, to facilitate Bayesian inference of locations and background rate parameters, we make a subtle yet crucial change to an established kernel-based rate model; and second, to facilitate the same Bayesian inference at scale, we develop a massively parallel implementation of the model's log-likelihood gradient with respect to locations and thus avoid its quadratic computational cost in the context of Hamiltonian Monte Carlo. Our examples involve thousands of observations and allow us to demonstrate practicality at moderate scales., Comment: To appear in AOAS
- Published
- 2020
21. Scalable Bayesian inference for self-excitatory stochastic processes applied to big American gunfire data
- Author
-
Holbrook, Andrew J., Loeffler, Charles E., Flaxman, Seth R., and Suchard, Marc A.
- Subjects
Statistics - Applications - Abstract
The Hawkes process and its extensions effectively model self-excitatory phenomena including earthquakes, viral pandemics, financial transactions, neural spike trains and the spread of memes through social networks. The usefulness of these stochastic process models within a host of economic sectors and scientific disciplines is undercut by the processes' computational burden: complexity of likelihood evaluations grows quadratically in the number of observations for both the temporal and spatiotemporal Hawkes processes. We show that, with care, one may parallelize these calculations using both central and graphics processing unit implementations to achieve over 100-fold speedups over single-core processing. Using a simple adaptive Metropolis-Hastings scheme, we apply our high-performance computing framework to a Bayesian analysis of big gunshot data generated in Washington D.C. between the years of 2006 and 2019, thereby extending a past analysis of the same data from under 10,000 to over 85,000 observations. To encourage wide-spread use, we provide hpHawkes, an open-source R package, and discuss high-level implementation and program design for leveraging aspects of computational hardware that become necessary in a big data setting., Comment: Submitted to Statistics and Computing
- Published
- 2020
22. Massive Parallelization Boosts Big Bayesian Multidimensional Scaling
- Author
-
Holbrook, Andrew J, Lemey, Philippe, Baele, Guy, Dellicour, Simon, Brockmann, Dirk, Rambaut, Andrew, and Suchard, Marc A
- Subjects
Mathematical Sciences ,Statistics ,Infectious Diseases ,Influenza ,Pneumonia & Influenza ,Emerging Infectious Diseases ,Infection ,Bayesian phylogeography ,Graphics processing unit ,Hamiltonian Monte Carlo ,Massive parallelization ,Single-instruction ,multiple-data ,GPU ,SIMD ,stat.CO ,Econometrics ,Statistics & Probability - Abstract
Big Bayes is the computationally intensive co-application of big data and large, expressive Bayesian models for the analysis of complex phenomena in scientific inference and statistical learning. Standing as an example, Bayesian multidimensional scaling (MDS) can help scientists learn viral trajectories through space-time, but its computational burden prevents its wider use. Crucial MDS model calculations scale quadratically in the number of observations. We partially mitigate this limitation through massive parallelization using multi-core central processing units, instruction-level vectorization and graphics processing units (GPUs). Fitting the MDS model using Hamiltonian Monte Carlo, GPUs can deliver more than 100-fold speedups over serial calculations and thus extend Bayesian MDS to a big data setting. To illustrate, we employ Bayesian MDS to infer the rate at which different seasonal influenza virus subtypes use worldwide air traffic to spread around the globe. We examine 5392 viral sequences and their associated 14 million pairwise distances arising from the number of commercial airline seats per year between viral sampling locations. To adjust for shared evolutionary history of the viruses, we implement a phylogenetic extension to the MDS model and learn that subtype H3N2 spreads most effectively, consistent with its epidemic success relative to other seasonal influenza subtypes. Finally, we provide MassiveMDS, an open-source, stand-alone C++ library and rudimentary R package, and discuss program design and high-level implementation with an emphasis on important aspects of computing architecture that become relevant at scale.
- Published
- 2021
23. Scalable Bayesian inference for self-excitatory stochastic processes applied to big American gunfire data
- Author
-
Holbrook, Andrew J, Loeffler, Charles E, Flaxman, Seth R, and Suchard, Marc A
- Subjects
Applied Mathematics ,Mathematical Sciences ,Numerical and Computational Mathematics ,Statistics ,Networking and Information Technology R&D (NITRD) ,Massive parallelization ,GPU ,SIMD ,Spatiotemporal Hawkes process ,stat.AP ,Computation Theory and Mathematics ,Statistics & Probability ,Applied mathematics ,Numerical and computational mathematics - Abstract
The Hawkes process and its extensions effectively model self-excitatory phenomena including earthquakes, viral pandemics, financial transactions, neural spike trains and the spread of memes through social networks. The usefulness of these stochastic process models within a host of economic sectors and scientific disciplines is undercut by the processes' computational burden: complexity of likelihood evaluations grows quadratically in the number of observations for both the temporal and spatiotemporal Hawkes processes. We show that, with care, one may parallelize these calculations using both central and graphics processing unit implementations to achieve over 100-fold speedups over single-core processing. Using a simple adaptive Metropolis-Hastings scheme, we apply our high-performance computing framework to a Bayesian analysis of big gunshot data generated in Washington D.C. between the years of 2006 and 2019, thereby extending a past analysis of the same data from under 10,000 to over 85,000 observations. To encourage widespread use, we provide hpHawkes, an open-source R package, and discuss high-level implementation and program design for leveraging aspects of computational hardware that become necessary in a big data setting.
- Published
- 2021
24. Anterolateral entorhinal cortex thickness as a new biomarker for early detection of Alzheimer's disease
- Author
-
Holbrook, Andrew J, Tustison, Nicholas J, Marquez, Freddie, Roberts, Jared, Yassa, Michael A, Gillen, Daniel L, and Initiative§, for the Alzheimer's Disease Neuroimaging
- Subjects
Biological Psychology ,Psychology ,Acquired Cognitive Impairment ,Neurosciences ,Alzheimer's Disease including Alzheimer's Disease Related Dementias (AD/ADRD) ,Alzheimer's Disease ,Dementia ,Neurodegenerative ,Brain Disorders ,Aging ,Aetiology ,4.1 Discovery and preclinical testing of markers and technologies ,Detection ,screening and diagnosis ,2.1 Biological and endogenous factors ,Neurological ,ADNI-1 ,Alzheimer's disease ,anterolateral entorhinal cortex ,biomarker ,brain imaging ,Clinical Dementia Rating ,cortical thickness ,cerebrospinal fluid amyloid ,linear mixed-effects models ,memory ,mild cognitive impairment ,Mini-Mental State Exam ,posteromedial entorhinal cortex ,receiver operating characteristic ,Alzheimer's Disease Neuroimaging Initiative§ ,ADNI‐1 ,Mini‐Mental State Exam ,linear mixed‐effects models ,Genetics ,Biological psychology - Abstract
IntroductionLoss of entorhinal cortex (EC) layer II neurons represents the earliest Alzheimer's disease (AD) lesion in the brain. Research suggests differing functional roles between two EC subregions, the anterolateral EC (aLEC) and the posteromedial EC (pMEC).MethodsWe use joint label fusion to obtain aLEC and pMEC cortical thickness measurements from serial magnetic resonance imaging scans of 775 ADNI-1 participants (219 healthy; 380 mild cognitive impairment; 176 AD) and use linear mixed-effects models to analyze longitudinal associations among cortical thickness, disease status, and cognitive measures.ResultsGroup status is reliably predicted by aLEC thickness, which also exhibits greater associations with cognitive outcomes than does pMEC thickness. Change in aLEC thickness is also associated with cerebrospinal fluid amyloid and tau levels.DiscussionThinning of aLEC is a sensitive structural biomarker that changes over short durations in the course of AD and tracks disease severity-it is a strong candidate biomarker for detection of early AD.
- Published
- 2020
25. Generating MCMC proposals by randomly rotating the regular simplex
- Author
-
Holbrook, Andrew J.
- Published
- 2023
- Full Text
- View/download PDF
26. Longitudinal Mapping of Cortical Thickness Measurements: An Alzheimer’s Disease Neuroimaging Initiative-Based Evaluation Study
- Author
-
Tustison, Nicholas J, Holbrook, Andrew J, Avants, Brian B, Roberts, Jared M, Cook, Philip A, Reagh, Zachariah M, Duda, Jeffrey T, Stone, James R, Gillen, Daniel L, Yassa, Michael A, and Initiative, for the Alzheimer’s Disease Neuroimaging
- Subjects
Biological Psychology ,Psychology ,Brain Disorders ,Aging ,Alzheimer's Disease including Alzheimer's Disease Related Dementias (AD/ADRD) ,Neurosciences ,Dementia ,Acquired Cognitive Impairment ,Bioengineering ,Neurodegenerative ,Alzheimer's Disease ,Neurological ,Alzheimer Disease ,Biomarkers ,Brain ,Cross-Sectional Studies ,Disease Progression ,Female ,Humans ,Linear Models ,Longitudinal Studies ,Male ,Neuroimaging ,Advanced normalization tools ,FreeSurfer ,linear mixed effects models ,longitudinal processing ,Alzheimer’s Disease Neuroimaging Initiative ,Clinical Sciences ,Cognitive Sciences ,Neurology & Neurosurgery ,Clinical sciences ,Biological psychology - Abstract
Longitudinal studies of development and disease in the human brain have motivated the acquisition of large neuroimaging data sets and the concomitant development of robust methodological and statistical tools for quantifying neurostructural changes. Longitudinal-specific strategies for acquisition and processing have potentially significant benefits including more consistent estimates of intra-subject measurements while retaining predictive power. Using the first phase of the Alzheimer's Disease Neuroimaging Initiative (ADNI-1) data, comprising over 600 subjects with multiple time points from baseline to 36 months, we evaluate the utility of longitudinal FreeSurfer and Advanced Normalization Tools (ANTs) surrogate thickness values in the context of a linear mixed-effects (LME) modeling strategy. Specifically, we estimate the residual variability and between-subject variability associated with each processing stream as it is known from the statistical literature that minimizing the former while simultaneously maximizing the latter leads to greater scientific interpretability in terms of tighter confidence intervals in calculated mean trends, smaller prediction intervals, and narrower confidence intervals for determining cross-sectional effects. This strategy is evaluated over the entire cortex, as defined by the Desikan-Killiany-Tourville labeling protocol, where comparisons are made with the cross-sectional and longitudinal FreeSurfer processing streams. Subsequent linear mixed effects modeling for identifying diagnostic groupings within the ADNI cohort is provided as supporting evidence for the utility of the proposed ANTs longitudinal framework which provides unbiased structural neuroimage processing and competitive to superior power for longitudinal structural change detection.
- Published
- 2019
27. The ANTsX ecosystem for quantitative biological and medical imaging
- Author
-
Tustison, Nicholas J., Cook, Philip A., Holbrook, Andrew J., Johnson, Hans J., Muschelli, John, Devenyi, Gabriel A., Duda, Jeffrey T., Das, Sandhitsu R., Cullen, Nicholas C., Gillen, Daniel L., Yassa, Michael A., Stone, James R., Gee, James C., and Avants, Brian B.
- Published
- 2021
- Full Text
- View/download PDF
28. Accelerating Bayesian inference of dependency between mixed-type biological traits
- Author
-
Zhang, Zhenyu, primary, Nishimura, Akihiko, additional, Trovão, Nídia S., additional, Cherry, Joshua L., additional, Holbrook, Andrew J., additional, Ji, Xiang, additional, Lemey, Philippe, additional, and Suchard, Marc A., additional
- Published
- 2023
- Full Text
- View/download PDF
29. On the surprising effectiveness of a simple matrix exponential derivative approximation, with application to global SARS-CoV-2.
- Author
-
Didier, Gustavo, Glatt-Holtz, Nathan E., Holbrook, Andrew J., Magee, Andrew F., and Suchard, Marc A.
- Subjects
MATRIX exponential ,SARS-CoV-2 ,RANDOM matrices ,MARKOV processes ,LIFE science education ,BIOLOGY education - Abstract
The continuous-time Markov chain (CTMC) is the mathematical workhorse of evolutionary biology. Learning CTMC model parameters using modern, gradientbased methods requires the derivative of the matrix exponential evaluated at the CTMC's infinitesimal generator (rate) matrix. Motivated by the derivative's extreme computational complexity as a function of state space cardinality, recent work demonstrates the surprising effectiveness of a naive, first-order approximation for a host of problems in computational biology. In response to this empirical success, we obtain rigorous deterministic and probabilistic bounds for the error accrued by the naive approximation and establish a "blessing of dimensionality" result that is universal for a large class of rate matrices with random entries. Finally, we apply the first-order approximation within surrogate-trajectory Hamiltonian Monte Carlo for the analysis of the early spread of Severe acute respiratory syndrome coronavirus 2 (SARSCoV- 2) across 44 geographic regions that comprise a state space of unprecedented dimensionality for unstructured (flexible) CTMC models within evolutionary biology. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. A Quantum Parallel Markov Chain Monte Carlo
- Author
-
Holbrook, Andrew J., primary
- Published
- 2023
- Full Text
- View/download PDF
31. Bayesian Spatial Longitudinal Modeling for Local Rates of Glaucoma Progression
- Author
-
Su, Erica, Weiss, Robert E1, Holbrook, Andrew J, Su, Erica, Su, Erica, Weiss, Robert E1, Holbrook, Andrew J, and Su, Erica
- Abstract
Timely detection of glaucoma progression is imperative to identify eyes for treatment and prevent further loss of vision. Methods to diagnose and monitor progression include routine use of structural and functional tests at multiple locations across the macula, the central part of the retina. Visual field (VF) testing provides functional measures of sensitivity to light while optical coherence tomography imaging gives structural thickness measurements of macular layers. In current practice, physicians assess progression by modeling functional or structural changes over time using simple linear regression (SLR) for each subject-location combination separately. This dissertation motivates and develops Bayesian hierarchical spatial longitudinal models to analyze structural and functional data from multiple subjects, borrowing strength across subjects and locations, to better detect glaucoma progression and predict future observations for individual subjects.Chapter 1 gives an overview of the study objectives and summarizes the contributions of this dissertation. Chapter 2 presents a novel Bayesian hierarchical spatial longitudinal (HSL) model and compares its performance in estimating macular rates of structural change to the performances of SLR and a conditional autoregressive model. Notably in the simulation study, the HSL model is more than three times as efficient as SLR in estimating local rates of change. To more explicitly model the spatial correlation in intercepts, slopes, and residual standard deviations (SD), Chapter 3 proposes a Bayesian hierarchical model with spatially varying random coefficients and visit effects. A comparison of the model to several nested models lacking different model components demonstrates the benefit of incorporating spatially varying visit effects in improving model fit and reducing prediction error. Chapter 4 extends the spatially varying coefficients approach to model VF data. This model simultaneously accounts for censoring and
- Published
- 2023
32. ANTsX neuroimaging-derived structural phenotypes of UK Biobank
- Author
-
Tustison, Nicholas J., primary, Yassa, Michael A, additional, Rizvi, Batool, additional, Cook, Philip A, additional, Holbrook, Andrew J., additional, Sathishkumar, Mithra, additional, Tustison, Mia G, additional, Gee, James C., additional, Stone, James R., additional, and Avants, Brian B., additional
- Published
- 2023
- Full Text
- View/download PDF
33. Principled, practical, flexible, fast: A new approach to phylogenetic factor analysis
- Author
-
Hassler, Gabriel W., primary, Gallone, Brigida, additional, Aristide, Leandro, additional, Allen, William L., additional, Tolkoff, Max R., additional, Holbrook, Andrew J., additional, Baele, Guy, additional, Lemey, Philippe, additional, and Suchard, Marc A., additional
- Published
- 2022
- Full Text
- View/download PDF
34. Author response for 'Principled, practical, flexible, fast: A new approach to phylogenetic factor analysis'
- Author
-
null Hassler, Gabriel W., null Gallone, Brigida, null Aristide, Leandro, null Allen, William L., null Tolkoff, Max R., null Holbrook, Andrew J., null Baele, Guy, null Lemey, Philippe, and null Suchard, Marc A.
- Published
- 2022
- Full Text
- View/download PDF
35. Bayesian mitigation of spatial coarsening for a Hawkes model applied to gunfire, wildfire and viral contagion
- Author
-
Holbrook, Andrew J., Ji, Xiang, and Suchard, Marc A.
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,Statistics and Probability ,Modeling and Simulation ,Statistics, Probability and Uncertainty ,Statistics - Methodology ,Article - Abstract
Self-exciting spatiotemporal Hawkes processes have found increasing use in the study of large-scale public health threats ranging from gun violence and earthquakes to wildfires and viral contagion. Whereas many such applications feature locational uncertainty, i.e., the exact spatial positions of individual events are unknown, most Hawkes model analyses to date have ignored spatial coarsening present in the data. Three particular 21st century public health crises -- urban gun violence, rural wildfires and global viral spread -- present qualitatively and quantitatively varying uncertainty regimes that exhibit (a) different collective magnitudes of spatial coarsening, (b) uniform and mixed magnitude coarsening, (c) differently shaped uncertainty regions and -- less orthodox -- (d) locational data distributed within the `wrong' effective space. We explicitly model such uncertainties in a Bayesian manner and jointly infer unknown locations together with all parameters of a reasonably flexible Hawkes model, obtaining results that are practically and statistically distinct from those obtained while ignoring spatial coarsening. This work also features two different secondary contributions: first, to facilitate Bayesian inference of locations and background rate parameters, we make a subtle yet crucial change to an established kernel-based rate model; and second, to facilitate the same Bayesian inference at scale, we develop a massively parallel implementation of the model's log-likelihood gradient with respect to locations and thus avoid its quadratic computational cost in the context of Hamiltonian Monte Carlo. Our examples involve thousands of observations and allow us to demonstrate practicality at moderate scales., Comment: To appear in AOAS
- Published
- 2022
- Full Text
- View/download PDF
36. Computational Statistics and Data Science in the Twenty‐First Century
- Author
-
Holbrook, Andrew J., primary, Nishimura, Akihiko, additional, Ji, Xiang, additional, and Suchard, Marc A., additional
- Published
- 2021
- Full Text
- View/download PDF
37. Scalable Bayesian inference for self-excitatory stochastic processes applied to big American gunfire data
- Author
-
Holbrook, Andrew J, Loeffler, Charles E, Flaxman, Seth R, and Suchard, Marc A
- Subjects
FOS: Computer and information sciences ,Networking and Information Technology R&D (NITRD) ,Spatiotemporal Hawkes process ,Statistics & Probability ,Statistics ,GPU ,Computation Theory and Mathematics ,Applications (stat.AP) ,stat.AP ,Statistics - Applications ,SIMD ,Massive parallelization - Abstract
The Hawkes process and its extensions effectively model self-excitatory phenomena including earthquakes, viral pandemics, financial transactions, neural spike trains and the spread of memes through social networks. The usefulness of these stochastic process models within a host of economic sectors and scientific disciplines is undercut by the processes' computational burden: complexity of likelihood evaluations grows quadratically in the number of observations for both the temporal and spatiotemporal Hawkes processes. We show that, with care, one may parallelize these calculations using both central and graphics processing unit implementations to achieve over 100-fold speedups over single-core processing. Using a simple adaptive Metropolis-Hastings scheme, we apply our high-performance computing framework to a Bayesian analysis of big gunshot data generated in Washington D.C. between the years of 2006 and 2019, thereby extending a past analysis of the same data from under 10,000 to over 85,000 observations. To encourage wide-spread use, we provide hpHawkes, an open-source R package, and discuss high-level implementation and program design for leveraging aspects of computational hardware that become necessary in a big data setting., Submitted to Statistics and Computing
- Published
- 2020
38. Massive Parallelization Boosts Big Bayesian Multidimensional Scaling
- Author
-
Holbrook, Andrew J., primary, Lemey, Philippe, additional, Baele, Guy, additional, Dellicour, Simon, additional, Brockmann, Dirk, additional, Rambaut, Andrew, additional, and Suchard, Marc A., additional
- Published
- 2020
- Full Text
- View/download PDF
39. ANTsX neuroimaging-derived structural phenotypes of UK Biobank.
- Author
-
Tustison NJ, Yassa MA, Rizvi B, Cook PA, Holbrook AJ, Sathishkumar MT, Tustison MG, Gee JC, Stone JR, and Avants BB
- Abstract
UK Biobank is a large-scale epidemiological resource for investigating prospective correlations between various lifestyle, environmental, and genetic factors with health and disease progression. In addition to individual subject information obtained through surveys and physical examinations, a comprehensive neuroimaging battery consisting of multiple modalities provides imaging-derived phenotypes (IDPs) that can serve as biomarkers in neuroscience research. In this study, we augment the existing set of UK Biobank neuroimaging structural IDPs, obtained from well-established software libraries such as FSL and FreeSurfer, with related measurements acquired through the Advanced Normalization Tools Ecosystem. This includes previously established cortical and subcortical measurements defined, in part, based on the Desikan-Killiany-Tourville atlas. Also included are morphological measurements from two recent developments: medial temporal lobe parcellation of hippocampal and extra-hippocampal regions in addition to cerebellum parcellation and thickness based on the Schmahmann anatomical labeling. Through predictive modeling, we assess the clinical utility of these IDP measurements, individually and in combination, using commonly studied phenotypic correlates including age, fluid intelligence, numeric memory, and several other sociodemographic variables. The predictive accuracy of these IDP-based models, in terms of root-mean-squared-error or area-under-the-curve for continuous and categorical variables, respectively, provides comparative insights between software libraries as well as potential clinical interpretability. Results demonstrate varied performance between package-based IDP sets and their combination, emphasizing the need for careful consideration in their selection and utilization.
- Published
- 2023
- Full Text
- View/download PDF
40. Random-effects substitution models for phylogenetics via scalable gradient approximations.
- Author
-
Magee AF, Holbrook AJ, Pekar JE, Caviedes-Solis IW, Iv FAM, Baele G, Wertheim JO, Ji X, Lemey P, and Suchard MA
- Abstract
Phylogenetic and discrete-trait evolutionary inference depend heavily on an appropriate characterization of the underlying character substitution process. In this paper, we present random-effects substitution models that extend common continuous-time Markov chain models into a richer class of processes capable of capturing a wider variety of substitution dynamics. As these random-effects substitution models often require many more parameters than their usual counterparts, inference can be both statistically and computationally challenging. Thus, we also propose an efficient approach to compute an approximation to the gradient of the data likelihood with respect to all unknown substitution model parameters. We demonstrate that this approximate gradient enables scaling of sampling-based inference, namely Bayesian inference via Hamiltonian Monte Carlo, under random-effects substitution models across large trees and state-spaces. Applied to a dataset of 583 SARS-CoV-2 sequences, an HKY model with random-effects shows strong signals of nonreversibility in the substitution process, and posterior predictive model checks clearly show that it is a more adequate model than a reversible model. When analyzing the pattern of phylogeographic spread of 1441 influenza A virus (H3N2) sequences between 14 regions, a random-effects phylogeographic substitution model infers that air travel volume adequately predicts almost all dispersal rates. A random-effects state-dependent substitution model reveals no evidence for an effect of arboreality on the swimming mode in the tree frog subfamily Hylinae. Simulations reveal that random-effects substitution models can accommodate both negligible and radical departures from the underlying base substitution model. We show that our gradient-based inference approach is over an order of magnitude more time efficient than conventional approaches.
- Published
- 2023
41. Anterolateral entorhinal cortex thickness as a new biomarker for early detection of Alzheimer's disease.
- Author
-
Holbrook AJ, Tustison NJ, Marquez F, Roberts J, Yassa MA, and Gillen DL
- Abstract
Introduction: Loss of entorhinal cortex (EC) layer II neurons represents the earliest Alzheimer's disease (AD) lesion in the brain. Research suggests differing functional roles between two EC subregions, the anterolateral EC (aLEC) and the posteromedial EC (pMEC)., Methods: We use joint label fusion to obtain aLEC and pMEC cortical thickness measurements from serial magnetic resonance imaging scans of 775 ADNI-1 participants (219 healthy; 380 mild cognitive impairment; 176 AD) and use linear mixed-effects models to analyze longitudinal associations among cortical thickness, disease status, and cognitive measures., Results: Group status is reliably predicted by aLEC thickness, which also exhibits greater associations with cognitive outcomes than does pMEC thickness. Change in aLEC thickness is also associated with cerebrospinal fluid amyloid and tau levels., Discussion: Thinning of aLEC is a sensitive structural biomarker that changes over short durations in the course of AD and tracks disease severity-it is a strong candidate biomarker for detection of early AD., Competing Interests: The authors declare no conflicts of interest relevant to this article., (© 2020 The Authors. Alzheimer's & Dementia: Diagnosis, Assessment & Disease Monitoring published by Wiley Periodicals, Inc. on behalf of Alzheimer's Association.)
- Published
- 2020
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.