1,032 results on '"statistical inference"'
Search Results
2. Estimating Causal Effects of Education Interventions Using a Two-Rating Regression Discontinuity Design: Lessons from a Simulation Study
- Author
-
MDRC, Porter, Kristin E., Reardon, Sean F., Unlu, Fatih, Bloom, Howard S., and Robinson-Cimpian, Joseph P.
- Abstract
A valuable extension of the single-rating regression discontinuity design (RDD) is a multiple-rating RDD (MRRDD). To date, four main methods have been used to estimate average treatment effects at the multiple treatment frontiers of an MRRDD: the "surface" method, the "frontier" method, the "binding-score" method, and the "fuzzy instrumental variables" method. This paper uses a series of simulations to evaluate the relative performance of each of these four methods under a variety of different data-generating models. Focusing on a two-rating RDD (2RRDD), the methods are compared in terms of their bias, precision, and mean squared error when implemented as they most likely would be in practice--using optimal bandwidth selection. The lessons learned from the simulations to a real-world example that uses data from a study of an English learner reclassification policy are also applied. Overall, this paper makes valuable contributions to the literature on MRRDDs in that it makes concrete recommendations for choosing among MRRDD estimation methods, for implementing any chosen method using local linear regression, and for providing accurate statistical inferences. Appended are: (1) Computing Average Frontier Effects for Models 3 and 4 in Our Simulations; (2) Estimation in Theory: Using Full Information; (3) Deriving Correct Model Specifications for the Fuzzy IV Method; (4) Domain or Bandwidth Selection; and (5) Simulation Results When Using Local Linear Regression, Not Including Other Rating as a Covariate.
- Published
- 2014
3. What Works Clearinghouse Procedures and Standards Handbook, Version 3.0
- Author
-
What Works Clearinghouse (ED)
- Abstract
This "What Works Clearinghouse Procedures and Standards Handbook (Version 3.0)" provides a detailed description of the standards and procedures of the What Works Clearinghouse (WWC). The remaining chapters of this Handbook are organized to take the reader through the basic steps that the WWC uses to develop a review protocol, identify the relevant literature, assess research quality, and summarize evidence of effectiveness. Organizational procedures used by the WWC to ensure an independent, systematic, and objective review are described in the appendices. Table I.1 provides a summary of the remaining chapters and associated appendices. The main differences between this version of the procedures and standards and the previous version (Version 2.1) are in clarity, detail, and scope. The organization of the Handbook, as well as all text, was reviewed and modified to support clarity; additionally, examples have been added throughout. There is more detail on the specific procedures and standards used by the WWC, including how to deal with missing data, random assignment probabilities, and cluster-level designs. Finally, whereas the previous version focused almost exclusively on intervention reports, this version provides information on other key WWC products, which include practice guides, single study reviews, and quick reviews. As the WWC continues to refine processes, develop new standards, and create new products, the "What Works Clearinghouse Procedures and Standards Handbook" will be revised to reflect these changes. Appended are: (1) Staffing, Reviewer Certification, and Quality Assurance; (2) Policies for Searching and Prioritizing Studies for Review; (3) The WWC Study Review Process; (4) Pilot Regression Discontinuity Design Standards; (5) Pilot Single-Case Design Standards; (6) Magnitude of Findings for Randomized Controlled Trials and Quasi-Experimental Designs; and (7) Statistical Significance for Randomized Controlled Trials and Quasi Experimental Designs. [See ED503772 to view previous version of this guide.]
- Published
- 2014
4. Estimating the Impacts of Educational Interventions Using State Tests or Study-Administered Tests. NCEE 2012-4016
- Author
-
National Center for Education Evaluation and Regional Assistance (ED), Olsen, Robert B., Unlu, Fatih, Price, Cristofer, and Jaciw, Andrew P.
- Abstract
This report examines the differences in impact estimates and standard errors that arise when these are derived using state achievement tests only (as pre-tests and post-tests), study-administered tests only, or some combination of state- and study-administered tests. State tests may yield different evaluation results relative to a test that is selected, and administered, by the research team for several reasons. For instance, (1) because state tests vary in content and emphasis, they also can vary in their coverage of the types of knowledge and skills targeted by any given intervention. In contrast, a study-administered test will correspond to the intervention being evaluated. In addition to differences in alignment with treatment, state tests may yield divergent evaluation results due to differences in (2) the stakes associated with the test, (3) missing data, (4) the timing of the tests, (5) reliability or measurement error, and (6) alignment between pre-test and post-test. Olsen, Unlu, Jaciw, and Price (2011) discuss how these six factors may differ between state- and study-administered tests to influence the findings from an impact evaluation. Specifically, Olsen et al. use data from three single-state, small-scale evaluations of reading interventions that collected outcomes data using both study-administered and state achievement tests to examine this and other issues. The authors found that (1) impact estimates based on study-administered tests had smaller standard errors than impact estimates based on state tests, (2) impacts estimates from models with "mismatched" pre-tests (e.g., a state pre-test used in combination with a study-administered post-test) had larger standard errors than impact estimates from models with matched pre-tests, and (3) impact estimates from models that included a second pre-test covariate had smaller standard errors than impact estimates from models that included a single pre-test covariate. Study authors caution that their results may not generalize to evaluations conducted in other states, with different study-administered tests, or with other student samples. Appended are: (1) Description of the Three Experiments; (2) Scatter Plots of Student Test Scores; (3) Quartiles of the Test Score Distribution; (4) Estimates from Other Evaluations; (5) Estimates from the Full Sample; (6) Hypothesis Tests and Minimum Detectable Differences; (7) Conceptual Approach to Generating Correlated Residuals for the Parametric Bootstrap; (8) Results from Bootstrapping and Hypothesis Testing; (9) Differences in Sample Size Requirements; (10) Correlations between Scores on State and Study-Administered Tests; and (11) Estimates of Key Statistical Power Parameters. (Contains 37 tables, 3 figures and 45 footnotes.)
- Published
- 2011
5. Illustrating Sampling Distribution of a Statistic: Minitab Revisited
- Author
-
Johnson, H. Dean and Evans, Marc A.
- Abstract
Understanding the concept of the sampling distribution of a statistic is essential for the understanding of inferential procedures. Unfortunately, this topic proves to be a stumbling block for students in introductory statistics classes. In efforts to aid students in their understanding of this concept, alternatives to a lecture-based mode of instruction have been introduced in the literature with some of these approaches using in-class activities, simulations using statistical software, and web-based applets. In this article, the use of statistical software, for the purpose of illustrating sampling distributions, is revisited through the use of Minitab macros, an approach that has not been observed in the literature. The result is a user-friendly way for students in introductory statistics classes to explore the concept of the sampling distribution of a statistic. Although the focus of this article will be on the sampling distribution, the methods described here are applicable to instruction of other statistical concepts. including confidence intervals and power. (Contains 4 figures.)
- Published
- 2008
6. Effects of Test Administrator Characteristics on Achievement Test Scores
- Author
-
Schafer, William D., Papapolydorou, Maria, Rahman, Taslima, and Parker, Lori
- Abstract
Possible relationships between five test examiner characteristics (gender, race, tenure, experience as a test administrator, and experience as a test developer or scorer) and six student achievement scores (reading, writing, language usage, mathematics, science, and social studies) were studied at the school level in a statewide assessment. The school-level results were aggregated using meta-analysis to explore the plausibility of examiner variables as threats to test validity. Very few of the average correlations across schools were statistically significant, and for all of them, even for those that were statistically significant, confidence intervals for the correlations were extremely small at both ends. Significant heterogeneity of effect sizes was found for virtually all of the 60 analyses, suggesting that further exploration is needed. Some directions for further research are discussed. (Contains 10 figures and 10 tables.) [This research was carried out by the Maryland Assessment Research Center for Education Success.]
- Published
- 2005
7. Using Commonly Available Software for Conducting Bootstrap Analyses.
- Author
-
Fan, Xitao
- Abstract
Bootstrap analysis, both for nonparametric statistical inference and for describing sample results stability and replicability, has been gaining prominence among quantitative researchers in educational and psychological research. Procedurally, however, it is often quite a challenge for quantitative researchers to implement bootstrap analysis in their research because bootstrap analysis is typically not an automated program option in statistical software programs. This paper uses a few heuristic analytical examples to show how bootstrap analysis can be accomplished through the use of some commonly available statistical software programs. Until bootstrap analysis becomes an automated program option in standard statistical software programs (e.g., the Statistical Package for the Social Sciences or the Statistical Analysis System), quantitative researchers may have to make do with these or other creative approaches to accomplish bootstrap analysis in their research. (Contains 4 tables, 10 figures, and 37 references.) (Author/SLD)
- Published
- 2001
8. Paper Towels, Baseball, Puzzles, Nuts & Bolts & the TI-83Plus STAT Tests Menu.
- Author
-
Carruth, Barbara
- Abstract
This collection of activities is designed to show how graphing calculators can be used to explore statistics. The activities address such topics as data representation, distributions, and statistical tests. Teaching notes and calculator instructions are included as are blackline masters. (MM)
- Published
- 2001
9. Statistics for Policymakers or Everything You Wanted To Know about Statistics but Thought You Could Never Understand. Working Paper Series.
- Author
-
National Center for Education Statistics (ED), Washington, DC. and Ahmed, Susan
- Abstract
This working paper contains the overheads used in a seminar designed to introduce some basic concepts of statistics to nonstatisticians. The seminar has been presented on several occasions. The first part of the seminar, and the first set of overheads, deals with the essentials of statistics, including: (1) population, sample, and inference; (2) standard errors and confidence intervals; (3) statistical significance; (4) correlation and linear regression; and (5) graphics. The second part of the seminar, and the second group of overheads, concerns basic principles of research design and analysis, including operationalizing terms, types of bias, confounding, and aspects of validity and reliability. (Contains 91 overheads.) (SLD)
- Published
- 1997
10. Bayesian Analysis for Linearized Multi-Stage Models in Quantal Bioassay.
- Author
-
Kuo, Lynn and Cohen, Michael P.
- Abstract
Bayesian methods for estimating dose response curves in quantal bioassay are studied. A linearized multi-stage model is assumed for the shape of the curves. A Gibbs sampling approach with data augmentation is employed to compute the Bayes estimates. In addition, estimation of the "relative additional risk" and the "risk specific dose" is studied. Model selection based on conditional predictive ordinates from cross-validated data is developed. Contains 22 references and 8 tables.) (Author)
- Published
- 1993
11. Controversies around the Role of Statistical Tests in Experimental Research.
- Author
-
Batanero, Carmen
- Abstract
Describes the logic of statistical testing in the Fisher and Neyman-Pearson approaches. Reviews some common misinterpretations of basic concepts behind statistical tests. Analyzes the philosophical and psychological issues that can contribute to these misinterpretations. Suggests possible ways in which statistical education might contribute to the better understanding and application of statistical inference. (Contains 53 references.) (Author/ASK)
- Published
- 2000
12. The Hispanic Population in the United States: March 1988 (Advance Report).
- Author
-
Bureau of the Census (DOC), Suitland, MD. Population Div.
- Abstract
This advance report based on the March 1988 supplement to the Current Population Survey (CPS) presents preliminary data on selected demographic, social, and economic characteristics of the Hispanic population of the United States. The Hispanic civilian noninstitutional population in March 1988 totaled about 19.4 million, a 34% increase since 1980; the non-Hispanic population increased 7% in that period. About half of Hispanic growth resulted from net migration and half from natural increase. The proportions of Hispanics completing 4 years of high school or more, and completing 4 or more years of college reached 51% and 10%, respectively, both records. About 55% of Hispanics resided in California and Texas. Married couple families decreased by 1988 to 70% from 74% in 1982. The unemployment rate among Hispanics 16 and over was 8.5%, its lowest level since the survey of March 1983, shortly after the end of the last recession. The poverty rate was 25.8%, little changed since 1982. The origins of Hispanic Americans were 62% Mexican, 13% Puerto Rican, 5% Cuban, and 12% Central or South American. These subgroups varied considerably in educational attainment, family composition, employment, and median family income. The report includes five graphs and four tables of selected social and economic characteristics, by type of Hispanic origin and by year from 1982 to 1988. Appendices discuss CPS data source, estimation procedure, and reliability of estimates, and they contain facsimiles of March 1988 CPS questionnaires. (SV)
- Published
- 1988
13. The Hispanic Population in the United States: March 1985.
- Author
-
Bureau of the Census (DOC), Suitland, MD. Population Div.
- Abstract
This report based on the March 1985 supplement to the Current Population Survey presents demographic, social, and economic data on the Hispanic population in the United States, and focuses on social and economic changes between 1982 and 1985. The Hispanic civilian noninstitutional population in March 1985 totaled about 17 million, a 17% increase since 1980; the non-Hispanic population increased 4% in that period. Between 1982 and 1985: (1) the median age of Hispanics rose from 23.9 to 25.0 years; (2) the proportion of Hispanics 25 and over with at least a high school diploma increased from 45% to 48%; (3) the proportion of married couple families fell from 74% to 72%; (4) the unemployment rate for Hispanics 16 and over fell from 13.4% to 11.3%; and (5) there was no change in the real median income of Hispanics, while that of non-Hispanics rose 3.5%. The origins of Hispanic Americans were about 61% Mexican, 15% Puerto Rican, 6% Cuban, and 10% Central or South American. These subgroups varied considerably in educational attainment, family size, family composition, employment, and median family income. The report includes 3 graphs and 27 detailed tables of data on demographic, social, and economic characteristics. Appendices cover the methodology for development of independent post-census estimates (component estimation technique), summary tables of selected characteristics for the years 1982 to 1985, definitions, and data source and reliability. (SV)
- Published
- 1988
14. The Hispanic Population in the United States: March 1988.
- Author
-
Bureau of the Census (DOC), Suitland, MD. Population Div.
- Abstract
This report presents data from the March 1988 Current Population Survey on the demographic, social, and economic characteristics of the Hispanic population of the United States, including age, sex, marital status, educational attainment, school enrollment, fertility, voting and registration, employment status, family composition and size, income, and poverty status. Data on educational attainment show 11.7% of Hispanic males, age 25 and over, completing less than 5 years of school compared to 2.6% of the total population. Fifty-two percent of Hispanic males completed 4 years of high school or more compared to 76.4% of the total population. Only 12% of Hispanic males completed 4 years of college or more, while 24% of the total population attained this much education. Relative percentages for females were very similar. Tables also show years of school completed by age, sex, and type of Hispanic origin and current educational enrollment. For those of Hispanic origin of both sexes, ages 3 to 34 years, 47.8% were enrolled in school in October, 1986, compared to 48.2% for those of non-Hispanic origin. The report includes five charts and 36 tables. Appendices discuss definitions, data sources, and the accuracy of estimates. (DHP)
- Published
- 1989
15. Teaching Students Inferential Statistics: A 'Tail' of Three Distributions.
- Author
-
Wybraniec, John and Wilmoth, Janet
- Abstract
Discusses the extent to which the existing literature offers insights into effectively teaching statistical inference. Describes an in-class exercise that helps students understand inferential statistics. Explains that students learn more about concepts such as population distribution, sampling distributions, and standard error of the estimate. (CMK)
- Published
- 1999
16. German Tanks: A Problem in Estimation.
- Author
-
Flaspoher, David C. and Dinkheller, Ann L.
- Abstract
Presents a problem that describes a real-world situation from World War II that tries to estimate the number of German tanks, half-tracks, and aircraft, and a simulation of that problem in which the selection of a suitable estimate is less apparent. (ASK)
- Published
- 1999
17. What Is Normal, Anyway?
- Author
-
Calzada, Maria E. and Scariano, Stephen M.
- Abstract
The study of statistics can be enhanced by using real-world data and problems obtained from the Internet. Uses data on HIV infections in Asian and Pacific Islander females to teach both descriptive and inferential statistics. (ASK)
- Published
- 1999
18. Statistical Methods in Psychology Journals.
- Author
-
Willkinson, Leland
- Abstract
Proposes guidelines for revising the American Psychological Association (APA) publication manual or other APA materials to clarify the application of statistics in research reports. The guidelines are intended to induce authors and editors to recognize the thoughtless application of statistical methods. Contains 54 references. (SLD)
- Published
- 1999
19. Championship Tennis as a Probabilistic Modelling Context.
- Author
-
Galbraith, Peter
- Abstract
Suggests ways for using data from championship tennis as a means for exploring probabilistic models, especially binomial probability. Examples include the probability of winning a service point and the probability of winning a service game using data from tables and graphs. (AIM)
- Published
- 1996
20. Bounded Population Growth: A Curve Fitting Lesson.
- Author
-
Mathews, John H.
- Abstract
Presents two mathematical methods for fitting the logistic curve to population data supplied by the U.S. Census Bureau utilizing computer algebra software to carry out the computations and plot graphs. (JKK)
- Published
- 1992
21. Some Cautions Concerning Inferences about Proportions, Differences between Proportions, and Quotients of Proportions.
- Author
-
Knapp, Thomas R. and Tam, Hak P.
- Abstract
Examines potential problems in the use of inferential statistics for single population proportions, differences between two population proportions, and quotients of two population proportions. Discusses hypothesis testing versus interval estimation. Emphasizes the importance of selecting the appropriate formula for the standard error and questioning the assumption of independent observations. Suggests cautionary procedures for educational researchers. Contains 21 references. (Author/SV)
- Published
- 1997
22. Using Quality Control Activities to Develop Scientific and Mathematical Literacy.
- Author
-
Miller, L. Diane and Mitchell, Charles E.
- Abstract
Suggests activities using large samples to simulate quality control processes in industry (specifically, manufacturing camera film). A series of small-group activities investigate the relationship of sample size to percent defective in batch sampling. Efficiency, cost of sampling, company reputation, and public relations are business factors considered. (RC)
- Published
- 1995
23. Variational inference in dynamical systems
- Author
-
Ialongo, Alessandro Davide and Rasmussen, Carl Edward
- Subjects
Variational Inference ,Gaussian Processes ,State Space Models ,Dynamical Systems ,Machine Learning ,Bayesian Methods ,Statistical Inference ,Normalizing Flows ,Gaussian Process State Space Model - Abstract
Dynamical systems are a powerful formalism to analyse the world around us. Many datasets are sequential in nature, and can be described by a discrete time evolution law. We are interested in approaching the analysis of such datasets from a probabilistic perspective. We would like to maintain justified beliefs about quantities which, though useful in explaining the behaviour of a system, may not be observable, as well as about the system's evolution itself, especially in regimes we have not yet observed in our data. The framework of statistical inference gives us the tools to do so, yet, for many systems of interest, performing inference exactly is not computationally or analytically tractable. The contribution of this thesis, then, is twofold: first, we uncover two sources of bias in existing variational inference methods applied to dynamical systems in general, and state space models whose transition function is drawn from a Gaussian process (GPSSM) in particular. We show bias can derive from assuming posteriors in non-linear systems to be jointly Gaussian, and from assuming that we can sever the dependence between latent states and transition function in state space model posteriors. Second, we propose methods to address these issues, undoing the resulting biases. We do this without compromising on computational efficiency or on the ability to scale to larger datasets and higher dimensions, compared to the methods we rectify. One method, the Markov Autoregressive Flow (Markov AF) addresses the Gaussian assumption, by providing a more flexible class of posteriors, based on normalizing flows, which can be easily evaluated, sampled, and optimised. The other method, Variationally Coupled Dynamics and Trajectories (VCDT), tackles the factorisation assumption, leveraging sparse Gaussian processes and their variational representation to reintroduce dependence between latent states and the transition function at no extra computational cost. Since the objective of inference is to maintain calibrated beliefs, if we employed approximations which are significantly biased in non-linear, noisy systems, or when there is little data available, we would have failed in our objective, as those are precisely the regimes in which uncertainty quantification is all the more important. Hence we think it is essential, if we wish to act optimally on such beliefs, to uncover, and, if possible, to correct, all sources of systematic bias in our inference methods.
- Published
- 2022
- Full Text
- View/download PDF
24. Multiplicative latent force models
- Author
-
Tait, Daniel J., Worton, Bruce, and Aitken, Colin
- Subjects
006.3 ,latent force models ,hybrid model ,Gaussian process ,statistical inference ,modelling ,multiplicative latent force model ,human motion capture data - Abstract
Latent force models (LFM) are a class of flexible models of dynamic systems, combining a simple mechanistic model with the flexibility of an additive inhomogeneous Gaussian process (GP) forcing term. These hybrid models achieve the dual goal of being flexible enough to be broadly applied, even for complex dynamic systems where a full mechanistic model may be hard to motivate, but by also encoding relevant properties of dynamic systems they are better able to model the underlying dynamics and so demonstrate superior generalisation. In this thesis, we consider an extension of this framework which keeps the same general form, a linear ordinary di↵erential equation with time-varying behaviour arising from a set of smooth GPs, but now we allow for multiplicative interactions between the state variables and the GP terms. The result is a semi-parametric modelling framework that allows for the embedding of rich topological structure. Following a brief review of the latent force model, which we note is a particular case of the GP regression model, we introduce our extension with multiplicative interactions which we refer to as the multiplicative latent force model (MLFM). We demonstrate that this class of models allows for the possibility of strong geometric constraints on the pathwise trajectories. This will enable the modelling of systems for which the GP trajectories of the LFM are unsatisfactory. Unfortunately, and as a direct consequence of the strong geometric constraints we have introduced, it is no longer straightforward to carry out inference in these models; therefore the remainder of this thesis is primarily devoted to constructing two methods for carrying out approximate inference for this class of models. The first is referred to as the Bayesian adaptive gradient matching method, and the second is a novel construction based on the method of successive approximations; a theoretical construct used in the standard classical existence and uniqueness theorems for ODEs. After introducing these methods, we demonstrate their accuracy on simulated data, which also allows for an investigation into the regimes in which each of the respective methods can be expected to perform well. Finally, we demonstrate the utility of the MLFM on motion capture data and show that, by using the framework developed in this thesis to allow for the sharing of a smaller number of latent forces between distinct trajectories with specific geometric constraints, we can achieve superior predictive performance than by the modelling of a single trajectory.
- Published
- 2020
- Full Text
- View/download PDF
25. Molecular evolution of biological sequences
- Author
-
Vázquez García, Ignacio and Mustonen, Ville
- Subjects
572.8 ,statistical inference ,genome sequencing ,population genomics ,clonal evolution ,mutation ,selection ,microbial evolution ,cancer evolution - Abstract
Evolution is an ubiquitous feature of living systems. The genetic composition of a population changes in response to the primary evolutionary forces: mutation, selection and genetic drift. Organisms undergoing rapid adaptation acquire multiple mutations that are physically linked in the genome, so their fates are mutually dependent and selection only acts on these loci in their entirety. This aspect has been largely overlooked in the study of asexual or somatic evolution and plays a major role in the evolution of bacterial and viral infections and cancer. In this thesis, we put forward a theoretical description for a minimal model of evolutionary dynamics to identify driver mutations, which carry a large positive fitness effect, among passenger mutations that hitchhike on successful genomes. We examine the effect this mode of selection has on genomic patterns of variation to infer the location of driver mutations and estimate their selection coefficient from time series of mutation frequencies. We then present a probabilistic model to reconstruct genotypically distinct lineages in mixed cell populations from DNA sequencing. This method uses Hidden Markov Models for the deconvolution of genetically diverse populations and can be applied to clonal admixtures of genomes in any asexual population, from evolving pathogens to the somatic evolution of cancer. To understand the effects of selection on rapidly adapting populations, we constructed sequence ensembles in a recombinant library of budding yeast (S. cerevisiae). Using DNA sequencing, we characterised the directed evolution of these populations under selective inhibition of rate-limiting steps of the cell cycle. We observed recurrent patterns of adaptive mutations and characterised common mutational processes, but the spectrum of mutations at the molecular level remained stochastic. Finally, we investigated the effect of genetic variation on the fate of new mutations, which gives rise to complex evolutionary dynamics. We demonstrate that the fitness variance of the population can set a selective threshold on new mutations, setting a limit to the efficiency of selection. In summary, we combined statistical analyses of genomic sequences, mathematical models of evolutionary dynamics and experiments in molecular evolution to advance our understanding of rapid adaptation. Our results open new avenues in our understanding of population dynamics that can be translated to a range of biological systems.
- Published
- 2018
- Full Text
- View/download PDF
26. Distributed parameter and state estimation for wireless sensor networks
- Author
-
Yu, Jia, Thompson, John, and Mulgrew, Bernie
- Subjects
distributed algorithms ,statistical inference ,wireless sensor networks ,WSNs applications ,EM algorithms ,Gaussian mixtures ,EM gradient algorithm ,Bernoulli model ,BFGS formula - Abstract
The research in distributed algorithms is linked with the developments of statistical inference in wireless sensor networks (WSNs) applications. Typically, distributed approaches process the collected signals from networked sensor nodes. That is to say, the sensors receive local observations and transmit information between each other. Each sensor is capable of combining the collected information with its own observations to improve performance. In this thesis, we propose novel distributed methods for the inference applications using wireless sensor networks. In particular, the efficient algorithms which are not computationally intensive are investigated. Moreover, we present a number of novel algorithms for processing asynchronous network events and robust state estimation. In the first part of the thesis, a distributed adaptive algorithm based on the component-wise EM method for decentralized sensor networks is investigated. The distributed component-wise Expectation-Maximization (EM) algorithm has been designed for application in a Gaussian density estimation. The proposed algorithm operates a component-wise EM procedure for local parameter estimation and exploit an incremental strategy for network updating, which can provide an improved convergence rate. Numerical simulation results have illustrated the advantages of the proposed distributed component-wise EM algorithm for both well-separated and overlapped mixture densities. The distributed component-wise EM algorithm can outperform other EM-based distributed algorithms in estimating overlapping Gaussian mixtures. In the second part of the thesis, a diffusion based EM gradient algorithm for density estimation in asynchronous wireless sensor networks has been proposed. Specifically, based on the asynchronous adapt-then-combine diffusion strategy, a distributed EM gradient algorithm that can deal with asynchronous network events has been considered. The Bernoulli model has been exploited to approximate the asynchronous behaviour of the network. Compared with existing distributed EM based estimation methods using a consensus strategy, the proposed algorithm can provide more accurate estimates in the presence of asynchronous networks uncertainties, such as random link failures, random data arrival times, and turning on or off sensor nodes for energy conservation. Simulation experiments have been demonstrated that the proposed algorithm significantly outperforms the consensus based strategies in terms of Mean-Square- Deviation (MSD) performance in an asynchronous network setting. Finally, the challenge of distributed state estimation in power systems which requires low complexity and high stability in the presence of bad data for a large scale network is addressed. A gossip based quasi-Newton algorithm has been proposed for solving the power system state estimation problem. In particular, we have applied the quasi-Newton method for distributed state estimation under the gossip protocol. The proposed algorithm exploits the Broyden- Fletcher-Goldfarb-Shanno (BFGS) formula to approximate the Hessian matrix, thus avoiding the computation of inverse Hessian matrices for each control area. The simulation results for IEEE 14 bus system and a large scale 4200 bus system have shown that the distributed quasi-Newton scheme outperforms existing algorithms in terms of Mean-Square-Error (MSE) performance with bad data.
- Published
- 2017
27. Inferencia estadística en los textos escolares: una aproximación al pensamiento estadístico
- Author
-
Rodríguez Alveal, Francisco, Aguerrea, Maitere, Rodríguez Alveal, Francisco, and Aguerrea, Maitere
- Abstract
[Objective] The main purpose of this study was to analyze the activities related to statistical inference present in secondary education textbooks in Chile, and their relationship to the development of statistical thinking. [Methodology] For the purposes of the study, a qualitative approach was used through a content analysis of the secondary education books disseminated free of charge by the Chilean Ministry of Education during the years 2016, 2018, 2020, 2021 and 2022, which were selected using intentional non-probabilistic sampling. [Results] Among the main results obtained, it was found that most of the activities related to statistical inference in the textbooks analyzed had to do with procedures related to the calculation of confidence intervals, without addressing their interpretation in the context of the problem. In addition, it was observed that there were few activities related to informal inference. Likewise, the activities and questions presented in the textbooks do not refer to the formulation of hypotheses or conjectures about the sample data made to guide the formulation of conclusions related to statistical inference. [Conclusions] The findings indicate that despite the fact that confidence intervals are among the topics related to statistical inference in the textbooks, these textbooks are not focused on decision-making at the population level, but rather on description., [Objetivo] O objetivo principal deste estudo foi analisar as atividades relacionadas à inferência estatística, presentes textos escolares do ensino médio no Chile, e sua relação com o desenvolvimento do pensamento estatístico. [Metodologia] Para efeitos do estudo, utilizou-se uma abordagem qualitativa através de uma análise de conteúdo dos livros do ensino secundário distribuídos gratuitamente pelo Ministério da Educação do Chile durante os anos de 2016, 2018, 2020, 2021 e 2022, que foram selecionados por meio de amostragem não probabilística do tipo intencional. [Resultados] Dentre os principais resultados obtidos, a maior parte das atividades relacionadas à inferência estatística nos livros analisados são do tipo processual, relacionadas ao cálculo de intervalos de confiança, sem abordar sua interpretação no contexto da situação problema. Além disso, há uma presença escassa de atividades relacionadas à inferência informal. Da mesma forma, as atividades e questões explicadas nos livros de texto não se referem à formulação de hipóteses ou conjecturas sobre os dados amostrais, de modo que a formulação de conclusões seja orientada para a inferência estatística. [Conclusões] Os resultados fornecem evidências de que, embora os intervalos de confiança estejam dentro dos tópicos relacionados à inferência estatística nos livros de texto, eles não se concentram na tomada de decisões em nível populacional, mas são tratados a partir de um vértice descritivo., [Objetivo] El presente estudio tuvo como objetivo principal analizar las actividades relacionadas con inferencia estadística, presentes en los textos escolares de enseñanza secundaria en Chile, y su relación con el desarrollo del pensamiento estadístico. [Metodología] Para efectos del estudio se hizo uso de un enfoque cualitativo mediante un análisis de contenido de los libros de enseñanza secundaria difundidos, gratuitamente, por el Ministerio de Educación chileno durante los años 2016, 2018, 2020, 2021 y 2022, los cuales fueron seleccionados mediante un muestreo no probabilístico del tipo intencionado. [Resultados] Entre los principales resultados obtenidos, la mayoría de las actividades afines a inferencia estadística en los libros analizados, son del tipo procedimental, relacionadas con el cálculo de intervalos de confianza, sin abordar la interpretación de estos en el contexto de la situación problema. Además, se observa una escasa presencia de actividades relacionadas con inferencia informal. Asimismo, las actividades e interpelaciones explicitadas en los libros de texto no hacen referencia a la formulación de hipótesis o conjeturas acerca de los datos muestrales, de manera que se oriente a la formulación de conclusiones hacia la inferencia estadística. [Conclusiones] Los hallazgos entregan evidencias que, a pesar de que los intervalos de confianza se encuentran dentro de las temáticas afines a inferencia estadística en los libros de texto, estas no se focalizan en la toma de decisiones en el nivel poblacional, más bien son tratados desde un vértice descriptivo.
- Published
- 2024
28. Addressing bias in statistical inference based on epidemiological registry data
- Author
-
Universitat Politècnica de Catalunya. Departament d'Estadística i Investigació Operativa, Gómez Melis, Guadalupe, Plana Ripoll, Oleguer, Gallego Alabanda, Dídac, Universitat Politècnica de Catalunya. Departament d'Estadística i Investigació Operativa, Gómez Melis, Guadalupe, Plana Ripoll, Oleguer, and Gallego Alabanda, Dídac
- Abstract
Hospital’s registry data is a widely used resource in Nordic countries to estimate parameters of interest in the public health and in epidemiology. This data allows the researcher to have an unbiased representation of the population, since it is collected for all the individuals that visit the hospitals in the country, and it is stored in databases that are available for the researchers. Despite being a powerful tool, this data has some drawbacks that have to be taken into account before making the study. This project aims to expose the problems that arise when hospital’s registers are used for estimating the incidence of a disease, and explain what it is usually done to correct (or partially correct) them. First, we provide an introduction (Chapter 1) where we discuss the importance of getting good estimations to study the incidence of mental disorders, why registers are a powerful tool to make population-wide research and the main problem we encounter in this type of data: the delayed entries. We continue with a summary on the methodologies that will be used during the rest of the project (Chapter 2). In Chapter 3 we develop further the strengths and limitations of using registration data to estimate the incidence of a disease, providing a self-derived theoretical description of the delayed-entries problematic. In this chapter we also include the methodology that it is usually applied to deal with this problematic: the washout-period method, and we relate this methodology with the provided theoretical description. We finish the chapter with an introduction to the history of the registration system in Denmark and its structure. Lastly, we provide a simulation study based on the data of women diagnosed with depression in Denmark. We have simulated a hospital register and studied the incidence in terms of the cumulative incidence function applying the washout period method.
- Published
- 2024
29. Data-driven Parameter Estimation of Stochastic Models with Applications in Finance
- Author
-
Ayorinde, Ayoola and Ayorinde, Ayoola
- Abstract
Parameter estimation is a powerful and adaptable framework that addresses the inherent complexities and uncertainties of financial data. We provide an overview of likelihood functions, and likelihood estimations, as well as the essential numerical approximations and techniques. In the financial domain, where unpredictable and non-stationary market dynamics prevail, parameter estimations of relevant SDE models prove highly relevant. We delve into practical applications, showcasing how SDEs can effectively capture the inherent uncertainties and dynamics of financial models related to time evolution of interest rates. We work with the Vašíček model and Cox-Ingersoll-Ross (CIR) model which describes the dynamics of interest rates over time. We incorporate the Maximum likelihood and Quasi-maximum likelihood estimation methods in estimating the parameters of our models.
- Published
- 2024
30. A new likelihood inequality for models with latent variables
- Author
-
Olsen, Niels Aske Lundtorp and Olsen, Niels Aske Lundtorp
- Abstract
Likelihood-based approaches are central in statistics and its applications, yet often challenging since likelihoods can be intractable. Many methods such as the EM algorithm have been developed to alleviate this. We present a new likelihood inequality involving posterior distributions of a latent variable that holds under conditions similar to the EM algorithm. Potential scopes of the inequality includes maximum-likelihood estimation, likelihood ratios tests and model selection. We demonstrate the latter by performing selection in a non-linear mixed-model using MCMC.
- Published
- 2024
31. Approximation methods and inference for stochastic biochemical kinetics
- Author
-
Schnoerr, David Benjamin, Grima, Ramon, and Sanguinetti, Guido
- Subjects
519.2 ,stochastic processes ,approximation methods ,statistical inference - Abstract
Recent experiments have shown the fundamental role that random fluctuations play in many chemical systems in living cells, such as gene regulatory networks. Mathematical models are thus indispensable to describe such systems and to extract relevant biological information from experimental data. Recent decades have seen a considerable amount of modelling effort devoted to this task. However, current methodologies still present outstanding mathematical and computational hurdles. In particular, models which retain the discrete nature of particle numbers incur necessarily severe computational overheads, greatly complicating the tasks of characterising statistically the noise in cells and inferring parameters from data. In this thesis we study analytical approximations and inference methods for stochastic reaction dynamics. The chemical master equation is the accepted description of stochastic chemical reaction networks whenever spatial effects can be ignored. Unfortunately, for most systems no analytic solutions are known and stochastic simulations are computationally expensive, making analytic approximations appealing alternatives. In the case where spatial effects cannot be ignored, such systems are typically modelled by means of stochastic reaction-diffusion processes. As in the non-spatial case an analytic treatment is rarely possible and simulations quickly become infeasible. In particular, the calibration of models to data constitutes a fundamental unsolved problem. In the first part of this thesis we study two approximation methods of the chemical master equation; the chemical Langevin equation and moment closure approximations. The chemical Langevin equation approximates the discrete-valued process described by the chemical master equation by a continuous diffusion process. Despite being frequently used in the literature, it remains unclear how the boundary conditions behave under this transition from discrete to continuous variables. We show that this boundary problem results in the chemical Langevin equation being mathematically ill-defined if defined in real space due to the occurrence of square roots of negative expressions. We show that this problem can be avoided by extending the state space from real to complex variables. We prove that this approach gives rise to real-valued moments and thus admits a probabilistic interpretation. Numerical examples demonstrate better accuracy of the developed complex chemical Langevin equation than various real-valued implementations proposed in the literature. Moment closure approximations aim at directly approximating the moments of a process, rather then its distribution. The chemical master equation gives rise to an infinite system of ordinary differential equations for the moments of a process. Moment closure approximations close this infinite hierarchy of equations by expressing moments above a certain order in terms of lower order moments. This is an ad hoc approximation without any systematic justification, and the question arises if the resulting equations always lead to physically meaningful results. We find that this is indeed not always the case. Rather, moment closure approximations may give rise to diverging time trajectories or otherwise unphysical behaviour, such as negative mean values or unphysical oscillations. They thus fail to admit a probabilistic interpretation in these cases, and care is needed when using them to not draw wrong conclusions. In the second part of this work we consider systems where spatial effects have to be taken into account. In general, such stochastic reaction-diffusion processes are only defined in an algorithmic sense without any analytic description, and it is hence not even conceptually clear how to define likelihoods for experimental data for such processes. Calibration of such models to experimental data thus constitutes a highly non-trivial task. We derive here a novel inference method by establishing a basic relationship between stochastic reaction-diffusion processes and spatio-temporal Cox processes, two classes of models that were considered to be distinct to each other to this date. This novel connection naturally allows to compute approximate likelihoods and thus to perform inference tasks for stochastic reaction-diffusion processes. The accuracy and efficiency of this approach is demonstrated by means of several examples. Overall, this thesis advances the state of the art of modelling methods for stochastic reaction systems. It advances the understanding of several existing methods by elucidating fundamental limitations of these methods, and several novel approximation and inference methods are developed.
- Published
- 2016
32. Hidden states, hidden structures : Bayesian learning in time series models
- Author
-
Murphy, James Kevin
- Subjects
620 ,Information engineering ,Statistics ,Bayesian statistics ,Monte Carlo methods ,Statistical inference ,Time series ,Network analysis ,Gaussian processes - Abstract
This thesis presents methods for the inference of system state and the learning of model structure for a number of hidden-state time series models, within a Bayesian probabilistic framework. Motivating examples are taken from application areas including finance, physical object tracking and audio restoration. The work in this thesis can be broadly divided into three themes: system and parameter estimation in linear jump-diffusion systems, non-parametric model (system) estimation and batch audio restoration. For linear jump-diffusion systems, efficient state estimation methods based on the variable rate particle filter are presented for the general linear case (chapter 3) and a new method of parameter estimation based on Particle MCMC methods is introduced and tested against an alternative method using reversible-jump MCMC (chapter 4). Non-parametric model estimation is examined in two settings: the estimation of non-parametric environment models in a SLAM-style problem, and the estimation of the network structure and forms of linkage between multiple objects. In the former case, a non-parametric Gaussian process prior model is used to learn a potential field model of the environment in which a target moves. Efficient solution methods based on Rao-Blackwellized particle filters are given (chapter 5). In the latter case, a new way of learning non-linear inter-object relationships in multi-object systems is developed, allowing complicated inter-object dynamics to be learnt and causality between objects to be inferred. Again based on Gaussian process prior assumptions, the method allows the identification of a wide range of relationships between objects with minimal assumptions and admits efficient solution, albeit in batch form at present (chapter 6). Finally, the thesis presents some new results in the restoration of audio signals, in particular the removal of impulse noise (pops and clicks) from audio recordings (chapter 7).
- Published
- 2014
- Full Text
- View/download PDF
33. Geometric context from single and multiple views
- Author
-
Flint, Alexander John, Reid, Ian, and Murray, David
- Subjects
006.37 ,Image understanding ,Robotics ,computer vision ,geometry ,artificial intelligence ,statistical inference - Abstract
In order for computers to interact with and understand the visual world, they must be equipped with reasoning systems that include high–level quantities such as objects, actions, and scenes. This thesis is concerned with extracting such representations of the world from visual input. The first part of this thesis describes an approach to scene understanding in which texture characteristics of the visual world are used to infer scene categories. We show that in the context of a moving camera, it is common to observe images containing very few individually salient image regions, yet overall texture structure often allows our system to derive powerful contextual cues about the environment. Our approach builds on ideas from texture recognition, and we show that our algorithm out–performs the well–known Gist descriptor on several classification tasks. In the second part of this thesis we we are interested in scene understanding in the context of multiple calibrated views of a scene, as might be obtained from a Structure–from–Motion or Simultaneous Localization and Mapping (SLAM) system. Though such systems are capable of localizing the camera robustly and efficiently, the maps produced are typically sparse point-clouds that are difficult to interpret and of little use for higher–level reasoning tasks such as scene understanding or human-machine interaction. In this thesis we begin to address this deficiency, presenting progress towards modeling scenes using semantically meaningful primitives such as floor, wall, and ceiling planes. To this end we adopt the indoor Manhattan representation, which was recently proposed for single–view reconstruction. This thesis presents the first in–depth description and analysis of this model in the literature. We describe a probabilistic model relating photometric features, stereo photo–consistencies, and 3D point clouds to Manhattan scene structure in a Bayesian framework. We then present a fast dynamic programming algorithm that solves exact MAP inference in this model in time linear in image size. We show detailed comparisons with the state–of–the art in both the single– and multiple–view contexts. Finally, we present a framework for learning within the indoor Manhattan hypothesis class. Our system is capable of extrapolating from labelled training examples to predict scene structure for unseen images. We cast learning as a structured prediction problem and show how to optimize with respect to two realistic loss functions. We present experiments in which we learn to recover scene structure from both single and multiple views — from the perspective of our learning algorithm these problems differ only by a change of feature space. This work constitutes one of the most complicated output spaces (in terms of internal constraints) yet considered within a structure prediction framework.
- Published
- 2012
34. Delay estimation in computer networks
- Author
-
Johnson, Nicholas Alexander, McLaughlin, Steven., and Thompson, John
- Subjects
004.6 ,network tomography ,statistical inference ,bottleneck-link detection ,Gaussian distribution ,Kullback-Leibler divergence - Abstract
Computer networks are becoming increasingly large and complex; more so with the recent penetration of the internet into all walks of life. It is essential to be able to monitor and to analyse networks in a timely and efficient manner; to extract important metrics and measurements and to do so in a way which does not unduly disturb or affect the performance of the network under test. Network tomography is one possible method to accomplish these aims. Drawing upon the principles of statistical inference, it is often possible to determine the statistical properties of either the links or the paths of the network, whichever is desired, by measuring at the most convenient points thus reducing the effort required. In particular, bottleneck-link detection methods in which estimates of the delay distributions on network links are inferred from measurements made at end-points on network paths, are examined as a means to determine which links of the network are experiencing the highest delay. Initially two published methods, one based upon a single Gaussian distribution and the other based upon the method-of-moments, are examined by comparing their performance using three metrics: robustness to scaling, bottleneck detection accuracy and computational complexity. Whilst there are many published algorithms, there is little literature in which said algorithms are objectively compared. In this thesis, two network topologies are considered, each with three configurations in order to determine performance in six scenarios. Two new estimation methods are then introduced, both based on Gaussian mixture models which are believed to offer an advantage over existing methods in certain scenarios. Computationally, a mixture model algorithm is much more complex than a simple parametric algorithm but the flexibility in modelling an arbitrary distribution is vastly increased. Better model accuracy potentially leads to more accurate estimation and detection of the bottleneck. The concept of increasing flexibility is again considered by using a Pearson type-1 distribution as an alternative to the single Gaussian distribution. This increases the flexibility but with a reduced complexity when compared with mixture model approaches which necessitate the use of iterative approximation methods. A hybrid approach is also considered where the method-of-moments is combined with the Pearson type-1 method in order to circumvent problems with the output stage of the former. This algorithm has a higher variance than the method-of-moments but the output stage is more convenient for manipulation. Also considered is a new approach to detection algorithms which is not dependant on any a-priori parameter selection and makes use of the Kullback-Leibler divergence. The results show that it accomplishes its aim but is not robust enough to replace the current algorithms. Delay estimation is then cast in a different role, as an integral part of an algorithm to correlate input and output streams in an anonymising network such as the onion router (TOR). TOR is used by users in an attempt to conceal network traffic from observation. Breaking the encryption protocols used is not possible without significant effort but by correlating the un-encrypted input and output streams from the TOR network, it is possible to provide a degree of certainty about the ownership of traffic streams. The delay model is essential as the network is treated as providing a pseudo-random delay to each packet; having an accurate model allows the algorithm to better correlate the streams.
- Published
- 2010
35. Statistical inference in population genetics using microsatellites
- Author
-
Csilléry, Katalin, Pemberton, Josephine., and Johnson, Toby
- Subjects
519 ,molecular population genetic data ,statistical inference ,demographic factors of modern populations ,genetic factors of modern populations ,microsatellite evolution ,statistical methods - Abstract
Statistical inference from molecular population genetic data is currently a very active area of research for two main reasons. First, in the past two decades an enormous amount of molecular genetic data have been produced and the amount of data is expected to grow even more in the future. Second, drawing inferences about complex population genetics problems, for example understanding the demographic and genetic factors that shaped modern populations, poses a serious statistical challenge. Amongst the many different kinds of genetic data that have appeared in the past two decades, the highly polymorphic microsatellites have played an important role. Microsatellites revolutionized the population genetics of natural populations, and were the initial tool for linkage mapping in humans and other model organisms. Despite their important role, and extensive use, the evolutionary dynamics of microsatellites are still not fully understood, and their statistical methods are often underdeveloped and do not adequately model microsatellite evolution. In this thesis, I address some aspects of this problem by assessing the performance of existing statistical tools, and developing some new ones. My work encompasses a range of statistical methods from simple hypothesis testing to more recent, complex computational statistical tools. This thesis consists of four main topics. First, I review the statistical methods that have been developed for microsatellites in population genetics applications. I review the different models of the microsatellite mutation process, and ask which models are the most supported by data, and how models were incorporated into statistical methods. I also present estimates of mutation parameters for several species based on published data. Second, I evaluate the performance of estimators of genetic relatedness using real data from five vertebrate populations. I demonstrate that the overall performance of marker-based pairwise relatedness estimators mainly depends on the population relatedness composition and may only be improved by the marker data quality within the limits of the population relatedness composition. Third, I investigate the different null hypotheses that may be used to test for independence between loci. Using simulations I show that testing for statistical independence (i.e. zero linkage disequilibrium, LD) is difficult to interpret in most cases, and instead a null hypothesis should be tested, which accounts for the “background LD” due to finite population size. I investigate the utility of a novel approximate testing procedure to circumvent this problem, and illustrate its use on a real data set from red deer. Fourth, I explore the utility of Approximate Bayesian Computation, inference based on summary statistics, to estimate demographic parameters from admixed populations. Assuming a simple demographic model, I show that the choice of summary statistics greatly influences the quality of the estimation, and that different parameters are better estimated with different summary statistics. Most importantly, I show how the estimation of most admixture parameters can be considerably improved via the use of linkage disequilibrium statistics from microsatellite data.
- Published
- 2009
36. Cuantificación y análisis estadístico de contaminantes en el Mar Menor
- Author
-
García Onsurbe, María del Carmen, Vicente Martínez, Yesica, Caravaca Garratón, Manuel, García Onsurbe, María del Carmen, Vicente Martínez, Yesica, and Caravaca Garratón, Manuel
- Abstract
[ESP] El Mar Menor es la laguna costera hipersalina más grande de Europa. La eutrofización es el proceso de contaminación más importante de las aguas de lagos, estanques, ríos y embalses. Este proceso es causado por el exceso de nutrientes en el agua, principalmente nitrógeno y fósforo, provenientes principalmente de la actividad humana. En este trabajo presentamos un análisis cuantitativo de contaminantes en las aguas del Mar Menor en diferentes playas de la laguna. Las muestras se analizaron con el sistema de cromatografía iónica. [ENG] The Mar Menor is the largest hypersaline coastal lagoon in Europe. Eutrophication is the most important contamination process of the waters of lakes, ponds, rivers, and reservoirs. This process is caused by the excess of nutrients in the water, mainly nitrogen and phosphorus, coming mainly from human activity. In this work we present a quantitative analysis of contaminants in the waters of the Mar Menor in different beaches of the lagoon. Samples were analyzed on the Metrohm 861 ion chromatography system.
- Published
- 2023
37. Predictive Modeling and Statistical Inference for CTA returns : A Hidden Markov Approach with Sparse Logistic Regression
- Author
-
Fransson, Oskar and Fransson, Oskar
- Abstract
This thesis focuses on predicting trends in Commodity Trading Advisors (CTAs), also known as trend-following hedge funds. The paper applies a Hidden Markov Model (HMM) for classifying trends. Additionally, by incorporating additional features, a regularized logistic regression model is used to enhance prediction capability. The model demonstrates success in identifying positive trends in CTA funds, with particular emphasis on precision and risk-adjusted return metrics. In the context of regularized regression models, techniques for statistical inference such as bootstrap resampling and Markov Chain Monte Carlo are applied to estimate the distribution of parameters. The findings suggest the model's effectiveness in predicting favorable CTA performance and mitigating equity market drawdowns. For future research, it is recommended to explore alternative classification models and extend the methodology to different markets and datasets.
- Published
- 2023
38. Causal Inference and Large Language Models from the Causal Invariance Framework
- Author
-
Wong, Emily Frances, Lu, Hongjing1, Wong, Emily Frances, Wong, Emily Frances, Lu, Hongjing1, and Wong, Emily Frances
- Abstract
Statistics serves as the grammar of all science, and central to the goal of science is understanding cause-effect relationships. Scientists rely on research methodology and statistical tools to uncover causal relationships, and engineers rely on statistical methods to create artificial assistants to aid daily life. Neither statistical learning nor next-word-prediction (used to train artificial general intelligence) are consistent with rational causal learning and reasoning in humans. The present thesis examines the fundamental goals and assumptions made in dominant statistical methods and discusses their implications for statistical inference and commonsense reasoning in artificial general intelligence (AGI). The first section introduces and evaluates a causal alternative to logistic regression, which estimates the causal power (from the causal invariance framework) of treatments among covariates. Causal invariance is defined as the influence of a candidate cause (elemental or conjunctive) that is independent of background causes, with the aspiration of acquiring knowledge that’s useable, in the minimalist sense being able to generalize from a learning context to an application context. The second and final section investigates current benchmark tasks used to evaluate causal reasoning in large language models (e.g., GPT-3, GPT-4), and introduces a stricter test informed by psychological literature on human causal cognition under the causal invariance framework.
- Published
- 2023
39. On the Value of Online Learning for Cognitive Radar Waveform Selection
- Author
-
Thornton III, Charles Ethridge and Thornton III, Charles Ethridge
- Abstract
Modern radar systems must operate in a wide variety of time-varying conditions. These include various types of interference from neighboring systems, self-interference or clutter, and targets with fluctuating responses. It has been well-established that the quality and nature of radar measurements depend heavily on the choice of signal transmitted by the radar. In this dissertation, we discuss techniques which may be used to adapt the radar's waveform on-the-fly while making very few a priori assumptions about the physical environment. By employing tools from reinforcement learning and online learning, we present a variety of algorithms which handle practical issues of the waveform selection problem that have been left open by previous works. In general, we focus on two key challenges inherent to the waveform selection problem, sample-efficiency and universality. Sample-efficiency corresponds to the number of experiences a learning algorithm requires to achieve desirable performance. Universality refers to the learning algorithm's ability to achieve desirable performance across a wide range of physical environments. Specifically, we develop a contextual bandit-based approach to vastly improve the sample-efficiency of learning compared to previous works. We then improve the generalization performance of this model by developing a Bayesian meta-learning technique. To handle the problem of universality, we develop a learning algorithm which is asymptotically optimal in any Markov environment having finite memory length. Finally, we compare the performance of learning-based waveform selection to fixed rule-based waveform selection strategies for the scenarios of dynamic spectrum access and multiple-target tracking. We draw conclusions as to when learning-based approaches are expected to significantly outperform rule-based strategies, as well as the converse.
- Published
- 2023
40. Are Small Effects the Indispensable Foundation for a Cumulative Psychological Science? A Reply to Götz et al. (2022)
- Author
-
Primbs, Maximilian A., Pennington, Charlotte R., Lakens, Daniël, Silan, Miguel Alejandro A., Lieck, Dwayne S.N., Forscher, Patrick S., Buchanan, Erin M., Westwood, Samuel J., Primbs, Maximilian A., Pennington, Charlotte R., Lakens, Daniël, Silan, Miguel Alejandro A., Lieck, Dwayne S.N., Forscher, Patrick S., Buchanan, Erin M., and Westwood, Samuel J.
- Abstract
In the January 2022 issue of Perspectives, Götz et al. argued that small effects are “the indispensable foundation for a cumulative psychological science.” They supported their argument by claiming that (a) psychology, like genetics, consists of complex phenomena explained by additive small effects; (b) psychological-research culture rewards large effects, which means small effects are being ignored; and (c) small effects become meaningful at scale and over time. We rebut these claims with three objections: First, the analogy between genetics and psychology is misleading; second, p values are the main currency for publication in psychology, meaning that any biases in the literature are (currently) caused by pressure to publish statistically significant results and not large effects; and third, claims regarding small effects as important and consequential must be supported by empirical evidence or, at least, a falsifiable line of reasoning. If accepted uncritically, we believe the arguments of Götz et al. could be used as a blanket justification for the importance of any and all “small” effects, thereby undermining best practices in effect-size interpretation. We end with guidance on evaluating effect sizes in relative, not absolute, terms.
- Published
- 2023
41. Learning Active Inference MODELs of Perception and Control: Application to Car Following Task
- Published
- 2023
42. Enhancing Automated Vehicle Safety Through Testing with Realistic Driver Models [supporting dataset]
- Published
- 2023
43. HCV E1 influences the fitness landscape of E2 and may enhance escape from E2-specific antibodies
- Author
-
Zhang, Hang, Bull, Rowena A., Quadeer, Ahmed Abdul, McKay, Matthew R., Zhang, Hang, Bull, Rowena A., Quadeer, Ahmed Abdul, and McKay, Matthew R.
- Abstract
The Hepatitis C virus (HCV) envelope glycoprotein E1 forms a non-covalent heterodimer with E2, the main target of neutralizing antibodies. How E1-E2 interactions influence viral fitness and contribute to resistance to E2-specific antibodies remain largely unknown. We investigate this problem using a combination of fitness landscape and evolutionary modeling. Our analysis indicates that E1 and E2 proteins collectively mediate viral fitness and suggests that fitness-compensating E1 mutations may accelerate escape from E2-targeting antibodies. Our analysis also identifies a set of E2-specific human monoclonal antibodies that are predicted to be especially resilient to escape via genetic variation in both E1 and E2, providing directions for robust HCV vaccine development. © The Author(s) 2023. Published by Oxford University Press.
- Published
- 2023
44. Data contamination versus model deviation
- Author
-
Fonseca, Viviane Grunert da
- Subjects
519.5 ,Statistical inference ,Robustness ,Diagnostics - Published
- 1999
45. Introducing Statistical Inference: Design of a Theoretically and Empirically Based Learning Trajectory
- Author
-
van Dijke-Droogers, Marianne, Drijvers, Paul, Bakker, Arthur, van Dijke-Droogers, Marianne, Drijvers, Paul, and Bakker, Arthur
- Abstract
This paper comprises the results of a design study that aims at developing a theoretically and empirically based learning trajectory on statistical inference for 9th-grade students. Based on theories of informal statistical inference, an 8-step learning trajectory was designed. The trajectory consisted of two similar four step sequences: (1) experimenting with a physical black box, (2) visualizing distributions, (3) examining sampling distributions using simulation software, and (4) interpreting sampling distributions to make inferences in real -life contexts. Sequence I included only categorical data and Sequence II regarded numerical data. The learning trajectory was implemented in an intervention among 267 students. To examine the effects of the trajectory on students’ understanding of statistical inference, we analyzed their posttest results after the intervention. To investigate how the stepwise trajectory fostered the learning process, students’ worksheets during each learning step were analyzed. The posttest results showed that students who followed the learning trajectory scored significantly higher on statistical inference and on concepts related to each step than students of a comparison group (n = 217) who followed the regular curriculum. Worksheet analysis demonstrated that the 8-step trajectory was beneficial to students’ learning processes. We conclude that ideas of repeated sampling with a black box and statistical modeling seem fruitful for introducing statistical inference. Both ideas invite more advanced follow-up activities, such as hypothesis testing and comparing groups. This suggests that statistics curricula with a descriptive focus can be transformed to a more inferential focus, to anticipate on subsequent steps in students’ statistics education.
- Published
- 2022
46. Urnings: A new method for tracking dynamically changing parameters in paired comparison systems
- Author
-
Bolsinova, Maria, Maris, Gunter, Hofman, Abe D., van der Maas, Han L.J., Brinkhuis, Matthieu J.S., Bolsinova, Maria, Maris, Gunter, Hofman, Abe D., van der Maas, Han L.J., and Brinkhuis, Matthieu J.S.
- Abstract
We introduce a new rating system for tracking the development of parameters based on a stream of observations that can be viewed as paired comparisons. Rating systems are applied in competitive games, adaptive learning systems and platforms for product and service reviews. We model each observation as an outcome of a game of chance that depends on the parameters of interest (e.g. the outcome of a chess game depends on the abilities of the two players). Determining the probabilities of the different game outcomes is conceptualized as an urn problem, where a rating is represented by a probability (i.e. proportion of balls in the urn). This setup allows for evaluating the standard errors of the ratings and performing statistical inferences about the development of, and relations between, parameters. Theoretical properties of the system in terms of the invariant distributions of the ratings and their convergence are derived. The properties of the rating system are illustrated with simulated examples and its potential for answering research questions is illustrated using data from competitive chess, a movie review system, and an adaptive learning system for math.
- Published
- 2022
47. Learning opportunities for pre-service teachers to develop pedagogical content knowledge for statistical inference
- Author
-
Blomberg, Per, Högström, Per, Liljekvist, Yvonne, Blomberg, Per, Högström, Per, and Liljekvist, Yvonne
- Abstract
Recently, researchers have encouraged the teaching of statistical inference to students at all levels. However, what constitutes pre-service teachers’ pedagogical content knowledge for statistical inference has not yet been given specific attention in research. This paper presents a qualitative study of pre-service teachers participating in a collaborative learning setup in a mathematics course to be prepared for teaching statistics in primary school aged 6–10 years. The study reported here is the first cycle of a design research project. This first-phase study explores how pre-service teachers’ pedagogical content knowledge for statistical inference can be developed during their mathematics course. The findings show that pre-service teachers’ learning opportunities regarding pedagogical content knowledge for statistical inference are insufficient. Based on the initial results, an initial conjecture map was constructed that guides the forthcoming design cycle.
- Published
- 2022
48. Misinterpretations of P-values and statistical tests persist among researchers and professionals working with statistics and epidemiology
- Author
-
Lytsy, Per, Hartman, Mikael, Pingel, Ronnie, Lytsy, Per, Hartman, Mikael, and Pingel, Ronnie
- Abstract
Background: The aim was to investigate inferences of statistically significant test results among persons with more or less statistical education and research experience. Methods: A total of 75 doctoral students and 64 statisticians/epidemiologist responded to a web questionnaire about inferences of statistically significant findings. Participants were asked about their education and research experience, and also whether a 'statistically significant' test result (P = 0.024, alpha-level 0.05) could be inferred as proof or probability statements about the truth or falsehood of the null hypothesis (H-0) and the alternative hypothesis (H-1). Results: Almost all participants reported having a university degree, and among statisticians/epidemiologist, most reported having a university degree in statistics and were working professionally with statistics. Overall, 9.4% of statisticians/epidemiologist and 24.0% of doctoral students responded that the statistically significant finding proved that H-0 is not true, and 73.4% of statisticians/epidemiologists and 53.3% of doctoral students responded that the statistically significant finding indicated that H(0 )is improbable. Corresponding numbers about inferences about the alternative hypothesis (H-1) were 12.0% and 6.2% about proving H-1 being true and 62.7 and 62.5% for the conclusion that H-1 is probable. Correct inferences to both questions, which is that a statistically significant finding cannot be inferred as either proof or a measure of a hypothesis' probability, were given by 10.7% of doctoral students and 12.5% of statisticians/epidemiologists. Conclusions: Misinterpretation of P-values and statistically significant test results persists also among persons who have substantial statistical education and who work professionally with statistics.
- Published
- 2022
- Full Text
- View/download PDF
49. Data sharpening for improving central limit theorem approximations for data envelopment analysis-type efficiency estimators
- Author
-
UCL - SSH/LIDAM/ISBA - Institut de Statistique, Biostatistique et Sciences Actuarielles, Nguyen, Bao Hoang, Simar, Léopold, Zelenyuk, Valentin, UCL - SSH/LIDAM/ISBA - Institut de Statistique, Biostatistique et Sciences Actuarielles, Nguyen, Bao Hoang, Simar, Léopold, and Zelenyuk, Valentin
- Abstract
Asymptotic statistical inference on productivity and production efficiency, using nonparametric envelopment estimators, is now available thanks to the basic central limit theorems (CLTs) developed in Kneip, Simar, and Wilson (2015). They provide asymptotic distributions of averages of Data Envelopment Analysis (DEA) and Free Disposal Hull (FDH) estimators of production efficiency. As shown in their Monte-Carlo experiments, due to the curse of dimensionality, the accuracy of the normal approximation is disappoint- ing when the sample size is not large enough. Simar and Zelenyuk (2020) have suggested a simple way to improve the approximation by using a more appropriate estimator of the variances. In this paper we suggest a novel way to improve the approximation, by smoothing out the spurious values of efficiency estimates when they are in a neighborhood of 1. This results in sharpening the data for observations near the estimated efficient frontier. The method is very easy to implement and does not require more computations than the original method. We compare our approach using Monte-Carlo experiments, both with the basic method and with the improved method suggested in Simar & Zelenyuk (2020) and in both cases we observe significant improvements. We show also that the Simar & Zelenyuk (2020) idea of the variance correction can also be adapted to our sharpening method, bringing additional improvements. We illustrate the method with some real data sets.
- Published
- 2022
50. Infants use emotion to infer intentionality from non-random sampling events.
- Author
-
Lopez, Lukas D, Lopez, Lukas D, Walle, Eric A, Lopez, Lukas D, Lopez, Lukas D, and Walle, Eric A
- Abstract
Infants use statistical information in their environment, as well as others' emotional communication, to understand the intentions of social partners. However, rarely do researchers consider these two sources of social information in tandem. This study assessed 2-year-olds' attributions of intentionality from non-random sampling events and subsequent discrete emotion reactions. Infants observed an experimenter remove five objects from either the non-random minority (18%) or random majority (82%) of a sample and express either joy, disgust, or sadness after each selection. Two-year-olds inferred the experimenter's intentionality by giving her the object that she had previously selected when she expressed joy or disgust after non-random sampling events, but not when she expressed sadness or sampled at random. These findings demonstrate that infants use both statistical regularities and discrete emotion communication to infer an agent's intentions. In particular, the present findings show that 2-year-olds infer that an agent can intentionally select a preferred or an undesired object from a sample as a function of the discrete emotion. Implications for the development of inferring intentionality from statistical sampling events and discrete emotion communication are discussed.
- Published
- 2022
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.