365 results on '"Standard errors"'
Search Results
102. Using the Kernel Method of Test Equating for Estimating the Standard Errors of Population Invariance Measures.
- Author
-
Moses, Tim
- Subjects
MATHEMATICAL symmetry ,EDUCATIONAL testing services ,FACTORIZATION ,POPULATION ,LOG-linear models ,MULTIVARIATE analysis ,EDUCATIONAL statistics ,PROBABILITY theory ,MATHEMATICS - Abstract
The article shows how to extend the framework of kernel equating so that the standard errors of the root mean square difference and of the difference between two subpopulations' equated scores can be estimated. It discusses an investigation of population invariance for the equivalent groups design. The accuracies of the derived standard errors are evaluated with respect to empirical standard errors. It shows that the accuracy of the standard error estimates for the equated score differences is better than for the root mean square difference and that accuracy for both standard error estimates is best when sample sizes are large.
- Published
- 2008
- Full Text
- View/download PDF
103. Markov Chain Monte Carlo: Can We Trust the Third Significant Figure?
- Author
-
Flegal, James M., Haran, Murali, and Jones, Galin L.
- Subjects
MARKOV processes ,MONTE Carlo method ,STOCHASTIC convergence ,ERRORS ,MATHEMATICAL statistics ,STOCHASTIC processes - Abstract
Current reporting of results based on Markov chain Monte Carlo computations could be improved. In particular, a measure of the accuracy of the resulting estimates is rarely reported. Thus we have little ability to objectively assess the quality of the reported estimates. We address this issue in that we discuss why Monte Carlo standard errors are important, how they can be easily calculated in Markov chain Monte Carlo and how they can be used to decide when to stop the simulation. We compare their use to a popular alternative in the context of two examples. [ABSTRACT FROM AUTHOR]
- Published
- 2008
- Full Text
- View/download PDF
104. Bootstrap Estimates of Standard Errors in Generalizability Theory.
- Author
-
Ye Tong and Brennan, Robert L.
- Subjects
- *
STATISTICAL bootstrapping , *DISTRIBUTION (Probability theory) , *GENERALIZABILITY theory , *ESTIMATES , *ERRORS , *STATISTICS , *SIMULATION methods & models , *VARIANCES , *MATHEMATICS - Abstract
Estimating standard errors of estimated variance components has long been a challenging task in generalizability theory. Researchers have speculated about the potential applicability of the bootstrap for obtaining such estimates, but they have identified problems (especially bias) in using the bootstrap. Using Brennan's bias-correcting procedures (which extend the work of Wiley), as well as a proposed set of rules for picking a bootstrap procedure, the authors examined the potential utility of the bootstrap technique with multifacet designs in generalizability theory. Using six simulation conditions (normal, dichotomous, and polytomous data with the p × i × h and p × [i:h] designs), the rules proposed in this article were empirically demonstrated to perform well for estimating standard errors of estimated variance components and relative error variance with random models. No single bootstrap procedure performed well for estimating standard errors for absolute error variance, but a combination of bootstrap procedures was identified and empirically demonstrated to have performed well. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
105. Some Posterior Standard Deviations in Item Response Theory.
- Author
-
Seock-Ho Kim
- Subjects
- *
ITEM response theory , *EDUCATIONAL tests & measurements , *PSYCHOLOGICAL tests , *SCALE analysis (Psychology) , *ANALYSIS of variance , *STANDARD deviations , *COMPUTER software , *COGNITIVE learning , *PSYCHOMETRICS - Abstract
The procedures required to obtain the approximate posterior standard deviations of the parameters in the three commonly used item response models for dichotomous items are described and used to generate values for some common situations. The results were compared with those obtained from maximum likelihood estimation. It is shown that the use of priors may reduce the instability of estimates of the item parameters, assuming the choice of priors is reasonable. The sample size required for acceptable accuracy for the purposes of practical applications of item response theory may be inferred from tables or computer programs. It is suggested that the careful selection of priors be exercised to obtain the required precision when three-parameter models are applied in these situations. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
106. Standard errors and confidence intervals in inverse problems: sensitivity and associated pitfalls.
- Author
-
Banks, H. T., Ernstberger, S. L., and Grove, S. L.
- Subjects
- *
LEAST squares , *INVERSE problems , *ASYMPTOTIC theory of algebraic ideals , *DIFFERENTIAL equations , *PARAMETER estimation , *MATRICES (Mathematics) , *DIFFERENTIABLE dynamical systems - Abstract
We review the asymptotic theory for standard errors in classical ordinary least squares (OLS) inverse or parameter estimation problems involving general nonlinear dynamical systems where sensitivity matrices can be used to compute the asymptotic covariance matrices. We discuss possible pitfalls in computing standard errors in regions of low parameter sensitivity and/or near a steady state solution of the underlying dynamical system. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
107. On the Estimation of Standard Errors in Cognitive Diagnosis Models
- Author
-
Carolin Strobl, Jimmy de la Torre, Michel Philipp, Achim Zeileis, and University of Zurich
- Subjects
Computer science ,media_common.quotation_subject ,Computation ,3301 Social Sciences (miscellaneous) ,Inference ,computer.software_genre ,01 natural sciences ,Education ,Set (abstract data type) ,information matrix ,010104 statistics & probability ,symbols.namesake ,C30 ,C52 ,0504 sociology ,ddc:330 ,Quality (business) ,0101 mathematics ,Fisher information ,G-DINA ,media_common ,Estimation ,10093 Institute of Psychology ,Covariance matrix ,05 social sciences ,050401 social sciences methods ,DoktoratPsych Erstautor ,Standard error ,cognitive diagnosis model ,standard errors ,symbols ,Data mining ,150 Psychology ,C87 ,Algorithm ,computer ,Social Sciences (miscellaneous) ,3304 Education - Abstract
Cognitive diagnosis models (CDMs) are an increasingly popular method to assess mastery or nonmastery of a set of fine-grained abilities in educational or psychological assessments. Several inference techniques are available to quantify the uncertainty of model parameter estimates, to compare different versions of CDMs, or to check model assumptions. However, they require a precise estimation of the standard errors (or the entire covariance matrix) of the model parameter estimates. In this article, it is shown analytically that the currently widely used form of calculation leads to underestimated standard errors because it only includes the item parameters but omits the parameters for the ability distribution. In a simulation study, we demonstrate that including those parameters in the computation of the covariance matrix consistently improves the quality of the standard errors. The practical importance of this finding is discussed and illustrated using a real data example.
- Published
- 2017
- Full Text
- View/download PDF
108. Asymptotic robustness of standard errors in multilevel structural equation models
- Author
-
Yuan, Ke-Hai and Bentler, Peter M.
- Subjects
- *
MULTIVARIATE analysis , *ASYMPTOTIC distribution , *STATISTICAL correlation , *PATH analysis (Statistics) - Abstract
Abstract: Data in social and behavioral sciences are often hierarchically organized. Multilevel statistical methodology was developed to analyze such data. Most of the procedures for analyzing multilevel data are derived from maximum likelihood based on the normal distribution assumption. Standard errors for parameter estimates in these procedures are obtained from the corresponding information matrix. Because practical data typically contain heterogeneous marginal skewnesses and kurtoses, this paper studies how nonnormally distributed data affect the standard errors of parameter estimates in a two-level structural equation model. Specifically, we study how skewness and kurtosis in one level affect standard errors of parameter estimates within its level and outside its level. We also show that, parallel to asymptotic robustness theory in conventional factor analysis, conditions exist for asymptotic robustness of standard errors in a multilevel factor analysis model. [Copyright &y& Elsevier]
- Published
- 2006
- Full Text
- View/download PDF
109. PARSIMONIOUS PERIODIC TIME SERIES MODELING.
- Author
-
Lund, Robert, Qin Shao, and Basawa, Ishwar
- Subjects
- *
FOURIER series , *FOURIER analysis , *MATHEMATICAL models , *ERRORS , *ESTIMATION theory - Abstract
This paper studies techniques for fitting parsimonious periodic time series models to periodic data. Large sample standard errors for the parameter estimates in a periodic autoregressive moving-average time series model under parametric constraints are derived. Likelihood ratio statistics for hypothesis testing are examined. The techniques are applied in modeling daily temperatures at Griffin, Georgia, USA. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
110. AVOIDING BOUNDARY ESTIMATES IN LATENT CLASS ANALYSIS BY BAYESIAN POSTERIOR MODE ESTIMATION.
- Author
-
Garre, Francisca Galindo and Vermunt, Jeroen K.
- Abstract
In maximum likelihood estimation of latent class models, it often occurs that one or more of tile parameter estimates are on the boundary of the parameter space: that is, that estimated probabilities equal 0 (or l) or, equivalently, that legit coefficients equal minus (or plus) infinity. This not only causes numerical problems in tile computation of the variance-covariance matrix, it also makes the reported confidence intervals and significance tests for the parameters concerned meaningless. Boundary estimates can, however, easily be prevented by the use of prior distributions for the model parameters, yielding a Bayesian procedure called posterior mode or maximum a posteriori estimation. This approach is implemented in, for example, the Latent GOLD software packages for latent class analysis (Vermont & Magidson. 2005). Little is, however, known about the quality of posterior mode estimates of the parameters of latent class models, nor about their sensitivity for the choice of the prior distribution. In this paper, we compare the quality of various types of posterior mode point and interval estimates for the parameters of latent class models with both the classical maximum likelihood estimates and the bootstrap estimates proposed by De Menezes (1999). Our simulation study shows that parameter estimates and standard errors obtained by the Bayesian approach are more reliable than the corresponding parameter estimates and standard errors obtained by maximum likelihood and parametric bootstrapping. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
111. Sex ratio, and probability of sexual maturity of females at size, of the blue swimmer crab, Portunus pelagicus Linneaus, off southern Australia
- Author
-
Xiao, Yongshun and Kumar, Martin
- Subjects
- *
PORTUNIDAE , *BLUE swimming crab , *FISH sex ratio , *CRAB fisheries - Abstract
Although the blue swimmer crab, Portunus pelagicus, supports a commercial fishery in South Australia, with an annual landed value of about A$2 million, little information is available on its biology in this region, and this has hindered the development of strategies for effectively managing the fishery. We used generalized linear models to examine the sex ratio of males and females and the probability of sexual maturity of females at size, based on original commercial catch data collected in a fishery monitoring project. Specifically, we modeled temporal changes in the sex ratio as a cosine function and the probability of sexual maturity of females at size as a logistic function through a logit transformation. We reparameterized both equations, and gave formulae for calculating the standard errors of the estimates of reparameterized parameters characterizing such temporal changes in sex ratio, sizes at x% sexual maturity, the mean and variance of the size of females at sexual maturity. The sex ratio of male crabs in commercial catches decreased from a maximum on 3 February to a minimum on 2 July, and then increased from the minimum to the next maximum on 3 February for another cycle. Some fishers caught proportionally more males (which have larger sizes than females) than others. The sex ratio of male blue swimmer crabs was related to their condition upon capture: dead (killed by sea lice) crabs had a higher proportion of males than live (commercially usable) ones, which in turn had a higher proportion of males than soft-shelled (moulted) ones. The sex ratio of males increased with water depth from January to September and decreased with water depth from October to December. The sex ratio of males increased with carapace width. Finally, the carapace width of females at 50% maturity was 58.5 (±1.0) mm, and that at 95% maturity was 66.3 (±1.9) mm; the mean and variance of the carapace width of females at sexual maturity were 58.5 (±1.0) mm and 22.9 (±9.0) mm2. Therefore, the current legal minimum carapace width of 110 mm in the commercial fishery allows most mature female crabs to spawn before attaining the legal size for commercial fishing. [Copyright &y& Elsevier]
- Published
- 2004
- Full Text
- View/download PDF
112. Understanding persistence
- Author
-
Kelly, Morgan
- Subjects
Deep origins ,Explanatory variables ,Räumliche Statistik ,Regressionsanalyse ,Standard errors ,Spatial noise ,ddc:330 ,Autokorrelation ,Streuungsmaß ,Robustness checks ,Nichtparametrische Schätzung - Abstract
A large literature on persistence finds that many modern outcomes strongly reflect characteristics of the same places in the distant past. These studies typically combine unusually high t statistics with severe spatial autocorrelation in residuals, suggesting that some findings may be artefacts of underestimating standard errors or of fitting spatial trends. For 25 studies in leading journals, I apply three basic robustness checks against spatial trends and find that effect sizes typically fall by over half, leaving most well known results insignificant at conventional levels. Turning to standard errors, there is currently no data-driven method for selecting an appropriate HAC spatial kernel. The paper proposes a simple procedure where a kernel with a highly flexible functional form is estimated by maximum likelihood. After correction, standard errors tend to rise substantially for cross sectional studies but to fall for panels. Overall, credible identification strategies tend to perform no better than naive regressions. Although the focus here is on historical persistence, the methods apply to regressions using spatial data more generally.
- Published
- 2020
113. Standard error correction in two-stage estimation with nested samples.
- Author
-
Karaca-Mandic, Pinar and Train, Kenneth
- Subjects
ANALYSIS of covariance ,HOUSEHOLDS ,TELEVISION - Abstract
Data at different levels of aggregation are often used in two-stage estimation, with estimates obtained at the higher level of aggregation entering the estimation at the lower level of aggregation. An example is customers within markets: first-stage estimates on market data provide variables that enter the second-stage model on customers. We derive the asymptotic covariance matrix of the second-stage estimates for situations such as these. We implement the formulae in the Petrin–Train application of households’ choice of TV reception and compare the calculated standard errors with those obtained without correction. In this application, ignoring the sampling variance in the first-stage estimates would be seriously misleading. [ABSTRACT FROM AUTHOR]
- Published
- 2003
- Full Text
- View/download PDF
114. Sum of Profiles Model with Exchangeably Distributed Errors.
- Author
-
Mentz, Graciela B. and Kshirsagar, Anant M.
- Subjects
- *
MATHEMATICAL models , *ERROR analysis in mathematics , *STATISTICS - Abstract
We have considered a generalization of the basic Pottho.-Roy growth curve model by combining a sum of profiles model with an exchangeably distributed errors model. Analytical expressions for the growth curve coe4cients and their standard errors are obtain specifically. An illustration where such a model is useful in practice is also supplied. [ABSTRACT FROM AUTHOR]
- Published
- 2003
- Full Text
- View/download PDF
115. Oblique factors and components with independent clusters.
- Author
-
Ogasawara, Haruhiko
- Subjects
FACTOR analysis ,CLUSTER analysis (Statistics) ,VARIANCES ,STATISTICAL correlation ,MATHEMATICAL models - Abstract
Relationships between the results of factor analysis and component analysis are derived when oblique factors have independent clusters with equal variances of unique factors. The factor loadings are analytically shown to be smaller than the corresponding component loadings while the factor correlations are shown to be greater than the corresponding component correlations. The condition for the inequality of the factor/component contributions is derived in the case with different variances for unique factors. Further, the asymptotic standard errors of parameter estimates are obtained for a simplified model with the assumption of multivariate normality, which shows that the component loading estimate is more stable than the corresponding factor loading estimate. [ABSTRACT FROM AUTHOR]
- Published
- 2003
- Full Text
- View/download PDF
116. A bootstrap procedure for mixture models: applied to multidimensional scaling latent class models.
- Author
-
Winsberg, Suzanne and De Soete, Geert
- Subjects
STATISTICAL bootstrapping ,MULTIDIMENSIONAL scaling ,STATISTICAL sampling ,DISTRIBUTION (Probability theory) ,ERROR analysis in mathematics ,PARAMETER estimation - Abstract
A bootstrap procedure useful in latent class, or more general mixture models has been developed to determine the sufficient number of latent classes or components required to account for systematic group differences in the data. The procedure is illustrated in the context of a multidimensional scaling latent class model, CLASCAL. Also presented is a bootstrap technique for determining standard errors for estimates of the stimulus co-ordinates, parameters of the multidimensional scaling model. Real and artificial data are presented. The bootstrap procedure for selecting a sufficient number of classes seems to correctly select the correct number of latent classes at both low and high error levels. At higher error levels it outperforms Hope's (J. Roy. Statist. Soc. Ser B 1968; 30: 582) procedure. The bootstrap procedures to estimate parameter stability appear to correctly re-produce Monte Carlo results. Copyright © 2002 John Wiley & Sons, Ltd. [ABSTRACT FROM AUTHOR]
- Published
- 2002
- Full Text
- View/download PDF
117. A unified approach to exploratory factor analysis with missing data, nonnormal data, and in the presence of outliers.
- Author
-
KE-HAI YUAN, MARSHALL, LINDA L., and BENTLER, PETER M.
- Subjects
STRUCTURAL equation modeling ,ANALYSIS of covariance ,MATRICES (Mathematics) ,STATISTICAL sampling ,MAXIMUM likelihood statistics - Abstract
Factor analysis is regularly used for analyzing survey data. Missing data, data with outliers and consequently nonnormal data are very common for data obtained through questionnaires. Based on covariance matrix estimates for such nonstandard samples, a unified approach for factor analysis is developed. By generalizing the approach of maximum likelihood under constraints, statistical properties of the estimates for factor loadings and error variances are obtained. A rescaled Bartlett-corrected statistic is proposed for evaluating the number of factors. Equivariance and invariance of parameter estimates and their standard errors for canonical, varimax, and normalized varimax rotations are discussed. Numerical results illustrate the sensitivity of classical methods and advantages of the proposed procedures. [ABSTRACT FROM AUTHOR]
- Published
- 2002
- Full Text
- View/download PDF
118. Some relationships between factors and components.
- Author
-
OGASAWARA, HARUHIKO
- Subjects
FACTOR analysis ,PRINCIPAL components analysis ,STATISTICAL correlation ,GAUSSIAN distribution ,MULTIVARIATE analysis - Abstract
The asymptotic correlations between the estimates of factor and component loadings are obtained for the exploratory factor analysis model with the assumption of a multivariate normal distribution for manifest variables. The asymptotic correlations are derived for the cases of unstandardized and standardized manifest variables with orthogonal and oblique rotations. Based on the above results, the asymptotic standard errors for estimated correlations between factors and components are derived. Further, the asymptotic standard error of the mean squared canonical correlation for factors and components, which is an overall index for the closeness of factors and components, is derived. The results of a Monte Carlo simulation are presented to show the usefulness of the asymptotic results in the data with a finite sample size. [ABSTRACT FROM AUTHOR]
- Published
- 2000
- Full Text
- View/download PDF
119. An illustration of the use of model-based bootstrapping for evaluation of uncertainty in kinetic information derived from dynamic PET
- Author
-
Qi Wu, Finbarr O'Sullivan, David A. Mankoff, Fengyun Gu, Mark Muzi, and Jian Huang
- Subjects
Positron emission tomography ,Standards ,Bootstrapped data ,Computer science ,Numerical models ,01 natural sciences ,030218 nuclear medicine & medical imaging ,Domain (software engineering) ,Data modeling ,Imaging ,010104 statistics & probability ,03 medical and health sciences ,0302 clinical medicine ,PET flow-metabolism imaging study ,Projection domain ,Kinetic information ,Breast cancer patient ,Analytical models ,0101 mathematics ,Medical image processing ,Fitted mixture model ,Image domain ,Projection (set theory) ,Mixture analysis ,Kinetic theory ,Spatial analysis ,Cancer ,Kinetic mapping ,Dynamic PET imaging ,Bootstrapping ,Standard errors ,Data models ,Uncertainty ,Model-based bootstrapping ,4-D PET data ,Mixture model ,Statistical copies ,Statistical analysis ,Model-based bootstrap ,Comprehensive voxel-level analysis ,Mixture models ,Algorithm ,Spatial autocorrelation ,Simulation - Abstract
Kinetic mapping via mixture analysis[8], [10] involves comprehensive voxel-level analysis of dynamic PET data. Bootstrapping from the fitted mixture model gives the ability to directly simulate statistical copies of the 4-D PET data, and following suitable analysis, subsequent simulations of the associated kinetic maps. This gives the ability to numerically evaluate uncertainties in inferences associated with kinetic information. We provide a simple introduction to the concept of the model-based bootstrap and an illustration of the use of the approach for kinetic mapping from dynamic PET using results from recent work in Huang et al.[4]. The illustration is from a PET flow-metabolism imaging study in a breast cancer patient. It involves separate dynamic PET imaging following injections of O-15 H2O and F-18 FDG. The bootstrapped data is created in the image domain rather than the projection domain, so there is no reconstruction requirement involved.
- Published
- 2019
120. Spatial auto-regressive analysis of correlation in 3-D PET with application to model-based simulation of data
- Author
-
Kevin O'Regan, Jian Huang, Finbarr O'Sullivan, and Tian Mou
- Subjects
Normalization (statistics) ,Scanner ,Gamma distribution ,Lung Neoplasms ,Computer science ,computer.software_genre ,Article ,030218 nuclear medicine & medical imaging ,03 medical and health sciences ,0302 clinical medicine ,Imaging, Three-Dimensional ,medicine ,Humans ,Computer Simulation ,Sensitivity (control systems) ,Electrical and Electronic Engineering ,Lung cancer ,Spatial analysis ,Lung ,Bootstrapping (statistics) ,Measure (data warehouse) ,Likelihood Functions ,Radiological and Ultrasound Technology ,Phantoms, Imaging ,Standard errors ,medicine.disease ,Quality assurance ,Computer Science Applications ,PET ,Autoregressive model ,Positron-Emission Tomography ,Model-based bootstrap ,Conditional likelihood ,Iterative EM reconstruction ,Data mining ,computer ,Software ,Simulation ,Spatial autocorrelation - Abstract
When a scanner is installed and begins to be used operationally, its actual performance may deviate somewhat from the predictions made at the design stage. Thus it is recommended that routine quality assurance (QA) measurements be used to provide an operational understanding of scanning properties. While QA data are primarily used to evaluate sensitivity and bias patterns, there is a possibility to also make use of such data sets for a more refined understanding of the 3-D scanning properties. Building on some recent work on analysis of the distributional characteristics of iteratively reconstructed PET data, we construct an auto-regression model for analysis of the 3-D spatial auto-covariance structure of iteratively reconstructed data, after normalization. Appropriate likelihood-based statistical techniques for estimation of the auto-regression model coefficients are described. The fitted model leads to a simple process for approximate simulation of scanner performance—one that is readily implemented in an [Formula: see text] script. The analysis provides a practical mechanism for evaluating the operational error characteristics of iteratively reconstructed PET images. Simulation studies are used for validation. The approach is illustrated on QA data from an operational clinical scanner and numerical phantom data. We also demonstrate the potential for use of these techniques, as a form of model-based bootstrapping, to provide assessments of measurement uncertainties in variables derived from clinical FDG-PET scans. This is illustrated using data from a clinical scan in a lung cancer patient, after a 3-minute acquisition has been re-binned into three consecutive 1-minute time-frames. An uncertainty measure for the tumor SUV(max) value is obtained. The methodology is seen to be practical and could be a useful support for quantitative decision making based on PET data.
- Published
- 2019
121. LMest: Generalized Latent Markov Models for longitudinal continuous and categorical data
- Author
-
Bartolucci, F, Bartolucci, F, Pandolfi, S, Pennoni, F, Farcomeni, A, Serafini, A, PANDOLFI, SARA, Bartolucci, F, Bartolucci, F, Pandolfi, S, Pennoni, F, Farcomeni, A, Serafini, A, and PANDOLFI, SARA
- Abstract
La libreria LMest permette di specificare e stimare i modelli Hidden o Latent Markov (LM) per l'analisi di dati longitudinali sia continui che categoriali. Le covariate sono presenti nei modelli in base ad adeguate parametrizzazioni. Varie tipologie di modelli possono essere stimati utilizzando la struttura dei dati sia in formato long che wide. La stima di massima verosimiglianza è ottenuta con l'algoritmo Expectation-Maximization implementato attraverso routines Fortran. La libreria permette di trattare risposte mancanti, drop-out e valori mancanti secondo una struttura non-monotona. Gli errori standard per i parametri stimati vengono calcolati attraverso la matrice di informazione ottenuta in modo esatto oppure approssimato. La libreria include alcuni esempi e dei dati sia reali che simulati., The package LMest is a framework for specifying and fitting Latent (or Hidden) Markov (LM) models, which are tailored for the analysis of longitudinal continuous and categorical data. Covariates are also included in the model specification through suitable parameterizations. Different LM models are estimated through specific functions requiring a data frame in long format. Maximum likelihood estimation of model parameters is performed through the Expectation-Maximization algorithm, which is implemented by relying on Fortran routines. The package allows us to deal with missing responses, including drop-out and non-monotonic missingness, under the missing-at-random assumption. Standard errors for the parameter estimates are obtained by exact computation of the information matrix or through reliable numerical approximations of this matrix. The package also provides some examples and real and simulated data sets.
- Published
- 2019
122. Variance models of the last age interval and their impact on life expectancy at subnational scales
- Author
-
Robert Bourbeau, Andrea Benedetti, Ernest Lo, and Dan Vatnik
- Subjects
Population ,Interval (mathematics) ,variance ,03 medical and health sciences ,0302 clinical medicine ,Overdispersion ,Statistics ,Econometrics ,030212 general & internal medicine ,delta method ,education ,Demography ,Mathematics ,Statistical hypothesis testing ,Chiang method ,education.field_of_study ,030505 public health ,overdispersion ,Variance (accounting) ,mortality ,health expectancy ,Delta method ,life table ,Standard error ,lcsh:HB848-3697 ,standard errors ,life expectancy ,Life expectancy ,lcsh:Demography. Population. Vital events ,0305 other medical science - Abstract
Background: The Chiang method is the most widely accepted standard for estimating life expectancy (LE) at subnational scales; it is the only method that provides an equation for the LE variance. However, the Chiang variance formula incorrectly omits the contribution of the last age interval. This error is largely unknown to practitioners, and its impact has not been rigorously assessed. Objective: We aim to demonstrate the potentially substantial role of the last age interval on LE variance. We further aim to provide formulae and tools for corrected variance estimation. Methods: The delta method is used to derive variance formulae for a range of variance models of the last age interval. Corrected variances are tested on 291 empirical, abridged life tables drawn from Canadian data (2004-2008) spanning provincial, regional, and intra-regional scales. Results: The last age interval death count can contribute substantially to the LE variance, leading to overestimates of precision and false positives in statistical tests when using the uncorrected Chiang variance. Overdispersion amplifies the contribution while error in population counts has minimal impact. Conclusions: Use of corrected variance formulae is essential for studies that use the Chiang LE. The important role of the last age interval , and hence the life table closure method, on LE variance is demonstrated. These findings extend to other LE-derived metrics such as health expectancy. Contribution: We demonstrate that the last age interval death count can contribute substantially to the LE variance, thus resolving an ambiguity in the scientific literature. We provide heretofore-unavailable formulae for correcting the Chiang LE variance equation.
- Published
- 2016
- Full Text
- View/download PDF
123. The Heavy Burden of "Dependent Children": An Italian Story.
- Author
-
Betti, Gianni, Gagliardi, Francesca, and Neri, Laura
- Abstract
This paper analyses multidimensional fuzzy monetary and non-monetary deprivation in households with children by using two different definitions: households with children under 14 years old, and the EU definition of households with dependent children. Eight dimensions of non-monetary deprivation were found using 34 items from the EU-SILC 2016 survey. Dealing with subpopulations, it is essential to compute standard errors for the presented estimators. Thus, a relevant added value of the paper is fuzzy poverty measures and associated standard errors, which were also computed. Moreover, a comparison was made between the measures obtained concerning the two subpopulations across countries. With a focus on Italy, an Italian macro-region is presented. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
124. Standard errors for the class of orthomax-rotated factor loadings: Some matrix results.
- Author
-
HAYASHI, KENTARO and YIU-FAI YUNG
- Subjects
FACTOR analysis ,MATRICES (Mathematics) ,ERROR ,MATHEMATICAL statistics ,MATHEMATICS problems & exercises - Abstract
The partial derivative matrices of the class of orthomax-rotated factor loadings with respect to the unrotated maximum likelihood factor loadings are derived. The reported results are useful for obtaining standard errors of the orthomax-rotated factor loadings, with or without row normalization (standardization) of the initial factor loading matrix for rotation. Using a numerical example, we verify our analytic formulas by comparing the obtained standard error estimates with that from some existing methods. Some advantages of the current approach are discussed. [ABSTRACT FROM AUTHOR]
- Published
- 1999
- Full Text
- View/download PDF
125. Impact measures in spatial autoregressive models
- Author
-
Arbia, Giuseppe, Anil, Bera, Osman, Dogan, and Suleyman, Taspinar
- Subjects
total effects ,inference ,spatialautoregressivemodels ,Settore SECS-S/03 - STATISTICA ECONOMICA ,direct effects ,standard errors ,impactmeasures ,asymptotic approximation ,spatialeconometricmodels,spatialautoregressivemodels,impactmeasures, asymptotic approximation, standard errors, inference, MLE, direct effects, indirect effects, total effects ,MLE ,spatialeconometricmodels ,indirect effects - Published
- 2019
126. A Bootstrap Method for Conducting Statistical Inference with Clustered Data.
- Author
-
Harden, Jeffrey J.
- Subjects
- *
STATISTICAL models , *MONTE Carlo method , *STATISTICAL correlation , *REGRESSION analysis , *STATISTICAL bias - Abstract
The article discusses a research on statistical analysis of data grouped in clusters as of 2011. The topics discussed include the presence of unmodeled correlations within clusters, bootstrap method that resamples clusters of observation (BCSE),, and estimation of statistical bias in the methods. Also discussed is the estimation of robust cluster standard errors (RCSE).
- Published
- 2011
- Full Text
- View/download PDF
127. Asymptotic Standard Error of Equipercentile Equating.
- Author
-
Liou, Michelle and Cheng, Philip E.
- Abstract
We propose simplified formulas for computing the standard errors of equiper-centile equating for continuous and discrete test scores. The suggested formulas are conceptually simple and easily extended to more complicated equating designs such as chained equipercentile equating, smoothed equipercentile equating, and equating using the frequency estimation method. Results from an empirical study indicate that the derived formulas work reasonably well for samples with moderate sizes (e.g., 1,000 examinees). [ABSTRACT FROM PUBLISHER]
- Published
- 1995
- Full Text
- View/download PDF
128. Standard Errors of Equipercentile Equating for the Common Item Nonequivalent Populations Design.
- Author
-
Jarjoura, David and Kolen, Michael J.
- Abstract
An equating design in which two groups of examinees from slightly different populations are administered different test forms that have a subset of items in common is widely used. A procedure for equipercentile equating under this design has been previously outlined, but standard errors for this rather complex procedure have not been provided. This paper provides these standard errors and a simulation that verifies the equations for large samples. A real data example is provided for considering issues involved in using these procedures. [ABSTRACT FROM PUBLISHER]
- Published
- 1985
- Full Text
- View/download PDF
129. Confidence regions for INDSCAL using the jackknife and bootstrap techniques.
- Author
-
Weinberg, Sharon, Carroll, J., and Cohen, Harvey
- Abstract
Bootstrap and jackknife techniques are used to estimate ellipsoidal confidence regions of group stimulus points derived from INDSCAL. The validity of these estimates is assessed through Monte Carlo analysis. Asymptotic estimates of confidence regions based on a MULTISCALE solution are also evaluated. Our findings suggest that the bootstrap and jackknife techniques may be used to provide statements regarding the accuracy of the relative locations of points in space. Our findings also suggest that MULTISCALE asymptotic estimates of confidence regions based on small samples provide an optimistic view of the actual statistical reliability of the solution. [ABSTRACT FROM AUTHOR]
- Published
- 1984
- Full Text
- View/download PDF
130. A feasible method for standard errors of estimate in maximum likelihood factor analysis.
- Author
-
Jennrich, R. and Clarkson, D.
- Abstract
A jackknife-like procedure is developed for producing standard errors of estimate in maximum likelihood factor analysis. Unlike earlier methods based on information theory, the procedure developed is computationally feasible on larger problems. Unlike earlier methods based on the jackknife, the present procedure is not plagued by the factor alignment problem, the Heywood case problem, or the necessity to jackknife by groups. Standard errors may be produced for rotated and unrotated loading estimates using either orthogonal or oblique rotation as well as for estimates of unique factor variances and common factor correlations. The total cost for larger problems is a small multiple of the square of the number of variables times the number of observations used in the analysis. Examples are given to demonstrate the feasibility of the method. [ABSTRACT FROM AUTHOR]
- Published
- 1980
- Full Text
- View/download PDF
131. A look, by simulation, at the validity of some asymptotic distribution results for rotated loadings.
- Author
-
Archer, Claude and Jennrich, Robert
- Abstract
In the last few years, a number of asymptotic results for the distribution of unrotated and rotated factor loadings have been given. This paper investigates the validity of some of these results based on simulation techniques. In particular, it looks at principal component extraction and quartimax rotation on a problem with 13 variables. The indication is that the asymptotic results are quite good. [ABSTRACT FROM AUTHOR]
- Published
- 1976
- Full Text
- View/download PDF
132. Bootstrapping an Econometric Model: Some Empirical Results.
- Author
-
Freedman, David A. and Peters, Stephen C.
- Subjects
LEAST squares ,ECONOMETRIC models ,ECONOMETRICS ,FORECASTING ,ERROR - Abstract
The bootstrap, like the jackknife, is a technique for estimating standard errors. The idea is to use Monte Carlo simulation, based on a nonparametric estimate of the underlying error distribution. The bootstrap will be applied to an econometric model describing the demand for capital, labor, energy, and materials. The model is fitted by three-stage least squares. In sharp contrast with previous results, the coefficient estimates and the estimated standard errors perform very well. However, the model's forecasts show serious bias and large random errors, significantly understated by the conventional standard error of forecast. ABSTRACT FROM AUTHOR [ABSTRACT FROM AUTHOR]
- Published
- 1984
- Full Text
- View/download PDF
133. Strong consistency of multivariate spectral variance estimators in Markov chain Monte Carlo
- Author
-
Dootika Vats, James M. Flegal, and Galin L. Jones
- Subjects
Statistics and Probability ,Multivariate statistics ,Markov chain ,010102 general mathematics ,Monte Carlo method ,Strong consistency ,Estimator ,Markov chain Monte Carlo ,01 natural sciences ,Normal distribution ,010104 statistics & probability ,symbols.namesake ,spectral methods ,standard errors ,symbols ,Applied mathematics ,Statistics::Methodology ,0101 mathematics ,Monte Carlo ,Mathematics ,Central limit theorem - Abstract
Markov chain Monte Carlo (MCMC) algorithms are used to estimate features of interest of a distribution. The Monte Carlo error in estimation has an asymptotic normal distribution whose multivariate nature has so far been ignored in the MCMC community. We present a class of multivariate spectral variance estimators for the asymptotic covariance matrix in the Markov chain central limit theorem and provide conditions for strong consistency. We examine the finite sample properties of the multivariate spectral variance estimators and its eigenvalues in the context of a vector autoregressive process of order 1.
- Published
- 2018
134. Statistical modeling and analysis of trace element concentrations in forensic glass evidence
- Author
-
Karen Kafadar and Karen D. H. Pan
- Subjects
Statistics and Probability ,Computer science ,error rates ,Population ,01 natural sciences ,Representativeness heuristic ,Polygraph ,03 medical and health sciences ,Lie detection ,0302 clinical medicine ,multivariate lognormal distribution ,Statistics ,030216 legal & forensic medicine ,education ,education.field_of_study ,Robust methods ,covariance matrix ,010401 analytical chemistry ,exploratory data analysis ,Statistical model ,Covariance ,ROC curve ,0104 chemical sciences ,Test (assessment) ,Exploratory data analysis ,Modeling and Simulation ,standard errors ,Statistics, Probability and Uncertainty - Abstract
The question of the validity of procedures used to analyze forensic evidence was raised many years ago by Stephen Fienberg, most notably when he chaired the National Academy of Sciences’ Committee that issued the report The Polygraph and Lie Detection [National Research Council (2003) The National Academies Press]; his role in championing this cause and drawing other statisticians to these issues continued throughout his life. We investigate the validity of three standards related to different test methods for forensic comparison of glass (micro $X$-ray fluorescence ($\mu $-XRF) spectrometry, ICP-MS, LA-ICP-MS], all of which include a series of recommended calculations from which “it may be concluded that [the samples] did not originate from the same source.” Using publicly available data and data from other sources, we develop statistical models based on estimates of means and covariance matrices of the measured trace element concentrations recommended in these standards, leading to population-based estimates of error rates for the comparison procedures stated in the standards. Our results therefore do not depend on internal comparisons between pairs of glass samples, the representativeness of which cannot be guaranteed: our results apply to any collection of glass samples that have been or can be measured via these technologies. They suggest potentially higher false positive rates than have been reported, and we propose alternative methods that will ensure lower error rates.
- Published
- 2018
135. Analysis of longitudinal data with intermittent missing values using the stochastic EM algorithm
- Author
-
Gad, Ahmed M. and Ahmed, Abeer S.
- Subjects
- *
MONTE Carlo method , *ALGORITHMS , *DATA analysis , *STOCHASTIC processes - Abstract
Abstract: Longitudinal data are not uncommon in many disciplines where repeated measurements on a response variable are collected for all subjects. Some intended measurements may not be available for some subjects resulting in a missing data pattern. Dropout pattern occurs when some subjects leave the study prematurely. The missing data pattern is defined as intermittent if a missing value followed by an observed value. When the probability of missingness depends on the missing value, and may be on the observed values, the missing data mechanism is termed as nonrandom. Ignoring the missing values in this case leads to biased inferences. The stochastic EM (SEM) algorithm is proposed and developed to find parameters estimates in the presence of intermittent missing values. Also, in this setting, the Monte Carlo method is developed to find the standard errors of parameters estimates. Finally, the proposed techniques are applied to a real data from the International Breast Cancer Study Group. [Copyright &y& Elsevier]
- Published
- 2006
- Full Text
- View/download PDF
136. Approaches to sample size determination for multivariate data : Applications to PCA and PLS-DA of omics data
- Author
-
Edoardo Saccenti, Marieke E. Timmerman, and Psychometrics and Statistics
- Subjects
0301 basic medicine ,Serum ,loading estimation ,Multivariate statistics ,METABOLOMICS DATA ,STOPPING RULES ,covariance estimation ,Multivariate analysis ,Swine ,STANDARD ERRORS ,power analysis ,Urine ,random matrix theory ,computer.software_genre ,dimensionality ,Biochemistry ,PRINCIPAL-COMPONENTS-ANALYSIS ,CONFIDENCE-INTERVALS ,03 medical and health sciences ,Estimation of covariance matrices ,Multivariate analysis of variance ,STATISTICAL POWER ,LARGEST EIGENVALUE ,eigenvalue distribution ,hypothesis testing ,Animals ,Humans ,Metabolomics ,Statistics::Methodology ,Systems and Synthetic Biology ,Least-Squares Analysis ,CROSS-VALIDATION ,Mathematics ,VLAG ,Principal Component Analysis ,Systeem en Synthetische Biologie ,CARDIOVASCULAR RISK ,Sparse PCA ,Discriminant Analysis ,General Chemistry ,Linear discriminant analysis ,030104 developmental biology ,multivariate analysis ,Sample size determination ,Sample Size ,Principal component analysis ,COVARIANCE MATRICES ,Data mining ,computer - Abstract
Sample size determination is a fundamental step in the design of experiments. Methods for sample size determination are abundant for univariate analysis methods, but scarce in the multivariate case. Omics data are multivariate in nature and are commonly investigated using multivariate statistical methods, such as principal component analysis (PCA) and partial least-squares discriminant analysis (PLS-DA). No simple approaches to sample size determination exist for PCA and PLS-DA. In this paper we will introduce important concepts and offer strategies for (minimally) required sample size estimation when planning experiments to be analyzed using PCA and/or PLS-DA.
- Published
- 2016
- Full Text
- View/download PDF
137. Inference in Linear Regression Models with Many Covariates and Heteroscedasticity
- Author
-
Cattaneo, MD, Cattaneo, MD, Jansson, M, Newey, WK, Cattaneo, MD, Cattaneo, MD, Jansson, M, and Newey, WK
- Abstract
The linear regression model is widely used in empirical work in economics, statistics, and many other disciplines. Researchers often include many covariates in their linear model specification in an attempt to control for confounders. We give inference methods that allow for many covariates and heteroscedasticity. Our results are obtained using high-dimensional approximations, where the number of included covariates is allowed to grow as fast as the sample size. We find that all of the usual versions of Eicker–White heteroscedasticity consistent standard error estimators for linear models are inconsistent under this asymptotics. We then propose a new heteroscedasticity consistent standard error formula that is fully automatic and robust to both (conditional) heteroscedasticity of unknown form and the inclusion of possibly many covariates. We apply our findings to three settings: parametric linear models with many covariates, linear panel models with many fixed effects, and semiparametric semi-linear models with many technical regressors. Simulation evidence consistent with our theoretical results is provided, and the proposed methods are also illustrated with an empirical application. Supplementary materials for this article are available online.
- Published
- 2018
138. Selection of Tuning Parameters, Solution Paths and Standard Errors for Bayesian Lassos
- Author
-
Sounak Chakraborty and Vivekananda Roy
- Subjects
Statistics and Probability ,Elastic net regularization ,Feature selection ,geometric ergodicity ,01 natural sciences ,010104 statistics & probability ,symbols.namesake ,62J07 ,60J05 ,Lasso (statistics) ,Frequentist inference ,0502 economics and business ,Statistics::Methodology ,0101 mathematics ,050205 econometrics ,Mathematics ,Markov chain ,business.industry ,Applied Mathematics ,05 social sciences ,Pattern recognition ,Markov chain Monte Carlo ,elastic net ,Statistics::Computation ,Bayesian lasso ,importance sampling ,shrinkage ,standard errors ,symbols ,Artificial intelligence ,62F15 ,business ,Algorithm ,Importance sampling ,empirical Bayes ,Gibbs sampling - Abstract
Penalized regression methods such as the lasso and elastic net (EN) have become popular for simultaneous variable selection and coefficient estimation. Implementation of these methods require selection of the penalty parameters. We propose an empirical Bayes (EB) methodology for selecting these tuning parameters as well as computation of the regularization path plots. The EB method does not suffer from the “double shrinkage problem” of frequentist EN. Also it avoids the difficulty of constructing an appropriate prior on the penalty parameters. The EB methodology is implemented by efficient importance sampling method based on multiple Gibbs sampler chains. Since the Markov chains underlying the Gibbs sampler are proved to be geometrically ergodic, Markov chain central limit theorem can be used to provide asymptotically valid confidence band for profiles of EN coefficients. The practical effectiveness of our method is illustrated by several simulation examples and two real life case studies. Although this article considers lasso and EN for brevity, the proposed EB method is general and can be used to select shrinkage parameters in other regularization methods.
- Published
- 2017
139. ESTIMATING AND TESTING BIOCONCENTRATION FACTORS.
- Author
-
Bailer, A. John, Walker, Sean E., and Venis, Kyle J.
- Subjects
- *
CHEMICAL equilibrium , *HYALELLA , *DDT (Insecticide) , *CONFIDENCE intervals , *STATISTICS - Abstract
Bioconcentration factors (BCFs) are commonly calculated to represent the equilibrium concentration of a substance in an organism relative to environmental concentrations of the same substance. The BCF is derived from parameters estimated in uptake and elimination experiments and is presented as a single value without error estimates or confidence intervals. However, it is desirable to know the variability/precision of the BCF estimate and to statistically compare BCFs among experimental conditions. In this study, the calculation of standard errors and confidence intervals for BCFs is presented. In addition, a statistical method for formally comparing the BCFs derived under two or more experimental conditions is discussed. These methods are illustrated using data from a study of DDT-exposed Hyalella azteca and Diporeia spp. [ABSTRACT FROM AUTHOR]
- Published
- 2000
- Full Text
- View/download PDF
140. Estimation of standard errors and treatment effects in empirical economics—methods and applications
- Author
-
Hübler, Olaf
- Published
- 2014
- Full Text
- View/download PDF
141. Information matrix for hidden Markov models with covariates
- Author
-
Alessio Farcomeni and Francesco Bartolucci
- Subjects
Statistics and Probability ,Mathematical optimization ,Basis (linear algebra) ,Posterior probability ,em algorithm ,Theoretical Computer Science ,symbols.namesake ,Observed information ,Computational Theory and Mathematics ,standard errors ,Expectation–maximization algorithm ,symbols ,Identifiability ,oakes’ identity ,Forward algorithm ,forward-backward recursions ,Statistics, Probability and Uncertainty ,Settore SECS-S/01 - Statistica ,Fisher information ,Hidden Markov model ,Algorithm ,Mathematics - Abstract
For a general class of hidden Markov models that may include time-varying covariates, we illustrate how to compute the observed information matrix, which may be used to obtain standard errors for the parameter estimates and check model identifiability. The proposed method is based on the Oakes' identity and, as such, it allows for the exact computation of the information matrix on the basis of the output of the expectation-maximization (EM) algorithm for maximum likelihood estimation. In addition to this output, the method requires the first derivative of the posterior probabilities computed by the forward-backward recursions introduced by Baum and Welch. Alternative methods for computing exactly the observed information matrix require, instead, to differentiate twice the forward recursion used to compute the model likelihood, with a greater additional effort with respect to the EM algorithm. The proposed method is illustrated by a series of simulations and an application based on a longitudinal dataset in Health Economics.
- Published
- 2014
- Full Text
- View/download PDF
142. Asymptotic Standard Errors of Parameter Scale Transformation Coefficients in Test Equating Under the Nominal Response Model.
- Author
-
Zhang Z
- Abstract
Researchers have developed a characteristic curve procedure to estimate the parameter scale transformation coefficients in test equating under the nominal response model. In the study, the delta method was applied to derive the standard error expressions for computing the standard errors for the estimates of the parameter scale transformation coefficients. This brief report presents the results of a simulation study that examined the accuracy of the derived formulas and compared the performance of this analytical method with that of the multiple imputation method. The results indicated that the standard errors produced by the delta method were very close to the criterion standard errors as well as those yielded by the multiple imputation method under all the simulation conditions., Competing Interests: Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article., (© The Author(s) 2020.)
- Published
- 2021
- Full Text
- View/download PDF
143. Multiple Equating of Separate IRT Calibrations
- Author
-
Michela Battauz
- Subjects
Psychometrics ,Scale (ratio) ,equating coefficients ,Generalization ,01 natural sciences ,Haebara ,item response theory ,linking ,mean-geometric mean ,mean-mean ,standard errors ,Stocking–Lord ,010104 statistics & probability ,0504 sociology ,Statistics ,Equating ,Item response theory ,Applied mathematics ,0101 mathematics ,General Psychology ,Mathematics ,Applied Mathematics ,05 social sciences ,050401 social sciences methods ,Equating coefficients ,Measurement scales ,Standard error - Abstract
When test forms are calibrated separately, item response theory parameters are not comparable because they are expressed on different measurement scales. The equating process includes the conversion of item parameter estimates on a common scale and the determination of comparable test scores. Various statistical methods have been proposed to perform equating between two test forms. This paper provides a generalization to multiple test forms of the mean-geometric mean, the mean-mean, the Haebara, and the Stocking–Lord methods. The proposed methods estimate simultaneously the equating coefficients that permit the scale transformation of the parameters of all forms to the scale of the base form. Asymptotic standard errors of the equating coefficients are derived. A simulation study is presented to illustrate the performance of the methods.
- Published
- 2017
144. Two-Level Designs to Estimate All Main Effects and Two-Factor Interactions
- Subjects
Optimal design ,Coordinate exchange ,Two-level designs ,Design ,Research ,D-optimal designs ,Standard errors ,RAPID - Risk Analysis for Products in Development ,Partial enumerations ,Ds-efficiency ,Optimal systems ,Efficiency ,ELSS - Earth ,Product design ,Life ,Orthogonal array ,Interaction model ,Life and Social Sciences ,D-efficiency ,Healthy for Life ,Healthy Living ,Partial enumeration - Abstract
We study the design of two-level experiments with N runs and n factors large enough to estimate the interaction model, which contains all the main effects and all the two-factor interactions. Yet, an effect hierarchy assumption suggests that main effect estimation should be given more prominence than the estimation of two-factor interactions. Orthogonal arrays (OAs) favor main effect estimation. However, complete enumeration becomes infeasible for cases relevant for practitioners. We develop a partial enumeration procedure for these cases and we establish upper bounds on the D-efficiency for the interaction model based on arrays that have not been generated by the partial enumeration. We also propose an optimal design procedure that favors main effect estimation. Designs created with this procedure have smaller D-efficiencies for the interaction model than D-optimal designs, but standard errors for the main effects in this model are improved. Generated OAs for 7–10 factors and 32–72 runs are smaller or have a higher D-efficiency than the smallest OAs from the literature. Designs obtained with the new optimal design procedure or strength-3 OAs (which have main effects that are not correlated with two-factor interactions) are recommended if main effects unbiased by possible two-factor interactions are of primary interest. D-optimal designs are recommended if interactions are of primary interest. Supplementary materials for this article are available online. © 2017 American Statistical Association and the American Society for Quality.
- Published
- 2017
145. Dummy variables and biases of allometric models when local estimating tree biomass (on an example of Picea L.)
- Subjects
PICEA L ,СТАНДАРТНЫЕ И СИСТЕМАТИЧЕСКИЕ ОШИБКИ ,ALLOMETRIC MODELS ,SAMPLE PLOTS ,ФИТОМАССА ДЕРЕВА ,TREE BIOMASS ,АЛЛОМЕТРИЧЕСКИЕ МОДЕЛИ ,STANDARD ERRORS ,BIASES ,РЕГИОНАЛЬНЫЕ РАЗЛИЧИЯ ,REGIONAL DIFFERENCES ,ПРОБНЫЕ ПЛОЩАДИ - Abstract
Леса играют важную роль в снижении количества парниковых газов в атмосфере и предотвращении изменения климата. Одним из способов количественной оценки углеродного обмена в лесном покрове является определение изменений в запасах его фитомассы и углерода со временем. Запас фитомассы на единице площади начинается с определения его на уровне отдельных деревьев. Известно строгое и устойчивое аллометрическое соотношение между фитомассой дерева и его диаметром (простая аллометрия), или между фитомассой дерева и несколькими массообразующими (морфометрическими) показателями (многофакторная аллометрия). В настоящее время в разных странах и континентах проводятся интенсивные исследования применимости так называемых «всеобщих» аллометрических моделей (generic, generalized, common models), которые обеспечивали бы аллометрической модели приемлемую точность при оценке фитомассы насаждений. В статье на основе сформированной базы данных о фитомассе деревьев Picea в количестве 1065 определений построены аллометрические модели четырёх видов, включающие в себя фиктивные переменные, которые дают возможность дать региональные оценки их фитомассы по известным морфометрическим показателям (диаметр ствола и кроны, высота дерева). Предложенные аллометрические модели свидетельствуют об их адекватности фактическим данным (коэффициент детерминации от 0,959 до 0,984) и могут применяться при региональных оценках фитомассы деревьев ели. Однако всеобщие аллометрические модели, построенные по всему массиву фактических данных, дают в экорегионах слишком большие стандартные ошибки (до 402%) и неприемлемые смещения обоих знаков (от +311 до -86%), что исключает воз-можность их применения на региональных уровнях. Forests play an important role in reducing the amount of greenhouse gases in the atmosphere and preventing climate change. One way to quantify сarbon exchange in forest cover is estimating changes in its biomass and carbon pools over time. Biomass estimating on the unit of area starts with harvesting sample trees and weighing their biomass. It is known the strong and sustainable relationship between tree biomass and its diameter (simple allometry), or between tree biomass and a number of mass-forming (morphometric) indices (multi-factor allometry). At present, in different countries and continents, the studies of the applicability of the so-called generic (generalized, common) allometric models are intensified that would give acceptable accuracy in estimating forest biomass. In the article on the basis of the compiled database of tree biomass of Picea at a number of 1065 trees, allometric models of the four modifications are designed, which include the block of independent dummy variables. These models provide an opportunity to give regional estimates of tree biomass when using some known mass-forming indices (stem and crown diameter and tree height). Allometric models proposed are indicative of their adequacy for the actual data (coefficients of determination are 0,959 to 0,984) and can be applied in regional estimating of spruce tree biomass. However, generic allometric models built using the total quantity of actual data give in different ecoregions too large standard errors (up to 402%) and unacceptable both positive and negative biases (from + 311 to -86%), that excludes any possibility of their application at regional levels
- Published
- 2017
146. Contributions to Machine Learning and Psychometrics
- Author
-
Philipp, Michel, University of Zurich, and Philipp, Michel
- Subjects
recursive partitioning ,cutpoint selection ,Wald test ,score test ,decision trees ,10093 Institute of Psychology ,stability ,R package stablelearner ,DINA model ,DoktoratPsych Erstautor ,information matrix ,resampling ,cognitive diagnosis model ,DINA ,UZHDISS UZH Dissertations ,standard errors ,differential item functioning ,150 Psychology ,variable selection ,Lagrange multiplier test - Published
- 2017
- Full Text
- View/download PDF
147. Biases of generic species-specific allometric models when local estimating tree biomass of firs and 2- or 5-needled pines (Abies Mill., Pinus sylvestris L., Pinus sibirica Du Tour)
- Subjects
СТАНДАРТНЫЕ И СИСТЕМАТИЧЕСКИЕ ОШИБКИ ,ПИХТЫ ,2- OR 5-NEEDLED PINES ,ALLOMETRIC MODELS ,SAMPLE PLOTS ,TREE BIOMASS ,STANDARD ERRORS ,BIASES ,РЕГИОНАЛЬНЫЕ РАЗЛИЧИЯ ,REGIONAL DIFFERENCES ,ДВУХ- И ПЯТИХВОЙНЫЕ СОСНЫ ,ФИТОМАССА ДЕРЕВА ,АЛЛОМЕТРИЧЕСКИЕ МОДЕЛИ ,ПРОБНЫЕ ПЛОЩАДИ ,FIRS - Abstract
На основе сформированной базы данных о биомассе деревьев двух- и пятихвойных сосен и пихт в количестве 1234 определений построены аллометрические модели четырёх структурных форм, включающие в себя фиктивные переменные, которые дают возможность дать региональные оценки их надземной биомассы по известным морфометрическим показателям (диаметр ствола и кроны, высота дерева). Предложенные аллометрические модели свидетельствуют об их адекватности фактическим данным с коэффициентом детерминации от 0,725 до 0,990 (исключение составила зависимость надземной биомассы от диаметра кроны пихты с коэффициентом детерминации 0,430) и могут применяться при региональных оценках биомассы деревьев трёх древесных пород. Однако всеобщие аллометрические модели, построенные по всему массиву фактических данных, дают в экорегионах слишком большие стандартные ошибки (до 572%) и неприемлемые смещения обоих знаков (от +315 до -92%), что исключает возможность их применения на региональных уровнях. Оn the basis of the compiled database of tree biomass of 2- or 5-needled pines and firs at a number of 1234 trees, allometric models of the four modifications are designed, which include the block of independent dummy variables. These models provide an opportunity to give regional estimates of tree above-ground biomass when using some known mass-forming indices (stem and crown diameter and tree height). Allometric models proposed are indicative of their adequacy for the actual data having the coefficients of determination between 0,725 and 0,990 (the only exception was the dependence of fir biomass upon crown diameter with a coefficient of determination 0,430) and can be applied in regional estimating above-ground biomass of the above-mentioned tree species. However, generic allometric models built using the total quantity of actual data give in different ecoregions too large standard errors (up to 572%) and unacceptable both positive and negative biases (from + 315 to -92%), that excludes any possibility of their application at regional levels.
- Published
- 2017
148. Inference in linear regression models with many covariates and heteroskedasticity
- Author
-
Cattaneo, Matias D., Jansson, Michael, and Newey, Whitney K.
- Subjects
standard errors ,many regressors ,ddc:330 ,linear regression ,Mathematics::Metric Geometry ,Statistics::Methodology ,heteroskedasticity ,high-dimensional models - Abstract
The linear regression model is widely used in empirical work in Economics, Statistics, and many other disciplines. Researchers often include many covariates in their linear model specification in an attempt to control for confounders. We give inference methods that allow for many covariates and heteroskedasticity. Our results are obtained using high-dimensional approximations, where the number of included covariates are allowed to grow as fast as the sample size. We find that all of the usual versions of Eicker-White heteroskedasticity consistent standard error estimators for linear models are inconsistent under this asymptotics. We then propose a new heteroskedasticity consistent standard error formula that is fully automatic and robust to both (conditional) heteroskedasticity of unknown form and the inclusion of possibly many covariates. We apply our findings to three settings: parametric linear models with many covariates, linear panel models with many fixed effects, and semiparametric semi-linear models with many technical regressors. Simulation evidence consistent with our theoretical results is also provided. The proposed methods are also illustrated with an empirical application.
- Published
- 2017
149. Sensitivity of Econometric Estimates to Item Non-response Adjustment
- Author
-
Sanchez, Juana
- Subjects
Item non-response ,LBD ,multivariate analysis ,non-sampling errors ,multiple imputation ,BRDIS ,standard errors ,business establishment survey data ,Physical Sciences and Mathematics ,statistical models ,unit non-response - Abstract
Non-response in establishment surveys is a very important problem that can bias results of statistical analysis. The bias can be considerable when the survey data is used to do multivariate analysis that involve several variables with different response rates, which can reduce the effective sample size considerably. Fixing the non-response, however, could potentially cause other econometric problems. This paper uses an operational approach to analyze the sensitivity of results of multivariate analysis to multiple imputation procedures applied to the U.S. Census Bureau/NSF‘s Business Research and Development and Innovation Survey (BRDIS) to address item non-response. Multiple imputation is first applied using data from all survey units and periods for which there is data, presenting scenario 1. A scenario 2 involves separate imputation for units that have participated in the survey only once and those that repeat. Scenario 3 involves no imputation. Sensitivity analysis is done by comparing the model estimates and their standard errors, and measures of the additional uncertainty created by the imputation procedure. In all cases, unit non-response is addressed by using the adjusted weights that accompany BRDIS micro data. The results suggest that substantial benefit may be derived from multiple imputation, not only because it helps provide more accurate measures of the uncertainty due to item non-response but also because it provides alternative estimates of effect sizes and population totals.
- Published
- 2016
150. The Cost of Fear : The Welfare Effect of the Risk of Violence in Northern Uganda
- Author
-
Rockmore, Marc
- Subjects
OBSERVER ,INFORMATION ,CITIES ,RIGHTS ,MEASUREMENT ,ROAD ,DESIGN ,SECURITY FORCES ,EMPLOYMENT ,MONITORING ,MEETINGS ,ASSOCIATIONS ,SAFETY NETS ,INCOME ,CIVIL SOCIETY ,OUTCOMES ,TERRORISM ,PRODUCTIVITY ,CIVIL WAR ,VIOLENT CONFLICT ,CRIME ,TACTICS ,STATISTICS ,FEMALE ,COUNTERTERRORISM ,REBELS ,HEALTH ,RENT ,FORCED LABOR ,WAR ,INTERVENTION ,VIOLENCE ,INTERVENTIONS ,ORGANIZATIONS ,LABOR SUPPLY ,STANDARD ERRORS ,PEACE RESEARCH ,DEVELOPMENT ECONOMICS ,MARKETS ,ECONOMIC COSTS ,ARMED CONFLICT ,PEACE ,CALCULATION ,PRICES ,RURAL AREAS ,RECONSTRUCTION ,PRODUCTION ,LABOR MARKET ,HOUSE PRICES ,DISTRICTS ,HOUSEHOLD ,SAMPLING ,CONSUMPTION ,SERVICES ,HOUSEHOLD COMPOSITION ,WARS ,RISKS ,GRANT ,PRECISION ,MARKET ,FACTORS ,MOBILITY ,SUPPLY ,HUMAN RIGHTS ,ECONOMIC CONDITIONS ,FEMALES ,REBEL ,ECONOMIC DEVELOPMENT ,CRISES ,DATA ,MILITARY EXPENDITURE ,CYCLE OF VIOLENCE ,DESCRIPTIVE STATISTICS ,GENOCIDE ,VARIABLES ,LABOUR ,BOOTSTRAP ,LABOR ALLOCATION ,ROADS ,ACCOUNTING ,PORTFOLIOS ,VALUE ,SECURITY ,RISK ,ECONOMIES ,FOOD SECURITY ,DEATH ,BOUNDARIES ,STANDARD DEVIATION ,VILLAGES ,POLICE ,COMMUNITY ,PROBABILITY ,YOUTH ,HUMAN CAPITAL ,SAFETY ,EFFECTS ,INSURANCE ,HOUSEHOLDS ,PROJECT ,RURAL COMMUNITIES ,FEMALE HEADED HOUSEHOLDS ,INDEPENDENT VARIABLES ,GRANTS ,ECONOMY ,ERRORS ,CREDIT ,FOOD AID ,INTERNATIONAL BANK ,EXPERTS ,ELECTIONS ,CONFLICTS ,LABOR ,POLITICS ,CONFLICT ,HOMES ,ECONOMICS ,MOTIVATION ,BIASES ,REFUGEE ,DISPLACED PERSONS ,URBAN AREAS ,WORLD DEVELOPMENT ,UNIVERSITY ,GENDER ,COMMUNITIES ,LAW - Abstract
Although the effects of insecurity are believed to be important, these have never been directly measured. Previous estimates of the costs of conflict have only captured the joint effect of violence and insecurity. The distinction is important for understanding the origins of the costs and for policy design. Spatially disaggregated measures of insecurity are created based on the spatial-temporal variation in the placement of violence. These are used to generate the first estimates of the relative causal contributions of the risk and experience of violence. The article also provides the first micro-data based counterpart to the cross-country literature on the costs of conflict.
- Published
- 2016
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.