1,210 results on '"Grouped data"'
Search Results
202. Statistical estimation for the parameters of Weibull distribution based on progressively type-I interval censored sample.
- Author
-
Ng, Hon Keung Tony and Wang, Zhu
- Subjects
- *
PROBABILITY theory , *ESTIMATION theory , *MONTE Carlo method , *DISTRIBUTION (Probability theory) , *MATHEMATICAL combinations - Abstract
In this paper, the estimation of parameters based on a progressively type-I interval censored sample from a two-parameter Weibull distribution is studied. Different methods of estimation are discussed. They include the maximum likelihood estimation, method of moments, estimation based on Weibull probability plot, mid-point approximation method and one-step approximation method. The estimation procedures are discussed in details and compared via Monte Carlo simulations in terms of their biases and mean square errors. Some recommendations are made from the simulation results and a numerical example is presented to illustrate all of the methods of estimation developed here. [ABSTRACT FROM AUTHOR]
- Published
- 2009
- Full Text
- View/download PDF
203. Run and Frequency Quota Rules in Process Monitoring and Acceptance Sampling.
- Author
-
BALAKRISHNAN, N., BERSIMIS, S., and KOUTRAS, M. V.
- Subjects
ACCEPTANCE sampling ,STATISTICAL sampling ,CONTROL theory (Engineering) ,MEASURING instruments ,FREQUENCIES of oscillating systems ,DISTRIBUTION (Probability theory) - Abstract
Motivated by certain models arising in the field of statistical process control and acceptance sampling, we study the waiting-time distribution for the first occurrence of a pattern belonging to a specific family of patterns in a bivariate sequence of trinomial trials. Such a setup can be applied in cases where the final quality of the output of an industrial process is determined by two correlated variables that are used to classify the two measured characteristics as "quite close to target quality," "of satisfactory quality," and "far from target quality." A typical case where this happens is whenever registering the exact value of the characteristic of interest is expensive or difficult (e.g., when the observations come in the form of grouped data, due to the use of m-step gauges). The exact distribution of the waiting time is derived by using a Markov chain-embedding technique, which is then used in order to determine the most appropriate design for the quality-control problems discussed. [ABSTRACT FROM AUTHOR]
- Published
- 2009
- Full Text
- View/download PDF
204. Planning of progressive group-censoring life tests with cost considerations.
- Author
-
Wu, Shuo-Jye, Chang, Chun-Tao, Liao, Kang-Jun, and Huang, Syuan-Rong
- Subjects
- *
ESTIMATION theory , *ALGORITHMS , *WEIBULL distribution , *DISTRIBUTION (Probability theory) , *REAL life test (Gender transition) - Abstract
This paper considers a life test under progressive type I group censoring with a Weibull failure time distribution. The maximum likelihood method is used to derive the estimators of the parameters of the failure time distribution. In practice, several variables, such as the number of test units, the number of inspections, and the length of inspection interval are related to the precision of estimation and the cost of experiment. An inappropriate setting of these decision variables not only wastes the resources of the experiment but also reduces the precision of estimation. One problem arising from designing a life test is the restricted budget of experiment. Therefore, under the constraint that the total cost of experiment does not exceed a pre-determined budget, this paper provides an algorithm to solve the optimal decision variables by considering three different criteria. An example is discussed to illustrate the proposed method. The sensitivity analysis is also studied. [ABSTRACT FROM AUTHOR]
- Published
- 2008
- Full Text
- View/download PDF
205. Semiparametric Analysis With Grouped Dependent Variables and Application to Physicians' Provision of Charity Care.
- Author
-
DAS, Mitali
- Subjects
REGRESSION analysis ,MATHEMATICAL statistics ,MULTIVARIATE analysis ,ESTIMATION theory ,ASYMPTOTIC theory in estimation theory ,STOCHASTIC processes - Abstract
We present estimators for semiparametric regression models where the dependent variable is grouped, that is, known to fall in a specified group with observable thresholds while its true value remains latent. Income, weeks unemployed, and treatment length are examples of such variables. Because the model is not amenable to direct estimation, estimators are derived from a transform in which the index emerges as the partially linear component in a vector of identities. √n asymptotic normality of the proposed estimator is derived. The analytical results are applied to study physicians' provision of charity care, using data in which charity care is grouped. [ABSTRACT FROM AUTHOR]
- Published
- 2008
- Full Text
- View/download PDF
206. Techniques for controlling bivariate grouped observations
- Author
-
Koutras, M.V., Maravelakis, P.E., and Bersimis, S.
- Subjects
- *
DISTRIBUTION (Probability theory) , *MARKOV processes , *MATHEMATICAL variables , *METHODOLOGY - Abstract
Abstract: The term grouped data refers to the case where the exact value of the characteristic of interest is either unknown or difficult to register. In the present article we study a model that can be used for the simultaneous control of two (possibly correlated) variables whose values have been registered in the form of grouped data. The exact distribution of the waiting time for an out of control signal through the suggested scheme and its ARL are investigated by using a Markov Chain embedding methodology and by establishing a recurrence scheme for the respective tail probabilities. A detailed study of the performance of the scheme is also carried out when the characteristics of interest follow the Marsall–Olkin’s bivariate Exponential Distribution. [Copyright &y& Elsevier]
- Published
- 2008
- Full Text
- View/download PDF
207. A note on the parameter estimation for the lognormal distribution based on progressively Type I interval censored samples.
- Author
-
Amin, Zeinab H.
- Subjects
LOGNORMAL distribution ,ESTIMATION theory ,BAYES' estimation ,DISTRIBUTION (Probability theory) ,PROBABILITY theory - Abstract
Estimation concerning the parameters of the lognormal distribution are discussed for progressively Type I interval censored samples. Maximum likelihood estimators as well as approximate Bayes estimators of the parameters are developed. The approximate asymptotic variance covariance matrix of the maximum likelihood estimates is given and the posterior risks for the Bayes estimates are derived. Illustrative examples involving real and simulated data sets show that the Bayes estimates provide more consistent results than the maximum likelihood estimates for the data discussed. [ABSTRACT FROM AUTHOR]
- Published
- 2008
- Full Text
- View/download PDF
208. THE INSPECTION OF ACCEPTANCE SAMPLING FOR STEP-STRESS TESTS WITH AN EQUALLY-SPACED INTERVAL CENSORING SCHEME.
- Author
-
TZONG-RUTSAI, HSIN-HAO CHEN, and WANBO LU
- Subjects
STATISTICAL sampling ,RAYLEIGH number ,ALGORITHMS ,NUMERICAL analysis ,PROCESS control systems - Abstract
For reducing the experimental time and considering the purpose of administrative convenience, the article establishes an acceptance sampling procedure for the Rayleigh lifetime distribution under a step-stress test with an equally-spaced interval censoring scheme. Both producer and consumer risks are considered and an algorithm is provided to develop the ordinary acceptance sampling plans. A numerical study is conducted for evaluating the performance of the proposed method. Moreover, a numerical example is used for illustration. [ABSTRACT FROM AUTHOR]
- Published
- 2008
- Full Text
- View/download PDF
209. Optimal step-stress test under type I progressive group-censoring with random removals
- Author
-
Wu, Shuo-Jye, Lin, Ying-Po, and Chen, Shyi-Tien
- Subjects
- *
DISTRIBUTION (Probability theory) , *MATHEMATICAL statistics , *EIGENVALUES , *MATRICES (Mathematics) , *FAILURE time data analysis - Abstract
Abstract: Some traditional life tests result in no or very few failures by the end of test. In such cases, one approach is to do life testing at higher-than-usual stress conditions in order to obtain failures quickly. This paper discusses a k-level step-stress accelerated life test under type I progressive group-censoring with random removals. An exponential failure time distribution with mean life that is a log-linear function of stress and a cumulative exposure model are considered. We derive the maximum likelihood estimators of the model parameters and establish the asymptotic properties of the estimators. We investigate four selection criteria which enable us to obtain the optimum test plans. One is to minimize the asymptotic variance of the maximum likelihood estimator of the logarithm of the mean lifetime at use-condition, and the other three criteria are to maximize the determinant, trace and the smallest eigenvalue of Fisher''s information matrix. Some numerical studies are discussed to illustrate the proposed criteria. [Copyright &y& Elsevier]
- Published
- 2008
- Full Text
- View/download PDF
210. CALCULATING THE EXTENDED GINI COEFFICIENT FROM GROUPED DATA - A COVARIANCE PRESENTATION.
- Author
-
Schechtman, Edna and Yitzhaki, Shlomo
- Subjects
GINI coefficient ,DATA ,LORENZ curve ,ANALYSIS of covariance ,MATHEMATICAL variables ,MATHEMATICAL statistics - Abstract
The basic approach to estimating the Gini and extended Gini indices is to approximate the Lorenz curve by a number of linear segments, and then estimate the Gini coefficients as the areas (or weighted areas) between the linear segments and the 45-degree line. We show that the estimator for the extended Gini, obtained from the Lorenz curve (Chotikapanich and Griffiths, 2001) is algebraically identical to a covariance-based estimator. The advantages of the covariance-based estimators are twofold; first, they are easy to compute, using any standard statistical software, and second, the covariance-based estimators allow for the decomposition of the Gini index of a sum of variables (or populations). [ABSTRACT FROM AUTHOR]
- Published
- 2008
211. The Estimation of Conditional Distributions From Large Databases.
- Author
-
Ord, J. Keith and Iglarsh, Harvey J.
- Subjects
ESTIMATION theory ,CONDITIONAL expectations ,DATABASES ,THEORY of distributions (Functional analysis) ,INFORMATION storage & retrieval systems ,HOME prices ,INCOME tax deductions ,WAGE taxation ,PAYMENT - Abstract
Modern databases enable the estimation of conditional distributions for Y given X in a specified interval. Confidentiality requirements typically mean that individual observations cannot be released and output may be restricted to Y values for X within specified intervals. The width of such an interval could have a major impact upon the quality of the estimates. In this article, we develop an improved estimation procedure to reduce the impact of interval width, using a new variable whose distribution does not vary over the X interval in question. The method is illustrated using data on housing prices and income tax payment. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
212. Degeneracy in the Maximum Likelihood Estimation of Univariate Gaussian Mixtures for Grouped Data and Behaviour of the EM Algorithm.
- Author
-
Biernacki, Christophe
- Subjects
- *
GAUSSIAN measures , *MAXIMUM likelihood statistics , *EXPECTATION-maximization algorithms , *ALGORITHMS , *STOCHASTIC processes , *DIRAC equation , *PARTIAL differential equations , *QUANTUM field theory , *STOCHASTIC convergence - Abstract
In the context of the univariate Gaussian mixture with grouped data, it is shown that the global maximum of the likelihood may correspond to a situation where a Dirac lies in any non-empty interval. Existence of a domain of attraction near such a maximizer is discussed and we establish that the expectation-maximization (EM) iterates move extremely slowly inside this domain. These theoretical results are illustrated both by some Monte-Carlo experiments and by a real data set. To help practitioners identify and discard these potentially dangerous degenerate maximizers, a specific stopping rule for EM is proposed. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
213. Tests for the Validity of the Assumption That the Underlying Distribution of Life Is Pareto.
- Author
-
Amin, ZeinabH.
- Abstract
This article considers the problem of testing the validity of the assumption that the underlying distribution of life is Pareto. For complete and censored samples, the relationship between the Pareto and the exponential distributions could be of vital importance to test for the validity of this assumption. For grouped uncensored data the classical Pearson χ2 test based on the multinomial model can be used. Attention is confined in this article to handle grouped data with withdrawals within intervals. Graphical as well as analytical procedures will be presented. Maximum likelihood estimators for the parameters of the Pareto distribution based on grouped data will be derived. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
214. A Poisson-multinomial mixture approach to grouped and right-censored counts
- Author
-
Kenneth C. Land, Qiang Fu, and Xin Guo
- Subjects
Statistics and Probability ,05 social sciences ,050401 social sciences methods ,Estimator ,Poisson distribution ,01 natural sciences ,Regression ,Grouped data ,010104 statistics & probability ,symbols.namesake ,Quasi-likelihood ,0504 sociology ,Statistics ,symbols ,Econometrics ,Multinomial distribution ,Poisson regression ,0101 mathematics ,Count data ,Mathematics - Abstract
Although count data are often collected in social, psychological, and epidemiological surveys in grouped and right-censored categories, there is a lack of statistical methods simultaneously taking both grouping and right-censoring into account. In this research, we propose a new generalized Poisson-multinomial mixture approach to model grouped and right-censored (GRC) count data. Based on a mixed Poisson-multinomial process for conceptualizing grouped and right-censored count data, we prove that the new maximum-likelihood estimator (MLE-GRC) is consistent and asymptotically normally distributed for both Poisson and zero-inflated Poisson models. The use of the MLE-GRC, implemented in an R function, is illustrated by both statistical simulation and empirical examples. This research provides a tool for epidemiologists to estimate incidence from grouped and right-censored count data and lays a foundation for regression analyses of such data structure.
- Published
- 2017
215. Stress Affection of Two Scale Truncated Generalized Logistic Parameters with Progressive Censoring
- Author
-
A. M. Abd-Elfattah and Salma Omar Bleed
- Subjects
Statistics and Probability ,Constant-stress Accelerated Life Test ,Scale (ratio) ,Truncated Generalized Logistic Distribution ,EM algorithm method ,Fisher Information Matrix ,Progressive Type-I Censored Grouped Data ,media_common.quotation_subject ,Generalized logistic distribution ,Management Science and Operations Research ,01 natural sciences ,010104 statistics & probability ,symbols.namesake ,0502 economics and business ,Statistics ,Expectation–maximization algorithm ,0101 mathematics ,Fisher information ,lcsh:Statistics ,lcsh:HA1-4737 ,Normality ,050205 econometrics ,Mathematics ,media_common ,lcsh:Mathematics ,05 social sciences ,Estimator ,lcsh:QA1-939 ,Censoring (statistics) ,Grouped data ,Modeling and Simulation ,symbols ,Statistics, Probability and Uncertainty ,62F10 - Abstract
This paper deals with non-Bayesian estimation problem of constant-stress Accelerated Life Tests (ALTs) when the lifetime of the items follow truncated Generalized Logistic Distribution (GLD). Some considerations on inference based on the use of asymptotically normality of the ML estimators are presented considering the stress effects on the two scale parameters of the truncated GLD with a k-level constant-stress ALT under progressive type-I censored grouped data. The EM algorithm method is used to obtain the estimators of the unknown parameters. In addition, estimator of the two scale parameters, reliability function under usual conditions and Fisher information matrix of the estimators are given. Finally, we present a Simulation Study to illustrate the proposed procedure.
- Published
- 2017
216. Multilevel Modeling in Family Violence Research
- Author
-
Roderick A. Rose
- Subjects
Research design ,Sociology and Political Science ,Management science ,05 social sciences ,Multilevel model ,Context (language use) ,Ecological systems theory ,Data science ,Grouped data ,Clinical Psychology ,050501 criminology ,Leverage (statistics) ,Domestic violence ,0501 psychology and cognitive sciences ,Set (psychology) ,Psychology ,Law ,Social Sciences (miscellaneous) ,050104 developmental & child psychology ,0505 law - Abstract
Family violence researchers often use an ecological perspective to describe persons nested within groups. Further, family violence researchers frequently investigate whether group characteristics impact individual outcomes. The theoretical orientation and research designs typically used therefore present opportunities to utilize multilevel modeling (MLM) for clustered designs. It is widely understood that MLM corrects standard errors for grouped data, though other approaches can address this issue. Importantly, MLM presents a structured approach to the examination of group differences in outcomes, group differences in the association between the characteristics of persons and these outcomes, and the explanation of group differences using group-level characteristics. This journal frequently receives studies that use MLM for clustered designs, and a set of analytical guidelines may assist authors in preparing such articles so as to properly implement and better leverage the power of MLM to advance family violence research. I describe MLM for the new user, providing guidance on estimation of these models in the context of two examples. In addition, for more experienced users of MLM, I argue for greater attention to between-group and compositional effects that may be prevalent in family violence research, and the opportunities they may raise for a better understanding of the complexities at the group level. In closing I discuss some extensions of MLM and place MLM in the context of research design, providing guidelines for designing, carrying out, and reporting findings from studies that use these methods.
- Published
- 2017
217. Conditional optimal spacing in exponential distribution.
- Author
-
Park, Sangun
- Abstract
In this paper, we propose the conditional optimal spacing defined as the optimal spacing after specifying a predetermined order statistic. If we specify a censoring time, then the optimal inspection times for grouped inspection can be determined from this conditional optimal spacing. We take an example of exponential distribution, and provide a simple method of finding the conditional optimal spacing. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
218. 3-Year-Olds’ Perseveration on the DCCS Explained: A Meta-Analysis
- Author
-
Ari Franklin, Oriane Landry, and Shems Al-Taie
- Subjects
Perseveration ,05 social sciences ,Separation (statistics) ,Experimental and Cognitive Psychology ,050105 experimental psychology ,Developmental psychology ,Task (project management) ,Grouped data ,Psychiatry and Mental health ,Card sorting ,Distraction ,Meta-analysis ,Statistics ,Developmental and Educational Psychology ,Negative priming ,medicine ,0501 psychology and cognitive sciences ,medicine.symptom ,Psychology ,050104 developmental & child psychology - Abstract
The Dimensional Change Card Sort (DCCS) task is a widely used measure of preschoolers’ executive function. We combined data for 3,290 3-year-olds from 37 unique studies reporting 130 experimental conditions. Using raw pass/fail counts, we computed the pass rates and chi-squared value for each against chance (50/50) performance. We grouped data according to DCCS variants and computed the standard pass rate and chi-squared and phi for each variant relative to standard. For all standard versions, the mean pass rate was 36%. We compared all other variants to the standard and found robust improvements in performance for manipulations that involved spatial separation of the conflicting dimensions, use of distraction between pre and post-switch, elimination of all conflict, and extra practice. We also found that negative priming offers a better explanation for 3-year-olds’ perseveration than attentional inertia. The results support a theoretical model of 3-year-olds’ performance based on inhibitory control.
- Published
- 2017
219. Weighted Kolmogrov–Smirnov type tests for grouped Rayleigh data
- Author
-
Baklizi, Ayman
- Subjects
- *
STATISTICAL bootstrapping , *ESTIMATION theory , *DISTRIBUTION (Probability theory) , *STATISTICAL sampling - Abstract
Abstract: We consider goodness of fit tests for the Rayleigh distribution with grouped data. New Kolmogrov–Smirnov type tests are suggested and compared with the traditional chi-square and likelihood ratio tests. The results show that some of the suggested tests have a good power performance as compared with the traditional ones. [Copyright &y& Elsevier]
- Published
- 2006
- Full Text
- View/download PDF
220. Robust and efficient estimation under data grouping.
- Author
-
Nan Lin and Xuming He
- Subjects
- *
ESTIMATION theory , *QUANTITATIVE research , *STATISTICAL correlation , *APPROXIMATION theory , *MATHEMATICAL analysis - Abstract
The minimum Hellinger distance estimator is known to have desirable properties in terms of robustness and efficiency. We propose an approximate minimum Hellinger distance estimator by adapting the approach to grouped data from a continuous distribution. It is easier to compute the approximate version for either the continuous data or the grouped data. Given certain conditions on the model distribution and reasonable grouping rules, the approximate minimum Hellinger distance estimator is shown to be consistent and asymptotically normal. Furthermore, it is robust and can be asymptotically as efficient as the maximum likelihood estimator. The merit of the estimator is demonstrated through simulation studies and real data examples. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
221. Efficient Laplacian and Adaptive Gaussian Quadrature Algorithms for Multilevel Generalized Linear Mixed Models.
- Author
-
Pinheiro, José C. and Chao, Edward C.
- Subjects
- *
LAPLACIAN operator , *GAUSSIAN quadrature formulas , *LINEAR statistical models , *MULTILEVEL models , *REGRESSION analysis , *MATHEMATICAL models , *COMPUTATIONAL complexity - Abstract
Mixed-effects models have become a popular approach for the analysis of grouped data that arise in many areas as diverse as clinical trials, epidemiology, and sociology. Examples of grouped data include longitudinal data, repeated measures, and multilevel data. In the case of linear mixed-effects (LME) models, the likelihood function can be expressed in closed form, with efficient computational algorithms having been proposed for maximum likelihood and restricted maximum likelihood estimation. For nonlinear mixed-effects (NLME) models and generalized linear mixed models (GLMM5), however, the likelihood function does not have a closed form. Different likelihood approximations, with varying degrees of accuracy and computational complexity, have been proposed for these models. This article describes algorithms for one such approximation, the adaptive Gaussian quadrature (AGQ), for GLMMs which scale up efficiently to multilevel models with arbitrary number of levels. The proposed algorithms greatly reduce the computational complexity and the memory usage for approximating the multilevel GLMM likelihood, when compared to a direct application of a single-level AGQ approximation algorithm to the multilevel case. The accuracy of the associated estimates is evaluated and compared to that of estimates obtained from other approximations via simulation studies. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
222. Planning step-stress life test with progressively type I group-censored exponential data.
- Author
-
Shuo-Jye Wu, Ying-Po Lin, and Yi-Ju Chen
- Subjects
- *
HUMAN life cycle , *ACCELERATED life testing , *OPTIMAL designs (Statistics) , *PROBABILITY theory , *ANALYSIS of variance , *STATISTICS - Abstract
Accelerated life testing of products is used to get information quickly on their lifetime distributions. This paper discusses a k-stage step-stress accelerated life test under progressive type I censoring with grouped data. An exponential lifetime distribution with mean life that is a log-linear function of stress is considered. A cumulative exposure model is also assumed. We use the maximum likelihood method to obtain the estimators of the model parameters. The methods for obtaining the optimum test plan are investigated using the variance-optimality and D-optimality criteria. Some numerical studies are discussed to illustrate the proposed criteria. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
223. The simultaneous assessment of normality and homoscedasticity in linear fixed effects models
- Author
-
Thomas Mathew and Ye Yang
- Subjects
Statistics and Probability ,Score test ,media_common.quotation_subject ,05 social sciences ,Score ,01 natural sciences ,Grouped data ,Normal distribution ,010104 statistics & probability ,Homoscedasticity ,0502 economics and business ,Statistics ,Null distribution ,0101 mathematics ,Legendre polynomials ,Normality ,050205 econometrics ,media_common ,Mathematics - Abstract
This article investigates the problem of simultaneously testing the normality and homoscedasticity assumptions in a linear fixed effects model when we have grouped data. This has been facilitated by the assumption of a smooth alternative to the normal distribution. The smooth alternative is specified using Legendre polynomials, and the score statistic is derived under two scenarios: a common smooth alternative across the different groups, or different smooth alternatives across the different groups. A data-driven approach available in the literature is used for determining the order of the polynomials. For the null distribution of the score statistic, the accuracy of the asymptotic chi-squared distribution is numerically investigated under a one-way fixed effects model with balanced and unbalanced data. The results are illustrated with an example.
- Published
- 2017
224. A Unified Approach to Estimating and Testing Income Distributions With Grouped Data
- Author
-
Yi-Ting Chen
- Subjects
Statistics and Probability ,Consumption (economics) ,Economics and Econometrics ,Income shares ,05 social sciences ,Monte Carlo method ,Sample (statistics) ,Grouped data ,Income distribution ,0502 economics and business ,Statistics ,Econometrics ,Economics ,050207 economics ,Statistics, Probability and Uncertainty ,Social Sciences (miscellaneous) ,050205 econometrics ,Parametric statistics ,Quantile - Abstract
We propose a unified approach that is flexibly applicable to various types of grouped data for estimating and testing parametric income distributions. To simplify the use of our approach, we also provide a parametric bootstrap method and show its asymptotic validity. We also compare this approach with existing methods for grouped income data, and assess their finite-sample performance by a Monte Carlo simulation. For empirical demonstrations, we apply our approach to recovering China's income/consumption distributions from a sequence of income/consumption share tables and the U.S. income distributions from a combination of income shares and sample quantiles. Supplementary materials for this article are available online.
- Published
- 2017
225. Simulation–Extrapolation for Bias Correction with Exposure Uncertainty in Radiation Risk Analysis Utilizing Grouped Data
- Author
-
Harry M. Cullings, Kyoji Furukawa, Munechika Misumi, and John B. Cologne
- Subjects
Statistics and Probability ,Regression calibration ,Observational error ,Dose distribution ,01 natural sciences ,Grouped data ,010104 statistics & probability ,03 medical and health sciences ,Radiation risk ,0302 clinical medicine ,Robustness (computer science) ,030220 oncology & carcinogenesis ,Statistics ,Econometrics ,Bias correction ,0101 mathematics ,Statistics, Probability and Uncertainty ,Mathematics ,Simulation extrapolation - Abstract
Summary In observational epidemiological studies, the exposure that is received by an individual often cannot be precisely observed, resulting in measurement error, and a common approach to dealing with measurement error is regression calibration (RC). Use of RC, which requires assumptions about the distribution of unknown error-free (true) variables, leads to concern about the possibility of bias due to misspecification of that distribution. The simulation–extrapolation (SIMEX) method, in contrast, does not require a distributional assumption. However, analyses of large cohorts may be performed by using grouped or person-year data, and application of SIMEX to grouped data is not straightforward, particularly when there is a mixture of classical and Berkson measurement errors. We compared RC and SIMEX with grouped data analyses to assess robustness of the RC method to misspecification of the true dose distribution. We also applied SIMEX assuming mixtures of classical and Berkson errors and compared the results with those obtained by using RC for classical error only. SIMEX had less bias than RC and performed well regardless of the true dose distribution, whereas RC based on a misspecified true dose distribution showed greater bias than when based on the correctly specified true dose distribution.
- Published
- 2017
226. Maximum likelihood estimators under progressive Type-I interval censoring
- Author
-
Debasis Sengupta, Sonal Budhiraja, and Biswabrata Pradhan
- Subjects
Statistics and Probability ,021103 operations research ,Maximum likelihood ,0211 other engineering and technologies ,Estimator ,Asymptotic distribution ,02 engineering and technology ,01 natural sciences ,Grouped data ,010104 statistics & probability ,Censoring (clinical trials) ,Statistics ,0101 mathematics ,Statistics, Probability and Uncertainty ,Mathematics - Abstract
The consistency and asymptotic normality of the maximum likelihood estimators (MLEs), based on progressively type-I interval censored (PIC-I) data are proved under appropriate regularity conditions. The information obtained in the PIC-I setup is compared with that of grouped data and also with progressively type-I censored (PC-I) data.
- Published
- 2017
227. A new method for calculating quantiles of grouped data based on the frequency polygon
- Author
-
Hyuk Joo Kim
- Subjects
business.industry ,Polygon ,Statistics ,Pattern recognition ,Artificial intelligence ,business ,Grouped data ,Mathematics ,Quantile - Published
- 2017
228. Utilizing patient geographic information system data to plan telemedicine service locations
- Author
-
Joseph DeWalle, Neelkamal Soares, and Ben Marsh
- Subjects
Service (systems architecture) ,Telemedicine ,Geospatial analysis ,Geographic information system ,020205 medical informatics ,Population ,Health Informatics ,02 engineering and technology ,Research and Applications ,computer.software_genre ,Health Services Accessibility ,03 medical and health sciences ,0302 clinical medicine ,Patient Load ,Health care ,0202 electrical engineering, electronic engineering, information engineering ,Electronic Health Records ,Humans ,030212 general & internal medicine ,education ,Retrospective Studies ,education.field_of_study ,Database ,Delivery of Health Care, Integrated ,business.industry ,Data Collection ,Pennsylvania ,Grouped data ,Health Planning ,Geography ,Geographic Information Systems ,Rural Health Services ,business ,computer ,Cartography - Abstract
Objective: To understand potential utilization of clinical services at a rural integrated health care system by generating optimal groups of telemedicine locations from electronic health record (EHR) data using geographic information systems (GISs). Methods: This retrospective study extracted nonidentifiable grouped data of patients over a 2-year period from the EHR, including geomasked locations. Spatially optimal groupings were created using available telemedicine sites by calculating patients’ average travel distance (ATD) to the closest clinic site. Results: A total of 4027 visits by 2049 unique patients were analyzed. The best travel distances for site groupings of 3, 4, 5, or 6 site locations were ranked based on increasing ATD. Each one-site increase in the number of available telemedicine sites decreased minimum ATD by about 8%. For a given group size, the best groupings were very similar in minimum travel distance. There were significant differences in predicted patient load imbalance between otherwise similar groupings. A majority of the best site groupings used the same small number of sites, and urban sites were heavily used. Discussion: With EHR geospatial data at an individual patient level, we can model potential telemedicine sites for specialty access in a rural geographic area. Relatively few sites could serve most of the population. Direct access to patient GIS data from an EHR provides direct knowledge of the client base compared to methods that allocate aggregated data. Conclusion: Geospatial data and methods can assist health care location planning, generating data about load, load balance, and spatial accessibility.
- Published
- 2017
229. Random-intercept misspecification in generalized linear mixed models for binary responses
- Author
-
Xianzheng Huang and Shun Yu
- Subjects
Statistics and Probability ,Binary response ,05 social sciences ,Binary number ,Estimator ,01 natural sciences ,Generalized linear mixed model ,Power (physics) ,Grouped data ,010104 statistics & probability ,Distribution (mathematics) ,0502 economics and business ,Econometrics ,Applied mathematics ,0101 mathematics ,Statistics, Probability and Uncertainty ,Random intercept ,050205 econometrics ,Mathematics - Abstract
We study properties of maximum likelihood estimators of parameters in generalized linear mixed models for a binary response in the presence of random-intercept model misspecification. Further exploiting the test proposed in an existing work initially designed for detecting general random-effects misspecification, we are able to reveal how the true random-intercept distribution deviates from the assumed. Besides this advance compared to the existing methods, we also provide theoretical insights on when and why the proposed test has low power to identify certain forms of misspecification. Large-sample numerical study and finite-sample simulation experiments are carried out to illustrate the theoretical findings.
- Published
- 2017
230. Bandwidth selection in kernel density estimation for interval-grouped data
- Author
-
Miguel Reyes, Mario Francisco-Fernández, and Ricardo Cao
- Subjects
Statistics and Probability ,Statistics::Theory ,Mathematical optimization ,Bandwidth (signal processing) ,Kernel density estimation ,Monte Carlo method ,Estimator ,04 agricultural and veterinary sciences ,01 natural sciences ,Grouped data ,010104 statistics & probability ,Rate of convergence ,Sample size determination ,040103 agronomy & agriculture ,0401 agriculture, forestry, and fisheries ,0101 mathematics ,Statistics, Probability and Uncertainty ,Smoothing ,Mathematics - Abstract
When interval-grouped data are available, the classical Parzen–Rosenblatt kernel density estimator has to be modified to get a computable and useful approach in this context. The new nonparametric grouped data estimator needs of the choice of a smoothing parameter. In this paper, two different bandwidth selectors for this estimator are analyzed. A plug-in bandwidth selector is proposed and its relative rate of convergence obtained. Additionally, a bootstrap algorithm to select the bandwidth in this framework is designed. This method is easy to implement and does not require Monte Carlo. Both proposals are compared through simulations in different scenarios. It is observed that when the sample size is medium or large and grouping is not heavy, both bandwidth selection methods have a similar and good performance. However, when the sample size is large and under heavy grouping scenarios, the bootstrap bandwidth selector leads to better results.
- Published
- 2017
231. A Weighted Exponential Model for Grouped Line Transect Data
- Author
-
Omar Eidous and Fahid Al Eibood
- Subjects
Statistics and Probability ,Economics and Econometrics ,010308 nuclear & particles physics ,Maximum likelihood ,Estimator ,01 natural sciences ,Population abundance ,Exponential function ,Grouped data ,0103 physical sciences ,Line (geometry) ,Statistics ,Parametric model ,Statistics, Probability and Uncertainty ,010306 general physics ,Transect ,Mathematics - Abstract
This paper considers a parametric model for grouped data collected via line transect technique. The weighted exponential model is studied and investigated when the data are assumed to be grouped in the intervals. The maximum likelihood method is adopted for purpose of estimation. The resultant estimator of the population abundance is compared with the corresponding estimator that developed for ungrouped data by using the Laake stakes real data.
- Published
- 2017
232. King-Devick and Pre-season Visual Function in Adolescent Athletes
- Author
-
Drew Ferguson, Mark W. Swanson, Matthew Heath Hale, Katherine K. Weise, and Kimberly Penix
- Subjects
Male ,medicine.medical_specialty ,Visual acuity ,Adolescent ,Cross-sectional study ,Visual Acuity ,Neuropsychological Tests ,Audiology ,Article ,Pupil ,03 medical and health sciences ,Vision Screening ,0302 clinical medicine ,medicine ,Humans ,Bland–Altman plot ,Brain Concussion ,biology ,business.industry ,Athletes ,Repeated measures design ,Repeatability ,Convergence, Ocular ,biology.organism_classification ,Grouped data ,Ophthalmology ,Cross-Sectional Studies ,Athletic Injuries ,030221 ophthalmology & optometry ,Female ,Seasons ,medicine.symptom ,business ,030217 neurology & neurosurgery ,Optometry - Abstract
PURPOSE The King-Devick test (KD) has been studied as a remove-from-play sideline test in college-age athletes and older; however, studies in younger athletes are limited. A cross-sectional study of the KD and other vision correlates was completed on school-aged athletes during pre-season physicals for a variety of sports to determine the repeatability of the KD. The study also evaluated how convergence, alignment, or pupil function contributed to a slower King-Devick baseline reading. METHODS Seven hundred eighty-five athletes underwent vision screenings in a hospital or school setting by trained/certified staff as part of pre-season physicals. Six hundred nineteen had KD testing completed per the manufacturer's suggested protocol and repeated. Other baseline vision testing included visual acuity, Modified Thorington testing for alignment, convergence testing, and pupil function using the NeurOptics (NPI-200) NPi. RESULTS The mean fastest, error-minimized KD time for all participants was 43.9 seconds(s) (SD ± 11.6, range 24-120). Median KD time got faster (+) with age (p < 0.0001). The inter-class correlation coefficient for all scores was 0.92. The absolute mean time difference for any two tests was 3.5 s (SD ± 2.5, range 0-23). There was no association between the best KD time and reduced NPC (p = 0.63), Modified Thorington measure of alignment (p = 0.55), or NPi pupil function (p = 0.79). The Bland Altman repeated measure limits of agreement was ±6.5 seconds for those in the 10th to12th grades, and ±10.2 seconds for those in the 6th to 9th grades. CONCLUSIONS King-Devick score in junior high and high school athletes is variable but gets faster and more repeatable with increasing age. The KD does not correlate significantly with reduced convergence, alignment, or pupil function. Based on grouped data, a slowing of 10 seconds for younger athletes and 6 seconds for older athletes on a second administration represents a true difference in testing speed. Within-player variability should be considered when removal-from-play decisions are influenced by KD results.
- Published
- 2017
233. AVERAGING INCOME DISTRIBUTIONS.
- Author
-
Griffiths, William E., Chotikapanich, Duangkamon, and Prasada Rao, D. S.
- Subjects
INCOME inequality ,PUBLIC welfare ,BAYESIAN analysis ,INCOME ,DISTRIBUTION (Probability theory) - Abstract
Various inequality and social welfare measures often depend heavily on the choice of a distribution of income. Picking a distribution that best fits the data involves throwing away information and does not allow for the fact that a wrong choice can be made. Instead, Bayesian model averaging utilizes a weighted average of the results from a number of income distributions, with each weight given by the probability that a distribution is ‘correct’. In this study, prior densities are placed on mean income, the mode of income and the Gini coefficient for Australian income units with one parent (1997–8). Then, using grouped sample data on incomes, posterior densities for the mean and mode of income and the Gini coefficient are derived for a variety of income distributions. The model-averaged results from these income distributions are obtained. [ABSTRACT FROM AUTHOR]
- Published
- 2005
- Full Text
- View/download PDF
234. Analyzing grouped data with hierarchical linear modeling
- Author
-
Guo, Shenyang
- Subjects
- *
CHILD welfare , *CAREGIVERS , *ACADEMIC achievement , *MULTILEVEL models - Abstract
Abstract: Grouped data are common but often improperly treated in welfare and child welfare research. Conventional regression models are not appropriate for analysis of this type of data, because the presence of intra-class correlation among study subjects from the same group violates the assumption that observations are independent of one another. This study demonstrates the advantages of using hierarchical linear modeling (HLM) to analyze grouped data found in the 1997 Child Development Supplement to the Panel Study of Income Dynamics. Specifically, this article presents an HLM example investigating intergenerational dependence on welfare and its relation to child academic achievement. Results show that HLM is a robust and flexible tool that can effectively test various types of research hypotheses, particularly those concerning multilevel influences and macro-to-micro relations. The study shows that early educational intervention is essential in improving child academic achievement for children receiving welfare, particularly for those who used welfare for most of their own childhood and whose caregivers never used welfare. [Copyright &y& Elsevier]
- Published
- 2005
- Full Text
- View/download PDF
235. The Trunsored Model and Its Applications to Lifetime Analysis: Unified Censored and Truncated Models.
- Author
-
Hirose, Hideo
- Subjects
- *
INFORMATION science , *MATHEMATICAL models , *MATHEMATICS , *MATHEMATICAL functions , *MATHEMATICAL statistics , *MATHEMATICAL analysis - Abstract
A new incomplete data model, the trunsored model, in lifetime analysis is introduced. This model can be regarded as a unified model of the censored and truncated models. Using the model, we can not only estimate the ratio of the fragile population to the mixed fragile and durable populations, but also test a hypothesis that the ratio is equal to a prescribed value. A central point of the paper is that such a test can easily be realized through the newly introduced trunsored model, because it has been difficult to do such a hypothesis test under only the framework of censored and truncated models. Therefore, the relationship of the trunsored model to the censored and truncated models is clarified because the trunsored model unifies the censored and truncated models. The paper also shows how to obtain the estimates of the parameters in lifetime estimation, and corresponding confidence intervals for the fragile population. Typical examples applied to electronic board failures. and to breast cancer data, for lifetime estimation are demonstrated, and successfully worked using the trunsored model. [ABSTRACT FROM AUTHOR]
- Published
- 2005
- Full Text
- View/download PDF
236. CONTRIBUTIONS TO ESTIMATION AND MODELING USING QUANTILES
- Author
-
Dilanka S. Dedduwakumara
- Subjects
Estimation ,General Mathematics ,Statistics ,Grouped data ,Mathematics ,Quantile - Published
- 2020
237. A Random-Effects Log-Linear Model with Poisson Distributions
- Author
-
Maria Alexandra Seco and António St. Aubyn
- Subjects
log-linear models ,grouped data ,random effects ,mixed models ,overdispersion ,iterative reweighted generalized least squares ,Statistics ,HA1-4737 ,Probabilities. Mathematical statistics ,QA273-280 - Abstract
In several applications data are grouped and there are within-group correlations. With continuous data, there are several available models that are often used; with counting data, the Poisson distribution is the natural choice. In this paper a mixed log-linear model based on a Poisson–Poisson conditional distribution is presented. The initial model is a conditional model for the mean of the response variable, and the marginal model is formed thereafter. Random effects with Poisson distribution are introduced and a variance-covariance matrix for the response vector is formed embodying the covariance structure induced by the grouping of the data.
- Published
- 2003
- Full Text
- View/download PDF
238. Linear mixed-effects models for weight–length relationships
- Author
-
Lai, Han-Lin and Helser, Thomas
- Subjects
- *
MASS (Physics) , *NATURAL resources , *AQUATIC resources , *BIOMASS - Abstract
Abstract: Length and weight data are often analyzed in fisheries science to derive a parametric weight–length relationship for estimating biomass and to develop indices of condition for comparing the ‘wellness’ of different populations of fish. However, analysis of such data often ignores the inherent spatial and temporal grouping of the observations, and hence, the data hierarchy. This paper proposes the use of linear mixed-effects models as an effective means of analyzing and comparing weight–length relationships and indices of condition when there are many groups. The use of simple linear regression (where grouping is ignored), ANCOVA (where group effects are incorporated as fixed-effects), and linear mixed-effects models (where group effects are random-effects) are compared using data for Atlantic sea scallops (Placopecten magellanicus). The group means of residuals is proposed as a measure of relative weight or index of population condition. Linear mixed-effects models should be used to analyze grouped data because the variability among groups is ignored in simple linear regression and ANCOVA. Also, it is important that explanatory variables be incorporated in analyses of grouped data because their influence may mask the true differences among groups. [Copyright &y& Elsevier]
- Published
- 2004
- Full Text
- View/download PDF
239. Inference based on the EM algorithm for the competing risks model with masked causes of failure.
- Author
-
Craiu, Radu V. and Duchesne, Thierry
- Subjects
- *
EXPECTATION-maximization algorithms , *RANDOM variables , *DISTRIBUTION (Probability theory) , *FAILURE time data analysis , *ALGORITHM research , *MATHEMATICAL statistics - Abstract
In this paper we propose inference methods based on the EM algorithm for estimating the parameters of a weakly parameterised competing risks model with masked causes of failure and second‐stage data. With a carefully chosen definition of complete data, the maximum likelihood estimation of the cause‐specific hazard functions and of the masking probabilities is performed via an EM algorithm. Both the E‐ and M‐steps can be solved in closed form under the full model and under some restricted models of interest. We illustrate the flexibility of the method by showing how grouped data and tests of common hypotheses in the literature on missing cause of death can be handled. The method is applied to a real dataset and the asymptotic and robustness properties of the estimators are investigated through simulation. [ABSTRACT FROM AUTHOR]
- Published
- 2004
- Full Text
- View/download PDF
240. Bayesian-Pearson Divergence Estimator Based on Grouped Data.
- Author
-
Baoxue Zhang and Qingxun Meng
- Subjects
BAYESIAN analysis ,DISTRIBUTION (Probability theory) ,PROBABILITY theory ,SYSTEMS theory ,MATHEMATICS ,POPULATION - Abstract
A new method along with Bayesian approach for estimating the parameter in the distribution function F(x;θ) by using grouped data is developed in this paper. The support of F(
x ;θ) is divided into disjointed intervals as -∞ = x0 < x1 < · · · < xk-1 < xk = +∞. Grouped data are the numbers of observations falling in the intervals. This method can be applied to estimate not only the parameter in one population model but also the parameters in multi-population model which is subject to the order restrictions. For it is not easy to present the prior distribution of the parameter θ in F(x; θ) by grouped data directly, it is considered the prior distribution of the probabilities of observations falling in the intervals denoted by Pj (θ) = F(xj ;θ) - F(xj-1 ;θ),j = 1, · · ·, k. The probabilities follow the multivariate distribution and can be regarded as the function of the parameter θ. Pearson divergence D(p; q) is introduced to scale the distance between the probabilities Pj (θ),j = 1 · · ·, k and the samples from the posterior distribution (Dirichlet distribution) of the probabilities. Then by minimizing the Pearson divergence D(p;q), the 'posterior' samples of the parameter θ can be obtained, through which statistical inference including Bayesian-Pearson Divergence Estimator of the parameter can be processed. Simulations and a numerical example employing this method are presented. [ABSTRACT FROM AUTHOR]- Published
- 2004
241. A Simple Estimator for the Shape Parameter of the Pareto Distribution with Economics and Medical Applications.
- Author
-
Ismaïl, Sanaa
- Subjects
- *
ESTIMATION theory , *MATHEMATICAL models , *WELFARE economics , *MATHEMATICAL physics , *ELASTICITY (Economics) , *ECONOMIC indicators - Abstract
In the present paper, an estimator of the shape parameter of the Pareto failure model is presented using grouped data. This estimator is based on obtaining the parameter in terms of the hazard rate, then replacing the unknown hazard rate by a grouped data estimator available in the literature. Death records are given as a numerical illustration in the medical context. The relation between the hazard rate and the income elasticity is derived. This relation allows the presentation of the same estimator in terms of the income elasticity so that it could be used in an economic context. Two illustrations are presented using income data. Simulated data are generated to compare the estimator with the maximum likelihood estimator. [ABSTRACT FROM AUTHOR]
- Published
- 2004
- Full Text
- View/download PDF
242. Estimating the Parameters of General Frequency Modulated Signals.
- Author
-
Luginbuhl, Tod and Willett, Peter
- Subjects
- *
ELECTRONIC modulation , *FREQUENCIES of oscillating systems , *ELECTRIC power , *ALGORITHMS , *DYNAMICS , *ESTIMATION theory - Abstract
A general frequency modulated (GFM) signal characterizes the vibrations produced by compressors, turbines, propellers, gears, and other rotating machines in a dynamic environment. A GFM signal is defined as the composition of a real or complex, periodic, or almost-periodic carrier function with a real, differentiable modulation function. A GFM signal therefore contains sinusoids whose frequencies are (possibly nonintegral) multiples of a fundamental; to distinguish a GFM signal from a set of unrelated sinusoids, it is necessary to track them as a group. This paper develops the general frequency modulation tracker (GFMT) for one or more GFM signals in noise using the expectation/conditional maximization (ECM) algorithm that is an extension of the expectation-maximization (EM) algorithm. Three advantages of this approach are that the ratios (harmonic numbers) of the carrier functions do not need to be known a priori, that the parameters of multiple signals are estimated simultaneously, and that the GFMT algorithm exploits knowledge of the noise spectrum so that a separate normalization procedure is not required. Several simulated examples are presented to illustrate the algorithm's performance. [ABSTRACT FROM AUTHOR]
- Published
- 2004
- Full Text
- View/download PDF
243. Recursive estimation in linear models with general errors and grouped data: a median-based procedure and related asymptotics
- Author
-
Anido, Carmen, Rivero, Carlos, and Valdés, Teófilo
- Subjects
- *
LINEAR statistical models , *ASYMPTOTIC expansions - Abstract
We introduce in this paper an iterative estimation procedure based on conditional medians valid to fit linear models when, on the one hand, the distribution of errors, assumed to be known, may be general and, on the other, the dependent data stem from different sources and, consequently, may be either non-grouped or grouped with different classification criteria. The procedure requires us at each step to interpolate the grouped data and is similar to the EM algorithm with normal errors. The expectation step has been replaced by a median-based step which avoids doing awkward integration with general errors and, also, we have substituted for the maximisation step, a natural one which only coincides with it when the errors are normally distributed. With these modifications, we have proved that the iterative estimating algorithm converges to a point which is unique and non-dependent on the starting values. Finally, our final estimate, being a Huber type
M -estimator, may enjoy good stochastic asymptotic properties which have also been investigated in detail. [Copyright &y& Elsevier]- Published
- 2003
- Full Text
- View/download PDF
244. On Reconstitution of Smooth Distributions from Grouped Data
- Author
-
K.K. Avilov
- Subjects
03 medical and health sciences ,0302 clinical medicine ,business.industry ,030220 oncology & carcinogenesis ,Applied Mathematics ,Biomedical Engineering ,Pattern recognition ,030212 general & internal medicine ,Artificial intelligence ,business ,Mathematics ,Grouped data - Abstract
In the paper, proposed is a simple nonparametric method of reconstitution of smooth distributions of additive quantities from grouped data. The method is based on the requirement of minimization of the norm of non-smoothness measure of the solution under the condition of exact equality of the group sums, which reduces the problem to the quadratic programming problem. The method was tested on the age-at-death data; its precision was shown to be comparable to and exceeding the precision of a method of other authors. After testing it on the cancer incidence data, some drawbacks and limitations of the nonparametric approach were determined. The advantages of the proposed method are algorithmic and computational simplicity, good flexibility of the mathematical model.
- Published
- 2016
245. Modal iterative estimation in linear models with unimodal errors and non-grouped and grouped data collected from different sources.
- Author
-
Anido, Carmen, Valdés, Teófilo, and Rivero, Carlos
- Abstract
In this paper we introduce an iterative estimation procedure based on conditional modes suitable to fit linear models when errors are known to be unimodal and, moreover, the dependent data stem from different sources and, consequently, may be either non-grouped or grouped with different classification criteria. The procedure requires, at each step, the imputation of the exact values of the grouped data and runs by means of a process that is similar to the EM algorithm with normal errors. The expectation step has been substituted with a mode step that avoids awkward integration with general errors and, in addition, we have substituted the maximisation step with a natural one which only coincides with it when the error distribution is normal. Notwithstanding the former modifications, we have proved that, on the one hand, the iterative estimating algorithm converges to a point which is unique and non-dependent on the starting values and, on the other hand, our final estimate, being an M-estimator, may enjoy good stochastic asymptotic properties such as consistency, boundness in L
2 , and limit normality. [ABSTRACT FROM AUTHOR]- Published
- 2000
- Full Text
- View/download PDF
246. Analysis of grouped data using conjugate generalized linear mixed models
- Author
-
Jarod Y. L. Lee, Peter H.R. Green, and Louise Ryan
- Subjects
Statistics and Probability ,Longitudinal data ,Statistics & Probability ,General Mathematics ,Unit-level model ,Multilevel model ,01 natural sciences ,Generalized linear mixed model ,010104 statistics & probability ,0502 economics and business ,Applied mathematics ,0103 Numerical and Computational Mathematics, 0104 Statistics, 1403 Econometrics ,0101 mathematics ,050205 econometrics ,Mathematics ,Applied Mathematics ,05 social sciences ,Random effects model ,Agricultural and Biological Sciences (miscellaneous) ,Grouped data ,Random effect ,Closed-form marginal likelihood ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Conjugate - Abstract
Summary This article concerns a class of generalized linear mixed models for two-level grouped data, where the random effects are uniquely indexed by groups and are independent. We derive necessary and sufficient conditions for the marginal likelihood to be expressed in explicit form. These models are unified under the conjugate generalized linear mixed models framework, where conjugate refers to the fact that the marginal likelihood can be expressed in closed form, rather than implying inference via the Bayesian paradigm. The proposed framework allows simultaneous conjugacy for Gaussian, Poisson and gamma responses, and thus can accommodate both unit- and group-level covariates. Only group-level covariates can be incorporated for the binomial distribution. In a simulation of Poisson data, our framework outperformed its competitors in terms of computational time, and was competitive in terms of robustness against misspecification of the random effects distributions.
- Published
- 2019
247. Presentation of grouped data
- Author
-
John Bird
- Subjects
Presentation ,business.industry ,media_common.quotation_subject ,Artificial intelligence ,computer.software_genre ,Psychology ,business ,computer ,Natural language processing ,Grouped data ,media_common - Published
- 2019
248. Analysis of interval‐grouped data in weed science: The binnednp Rcpp package
- Author
-
Ramón Doallo, Basilio B. Fraguela, José Luis González-Andújar, Mario Francisco-Fernández, Daniel Barreiro-Ures, Ricardo Cao, Miguel Reyes, Ministerio de Economía y Competitividad (España), Ministerio de Ciencia, Innovación y Universidades (España), Agencia Estatal de Investigación (España), Xunta de Galicia, and European Commission
- Subjects
0106 biological sciences ,weed emergence model ,Nonparametric kernel estimation ,010603 evolutionary biology ,01 natural sciences ,03 medical and health sciences ,Agricultural science ,Regional development ,lcsh:QH540-549.5 ,hydrothermal time ,Weed emergence model ,Ecology, Evolution, Behavior and Systematics ,030304 developmental biology ,Nature and Landscape Conservation ,Original Research ,0303 health sciences ,Ecology ,Weed science ,Bandwidth selection ,Grouped data ,Hydrothermal time ,Geography ,Interval (graph theory) ,lcsh:Ecology ,nonparametric kernel estimation ,bandwidth selection - Abstract
Weed scientists are usually interested in the study of the distribution and density functions of the random variable that relates weed emergence with environmental indices like the hydrothermal time (HTT). However, in many situations, experimental data are presented in a grouped way and, therefore, the standard nonparametric kernel estimators cannot be computed. Kernel estimators for the density and distribution functions for interval‐grouped data, as well as bootstrap confidence bands for these functions, have been proposed and implemented in the binnednp package. Analysis with different treatments can also be performed using a bootstrap approach and a Cramér‐von Mises type distance. Several bandwidth selection procedures were also implemented. This package also allows to estimate different emergence indices that measure the shape of the data distribution. The values of these indices are useful for the selection of the soil depth at which HTT should be measured which, in turn, would maximize the predictive power of the proposed methods. This paper presents the functions of the package and provides an example using an emergence data set of Avena sterilis (wild oat). The binnednp package provides investigators with a unique set of tools allowing the weed science research community to analyze interval‐grouped data., This research has been partially supported by MINECO grants MTM2014‐52876‐R and MTM2017‐82724‐R, for the first three authors, and by the Xunta de Galicia (Grupos de Referencia Competitiva ED431C‐2016‐015 and Centro Singular de Investigación de Galicia ED431G/01), for the first fifth authors, all of them through the European Regional Development Funds (ERDF). The research of sixth author has been partially funded by FEDER (ERDF) and MINECO grants AGL2012‐33736 and AGL2015‐64130‐R.
- Published
- 2019
249. Reconstruction of Continuous Image Using Maximum Likelihood Estimates from Grouped Data for Measuring Light Intensity and Interpolation by Atomic Functions According to Aperture of Photosensitive Element of Sensor
- Author
-
Irina Efremova, Alexander Malyshev, and Vladislav Efremov
- Subjects
Digital image ,Light intensity ,business.product_category ,Optics ,business.industry ,Aperture ,business ,Signal ,Intensity (heat transfer) ,Grouped data ,Digital camera ,Mathematics ,Interpolation - Abstract
This paper considers the optical signal acquired by a digital camera as a continuous function varying in the spatial domain that uses the light intensity distribution to carry information. Two consistently applied techniques for representing a digital image as a continuous approximation to the original signal values are presented. The quantized light intensity is considered as grouped data. The maximum likelihood method on the grouped data can get the best estimate of the original analog value of the light intensity for each photosensitive element of the sensor. The intensity of light, measured by each photosensitive element, is considered as the double integral of the original optical signal over the region defined by the aperture of the element. Interpolation by atomic functions can use the integrated data to produce an approximation of the original continuous function, which is an optical signal.
- Published
- 2019
250. Bayesian approach to Lorenz curve using time series grouped data
- Author
-
Yuki Kawakubo, Genya Kobayashi, Yuta Yamauchi, Shonosuke Sugasawa, and Kazuhiko Kakamu
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Economics and Econometrics ,Gini coefficient ,State-space representation ,Series (mathematics) ,Bayesian probability ,Markov chain Monte Carlo ,Grouped data ,Methodology (stat.ME) ,symbols.namesake ,Income distribution ,Statistics ,symbols ,Statistics, Probability and Uncertainty ,Lorenz curve ,Social Sciences (miscellaneous) ,Statistics - Methodology ,Mathematics - Abstract
This study is concerned with estimating the inequality measures associated with the underlying hypothetical income distribution from the times series grouped data on the Lorenz curve. We adopt the Dirichlet pseudo likelihood approach where the parameters of the Dirichlet likelihood are set to the differences between the Lorenz curve of the hypothetical income distribution for the consecutive income classes and propose a state space model which combines the transformed parameters of the Lorenz curve through a time series structure. Furthermore, the information on the sample size in each survey is introduced into the originally nuisance Dirichlet precision parameter to take into account the variability from the sampling. From the simulated data and real data on the Japanese monthly income survey, it is confirmed that the proposed model produces more efficient estimates on the inequality measures than the existing models without time series structures.
- Published
- 2019
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.