22 results
Search Results
2. An error analysis for deep binary classification with sigmoid loss.
- Author
-
Li, Changshi, Jiao, Yuling, and Yang, Jerry Zhijian
- Subjects
- *
ARTIFICIAL neural networks , *CONVEX functions , *BUSINESS losses - Abstract
Deep neural networks have demonstrated remarkable efficacy in diverse classification tasks. In this paper, we specifically focus on the predictive performance in deep binary classification problems with the sigmoid loss. Given that sigmoid loss is categorized as a non-convex and bounded loss function, it exhibits potential resilience against the disruptive impact of outlier noises. We first derive the convergence rate of the excess misclassification risk for deep ReLU neural networks with the sigmoid loss, a result that attains minimax optimality. To the best of our acknowledge, we are the first to derive the convergence rate for the sigmoid loss. Moreover, we extend our analysis to derive a faster convergence rate under margin assumptions. This achievement renders our findings comparable to those of commonly employed convex loss functions operating under analogous assumptions. Lastly, we undertake a comprehensive validation of the robustness inherent in the sigmoid loss across diverse datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Spatio-temporal variability and possible source identification of criteria pollutants from Ahmedabad-a megacity of Western India
- Author
-
Bano, Shahana, Anand, Vrinda, Kalbande, Ritesh, Beig, Gufran, and Rathore, Devendra Singh
- Published
- 2024
- Full Text
- View/download PDF
4. An observational study on risk of secondary cancers in chronic myeloid leukemia patients in the TKI era in the United States.
- Author
-
Kumar, Vivek, Garg, Mohit, Chaudhary, Neha, and Chandra, Abhinav Binod
- Subjects
CHRONIC myeloid leukemia ,PROTEIN-tyrosine kinase inhibitors ,EPIDEMIOLOGY ,DISEASE incidence ,MEDICAL databases - Abstract
Introduction: The treatment with tyrosine kinase inhibitors (TKIs) has drastically improved the outcome of chronic myeloid leukemia (CML) patients. This study was conducted to examine the risk of secondary cancers (SCs) in the CML patients who were diagnosed and treated in the TKI era in the United States. Methods: The surveillance epidemiology and end results (SEER) database was used to identify CML patients who were diagnosed and received treatment during January 2002-December 2014. Standardized incidence ratios (SIRs) and absolute excess risks (AER) were calculated. Results: Overall, 511 SCs (excluding acute leukemia) developed in 9,200 CML patients followed for 38,433 person-years. The risk of developing SCs in the CML patients was 30% higher than the age, sex and race matched standard population (SIR 1.30, 95% CI: 1.2-1.40; p < 0.001). The SIRs for CLL (SIR 3.4, 95% CI: 2-5.5; p < 0.001), thyroid (SIR 2.2, 95% CI: 1.2-3.5; p < 0.001), small intestine (SIR 3.1, 95% CI: 1.1-7; p = 0.004), gingiva (SIR 3.7, 95% CI: 1.2-8.7; p = 0.002), stomach (SIR 2.1, 95% CI: 1.1-3.5; p = 0.005), lung (SIR 1.4, 95% CI: 1.1-1.7; p = 0.006) and prostate (SIR 1.3, 95% CI: 1.02-1.6; p = 0.026) cancer among CML patients were significantly higher than the general population. The risk of SCs was higher irrespective of age and it was highest in the period 2-12 months after the diagnosis of CML. The risk of SCs in women was similar to that of the general population. Conclusion: CML patients diagnosed and treated in the TKI era in the United States are at an increased risk of developing a second malignancy. The increased risk of SCs in the early period after CML diagnosis suggests that the risk of SCs may be increased due to the factors other than TKIs treatment. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
5. Minimum Excess Risk in Bayesian Learning.
- Author
-
Xu, Aolin and Raginsky, Maxim
- Subjects
PROBABILISTIC generative models ,EPISTEMIC uncertainty ,INFORMATION modeling ,PARAMETRIC modeling ,PREDICTION models - Abstract
We analyze the best achievable performance of Bayesian learning under generative models by defining and upper-bounding the minimum excess risk (MER): the gap between the minimum expected loss attainable by learning from data and the minimum expected loss that could be achieved if the model realization were known. The definition of MER provides a principled way to define different notions of uncertainties in Bayesian learning, including the aleatoric uncertainty and the minimum epistemic uncertainty. Two methods for deriving upper bounds for the MER are presented. The first method, generally suitable for Bayesian learning with a parametric generative model, upper-bounds the MER by the conditional mutual information between the model parameters and the quantity being predicted given the observed data. It allows us to quantify the rate at which the MER decays to zero as more data becomes available. Under realizable models, this method also relates the MER to the richness of the generative function class, notably the VC dimension in binary classification. The second method, particularly suitable for Bayesian learning with a parametric predictive model, relates the MER to the minimum estimation error of the model parameters from data via various continuity arguments. We also extend the definition and analysis of MER to the setting with multiple model families and the setting with nonparametric models. Along the discussions we draw some comparisons between the MER in Bayesian learning and the excess risk in frequentist learning. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
6. Learning rate of support vector machine for ranking.
- Author
-
Chen, Heng and Chen, Di-Rong
- Subjects
- *
SUPPORT vector machines , *RANKING (Statistics) , *MACHINE learning , *U-statistics , *PROBLEM solving , *LEARNING curve - Abstract
The ranking problem has received increasing attention in both the statistical and machine learning literature. This paper considers support vector machines for ranking. Under some mild conditions, a learning rate is established. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
7. On Ranking and Generalization Bounds.
- Author
-
Rejchel, Wojciech
- Subjects
- *
MATHEMATICAL bounds , *EMPIRICAL research , *DATA analysis , *GENERALIZATION , *RANKING (Statistics) , *ESTIMATION theory - Abstract
The problem of ranking is to predict or to guess the ordering between objects on the basis of their observed features. In this paper we consider ranking estimators that minimize the empirical convex risk. We prove generalization bounds for the excess risk of such estimators with rates that are faster than 1/√n. We apply our results to commonly used ranking algorithms, for instance boosting or support vector machines. Moreover, we study the performance of considered estimators on real data sets. [ABSTRACT FROM AUTHOR]
- Published
- 2012
8. Spatiotemporal analysis of dengue fever in Burkina Faso from 2016 to 2019
- Author
-
Ouattara, Cheick Ahmed, Traore, Seydou, Sangare, Ibrahim, Traore, Tiandiogo Isidore, Meda, Ziemlé Clément, and Savadogo, Léon G. Blaise
- Published
- 2022
- Full Text
- View/download PDF
9. Weighted least squares estimation for exchangeable binary data.
- Author
-
Bowman, Dale and George, E.
- Subjects
LEAST squares ,MAXIMUM likelihood statistics ,COMPUTATIONAL complexity ,PARAMETRIC equations ,NEWTON-Raphson method - Abstract
Parametric models of discrete data with exchangeable dependence structure present substantial computational challenges for maximum likelihood estimation. Coordinate descent algorithms such as the Newton's method are usually unstable, becoming a hit or miss adventure on initialization with a good starting value. We propose a method for computing maximum likelihood estimates of parametric models for finitely exchangeable binary data, formalized as an iterative weighted least squares algorithm. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
10. Adaptive spectral regularizations of high dimensional linear models
- Author
-
Yuri Golubev
- Subjects
Statistics and Probability ,62C10 ,empirical risk minimization ,Linear model ,Order (ring theory) ,Sigma ,Mathematics - Statistics Theory ,ordered smoother ,Statistics Theory (math.ST) ,High dimensional ,Regularization (mathematics) ,oracle inequality ,spectral regularization ,symbols.namesake ,Additive white Gaussian noise ,FOS: Mathematics ,symbols ,Applied mathematics ,62G05 ,Empirical risk minimization ,Statistics, Probability and Uncertainty ,excess risk ,Oracle inequality ,Mathematics - Abstract
This paper focuses on recovering an unknown vector $\beta$ from the noisy data $Y=X\beta +\sigma\xi$, where $X$ is a known $n\times p$-matrix, $\xi $ is a standard white Gaussian noise, and $\sigma$ is an unknown noise level. In order to estimate $\beta$, a spectral regularization method is used, and our goal is to choose its regularization parameter with the help of the data $Y$. In this paper, we deal solely with regularization methods based on the so-called ordered smoothers and provide some oracle inequalities in the case, where the noise level is unknown.
- Published
- 2011
11. Generalized Scalar-on-Image Regression Models via Total Variation.
- Author
-
Wang, Xiao and Zhu, Hongtu
- Subjects
SCALAR field theory ,LINEAR statistical models ,FUNCTIONAL analysis ,REGRESSION analysis ,ALZHEIMER'S disease ,BRAIN imaging - Abstract
The use of imaging markers to predict clinical outcomes can have a great impact in public health. The aim of this article is to develop a class of generalized scalar-on-image regression models via total variation (GSIRM-TV), in the sense of generalized linear models, for scalar response and imaging predictor with the presence of scalar covariates. A key novelty of GSIRM-TV is that it is assumed that the slope function (or image) of GSIRM-TV belongs to the space of bounded total variation to explicitly account for the piecewise smooth nature of most imaging data. We develop an efficient penalized total variation optimization to estimate the unknown slope function and other parameters. We also establish nonasymptotic error bounds on the excess risk. These bounds are explicitly specified in terms of sample size, image size, and image smoothness. Our simulations demonstrate a superior performance of GSIRM-TV against many existing approaches. We apply GSIRM-TV to the analysis of hippocampus data obtained from the Alzheimers Disease Neuroimaging Initiative (ADNI) dataset. Supplementary materials for this article are available online. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
12. Weighted least squares estimation for exchangeable binary data
- Author
-
Bowman, Dale and George, E. Olusegun
- Published
- 2016
- Full Text
- View/download PDF
13. Spatiotemporal analysis of dengue fever in Nepal from 2010 to 2014.
- Author
-
Acharya, Bipin Kumar, ChunXiang Cao, Lakes, Tobia, Wei Chen, Naeem, Shahid, Cao, ChunXiang, and Chen, Wei
- Subjects
SPATIOTEMPORAL processes ,DENGUE ,CLUSTER analysis (Statistics) ,SPACETIME ,COMPUTER software ,PUBLIC health ,PUBLIC health surveillance ,STATISTICS ,DISEASE incidence - Abstract
Background: Due to recent emergence, dengue is becoming one of the major public health problems in Nepal. The numbers of reported dengue cases in general and the area with reported dengue cases are both continuously increasing in recent years. However, spatiotemporal patterns and clusters of dengue have not been investigated yet. This study aims to fill this gap by analyzing spatiotemporal patterns based on monthly surveillance data aggregated at district.Methods: Dengue cases from 2010 to 2014 at district level were collected from the Nepal government's health and mapping agencies respectively. GeoDa software was used to map crude incidence, excess hazard and spatially smoothed incidence. Cluster analysis was performed in SaTScan software to explore spatiotemporal clusters of dengue during the above-mentioned time period.Results: Spatiotemporal distribution of dengue fever in Nepal from 2010 to 2014 was mapped at district level in terms of crude incidence, excess risk and spatially smoothed incidence. Results show that the distribution of dengue fever was not random but clustered in space and time. Chitwan district was identified as the most likely cluster and Jhapa district was the first secondary cluster in both spatial and spatiotemporal scan. July to September of 2010 was identified as a significant temporal cluster.Conclusion: This study assessed and mapped for the first time the spatiotemporal pattern of dengue fever in Nepal. Two districts namely Chitwan and Jhapa were found highly affected by dengue fever. The current study also demonstrated the importance of geospatial approach in epidemiological research. The initial result on dengue patterns and risk of this study may assist institutions and policy makers to develop better preventive strategies. [ABSTRACT FROM AUTHOR]- Published
- 2016
- Full Text
- View/download PDF
14. Excess risk bounds in robust empirical risk minimization
- Author
-
Timothée Mathieu, Stanislav Minsker, Department of Mathematics [Los Angeles], University of Southern California (USC), Laboratoire de Mathématiques d'Orsay (LMO), Université Paris-Sud - Paris 11 (UP11)-Centre National de la Recherche Scientifique (CNRS), Statistique mathématique et apprentissage (CELESTE), Inria Saclay - Ile de France, Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire de Mathématiques d'Orsay (LMO), and Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS)-Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS)
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Computer Science - Machine Learning ,Machine Learning (stat.ML) ,Sample (statistics) ,02 engineering and technology ,01 natural sciences ,Machine Learning (cs.LG) ,Combinatorics ,010104 statistics & probability ,[STAT.ML]Statistics [stat]/Machine Learning [stat.ML] ,Statistics - Machine Learning ,0202 electrical engineering, electronic engineering, information engineering ,Empirical risk minimization ,0101 mathematics ,Mathematics ,Numerical Analysis ,Median-of-means ,Stochastic process ,Applied Mathematics ,Excess risk ,Estimator ,Classification ,Regression ,Robust estimation ,Computational Theory and Mathematics ,Sample size determination ,Outlier ,020201 artificial intelligence & image processing ,62G35 ,Marginal distribution ,Analysis - Abstract
This paper investigates robust versions of the general empirical risk minimization algorithm, one of the core techniques underlying modern statistical methods. Success of the empirical risk minimization is based on the fact that for a ‘well-behaved’ stochastic process $\left \{ f(X), \ f\in \mathscr F\right \}$ indexed by a class of functions $f\in \mathscr F$, averages $\frac{1}{N}\sum _{j=1}^N f(X_j)$ evaluated over a sample $X_1,\ldots ,X_N$ of i.i.d. copies of $X$ provide good approximation to the expectations $\mathbb E f(X)$, uniformly over large classes $f\in \mathscr F$. However, this might no longer be true if the marginal distributions of the process are heavy tailed or if the sample contains outliers. We propose a version of empirical risk minimization based on the idea of replacing sample averages by robust proxies of the expectations and obtain high-confidence bounds for the excess risk of resulting estimators. In particular, we show that the excess risk of robust estimators can converge to $0$ at fast rates with respect to the sample size $N$, referring to the rates faster than $N^{-1/2}$. We discuss implications of the main results to the linear and logistic regression problems and evaluate the numerical performance of proposed methods on simulated and real data.
- Published
- 2021
15. Deep determinism and the assessment of mechanistic interaction.
- Author
-
Berzuini, Carlo and Dawid, A. Philip
- Subjects
HEART diseases ,SYMMETRY (Biology) ,ACQUISITION of data ,BIOMECHANICS ,EPISTASIS (Genetics) ,BIOMETRY - Abstract
Given two variables that causally influence a binary response, we formalize the idea that their effects operate through a common mechanism, in which case we say that the two variables interact mechanistically. We introduce a mechanistic interaction relationship of “interference” that is asymmetric in the two causal factors. Conditions and assumptions under which such mechanistic interaction can be tested under a given regime of data collection, be it interventional or observational, are expressed in terms of conditional independence relationships between the problem variables, which can be manipulated with the aid of causal diagrams. The proposed method is able, under appropriate conditions, to test for interaction between direct effects, and to deal with the situation where one of the two factors is a dichotomized version of a continuous variable. The method is illustrated with the aid of a study on heart disease. [ABSTRACT FROM AUTHOR]
- Published
- 2013
- Full Text
- View/download PDF
16. Two-sample goodness-of-fit tests for additive risk models with censored observations.
- Author
-
KIM, JINHEUM and LEE, SEUNG-YEOUN
- Subjects
GOODNESS-of-fit tests ,STATISTICS ,STATISTICAL hypothesis testing ,CENSORING (Statistics) ,MARTINGALES (Mathematics) - Abstract
The additive risk model assumes that the hazard function associated with a set of covariates is the sum of the baseline hazard function and the regression function of covariates. We propose two different test procedures for checking the adequacy of two-sample additive risk models for randomly censored observations. One is based on the martingale residuals and the other on the difference between weighted estimators of the excess risk. The test statistics are shown to be asymptotically normal under appropriate regularity conditions and consistent under any model misspecifications. Finally, two real examples are provided, along with results of a simulation study. [ABSTRACT FROM AUTHOR]
- Published
- 1998
- Full Text
- View/download PDF
17. Improved classification rates under refined margin conditions
- Author
-
Ingo Steinwart and Ingrid Blaschzyk
- Subjects
Statistics and Probability ,fast rates of convergence ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,62H30, 62G20, 68T05 ,Statistical learning ,68T05 ,Support vector machine ,Set (abstract data type) ,Statistical classification ,histogram rule ,classification ,Margin (machine learning) ,Histogram ,FOS: Mathematics ,Decision boundary ,Noise (video) ,Statistics, Probability and Uncertainty ,excess risk ,Algorithm ,62H30 ,62G20 ,Ansatz ,Mathematics - Abstract
In this paper we present a simple partitioning based technique to refine the statistical analysis of classification algorithms. The core of our idea is to divide the input space into two parts such that the first part contains a suitable vicinity around the decision boundary, while the second part is sufficiently far away from the decision boundary. Using a set of margin conditions we are then able to control the classification error on both parts separately. By balancing out these two error terms we obtain a refined error analysis in a final step. We apply this general idea to the histogram rule and show that even for this simple method we obtain, under certain assumptions, better rates than the ones known for support vector machines, for certain plug-in classifiers, and for a recently analyzed tree based adaptive-partitioning ansatz. Moreover, we show that a margin condition which sets the critical noise in relation to the decision boundary makes it possible to improve the optimal rates proven for distributions without this margin condition., 32 pages
- Published
- 2018
18. Communicating epidemiological results through alternative indicators: Cognitive interviewing to assess a questionnaire on risk perception in a high environmental risk area
- Author
-
Domenica Farinella, Annibale Biggeri, Gianna Terni, and Michela Baccini
- Subjects
medicine.medical_specialty ,Applied psychology ,cognitive interviews ,050109 social psychology ,high risk area ,cognitive interviewing ,lcsh:Social Sciences ,risk communication, questionnaire validation, cognitive interviews, high risk area, Livorno, risk perception, excess risk, time needed to harm ,03 medical and health sciences ,0302 clinical medicine ,Environmental risk ,risk communication ,risk perception ,Epidemiology ,medicine ,Risk communication ,pollution ,0501 psychology and cognitive sciences ,030212 general & internal medicine ,Cognitive interview ,questionnaire validation ,05 social sciences ,General Social Sciences ,statistical uncertainty ,Cognition ,health ,Summary statistics ,Risk perception ,lcsh:H ,Livorno ,excess risk ,time needed to harm ,Cognitive interviewing ,environment ,health impact ,Psychology ,Social psychology ,Environmental epidemiology - Abstract
Participatory approaches to environmental research and decision-making require that all social stakeholders are involved from the onset of the debate. In such a setting, communication among different expertise is crucial, but language and technicalities may represent a barrier. In the clinical setting, decisions regarding treatment preferences may be influenced by the summary statistics used, but, according to the literature, no study has compared different statistical indicators for risk communication in environmental epidemiology. In this paper, we report on the qualitative results of the cognitive interviews conducted for assessing two questionnaires devoted to investigating risk perception when selected epidemiological results are communicated, by using different statistical indicators of health impact and uncertainty. The initial questionnaires were tested on 15 people residing in the high environmental risk area of Livorno (Italy). Cognitive interviewing led to substantial revision of the initial drafts. Moreover, it highlighted the difficulty of communicating statistical uncertainty and the need to account for the complex interaction between mathematical skills, affective factors and individual a priori knowledge on environmental risk perception.
- Published
- 2017
19. Binomial Distribution Sample Confidence Intervals Estimation 6. Excess Risk
- Author
-
Sorana BOLBOACĂ and Andrei ACHIMAŞ CADARIU
- Subjects
Excess risk ,Confidence interval estimation ,lcsh:Electronic computers. Computer science ,Risk factors assessments ,lcsh:QA75.5-76.95 - Abstract
We present the problem of the confidence interval estimation for excess risk (Y/n-X/m fraction), a parameter which allows evaluating of the specificity of an association between predisposing or causal factors and disease in medical studies. The parameter is computes based on 2x2 contingency table and qualitative variables. The aim of this paper is to introduce four new methods of computing confidence intervals for excess risk called DAC, DAs, DAsC, DBinomial, and DBinomialC and to compare theirs performance with the asymptotic method called here DWald.In order to assess the methods, we use the PHP programming language and a PHP program was creates. The performance of each method for different sample sizes and different values of binomial variables were assess using a set of criterions. First, the upper and lower boundaries for a given X, Y and a specified sample size for choused methods were compute. Second, were assessed the average and standard deviation of the experimental errors, and the deviation relative to imposed significance level α = 5%. Four methods were assessed on random numbers for binomial variables and for sample sizes from 4 to 1000 domain.The experiments show that the DAC methods obtain performances in confidence intervals estimation for excess risk.
- Published
- 2004
20. On universal oracle inequalities related to high-dimensional linear models
- Author
-
Yuri Golubev
- Subjects
Statistics and Probability ,Spectral regularization ,62C10 ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,computer.software_genre ,oracle inequality ,Regularization (mathematics) ,symbols.namesake ,FOS: Mathematics ,Applied mathematics ,62G05 ,Empirical risk minimization ,Mathematics ,Numerical linear algebra ,empirical risk minimization ,Linear model ,ordered smoother ,White noise ,Additive white Gaussian noise ,Gaussian noise ,symbols ,Statistics, Probability and Uncertainty ,excess risk ,Spectral method ,computer - Abstract
This paper deals with recovering an unknown vector $\theta$ from the noisy data $Y=A\theta+\sigma\xi$, where $A$ is a known $(m\times n)$-matrix and $\xi$ is a white Gaussian noise. It is assumed that $n$ is large and $A$ may be severely ill-posed. Therefore, in order to estimate $\theta$, a spectral regularization method is used, and our goal is to choose its regularization parameter with the help of the data $Y$. For spectral regularization methods related to the so-called ordered smoothers [see Kneip Ann. Statist. 22 (1994) 835--866], we propose new penalties in the principle of empirical risk minimization. The heuristical idea behind these penalties is related to balancing excess risks. Based on this approach, we derive a sharp oracle inequality controlling the mean square risks of data-driven spectral regularization methods., Comment: Published in at http://dx.doi.org/10.1214/10-AOS803 the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org)
- Published
- 2010
21. Nonparametric Tests of the Markov Model for Survival Data
- Author
-
Jones, Michael P. and Crowley, John
- Published
- 1992
- Full Text
- View/download PDF
22. Semiparametric Analysis of the Additive Risk Model
- Author
-
Lin, D. Y. and Ying, Zhiliang
- Published
- 1994
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.