Back to Search
Start Over
Scalable kernel two-sample tests via empirical likelihood and jackknife.
- Source :
-
Communications in Statistics: Simulation & Computation . 2023, Vol. 52 Issue 12, p5975-5990. 16p. - Publication Year :
- 2023
-
Abstract
- Two-sample hypothesis testing is a classical problem in statistics. Given i.i.d. samples X1, :::, Xm ~ P and Y1, :::, Yn ~ Q, the two-sample testing is to test whether P is significantly different from Q or not. Among the various statistical tests, the maximum mean discrepancy (MMD) has been proven to have excellent performance compared to traditional methods. In application of MMD, we are confronted with two challenges: (I) For large data set, MMD is computationally demanding with computational cost O((m + n)²); (II) The limiting distribution under the null hypothesis usually contains infinitely many unknown parameters. In this paper, we propose two testing methods that solve (I) and (II) respectively. To address (I), the data are independently and evenly divided into k groups, calculate the MMD on each group and then combine them by empirical likelihood to yield a test statistic Tk: It converges in distribution to the chi-square distribution with degree of freedom one (X1²) under the null hypothesis as k goes to infinity. To address (II), we propose to compute the two-sample jackknife pseudo-values on each group and combine them by empirical likelihood to get a new test statistic Rnm: Under the null hypothesis, Rnm converges in distribution to X1² if k = o(n): The simulation and real data application indicates that the proposed tests performs well. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 03610918
- Volume :
- 52
- Issue :
- 12
- Database :
- Academic Search Index
- Journal :
- Communications in Statistics: Simulation & Computation
- Publication Type :
- Academic Journal
- Accession number :
- 174588689
- Full Text :
- https://doi.org/10.1080/03610918.2021.2005096