Back to Search Start Over

Scalable kernel two-sample tests via empirical likelihood and jackknife.

Authors :
Qian Wen
Mingao Yuan
Source :
Communications in Statistics: Simulation & Computation. 2023, Vol. 52 Issue 12, p5975-5990. 16p.
Publication Year :
2023

Abstract

Two-sample hypothesis testing is a classical problem in statistics. Given i.i.d. samples X1, :::, Xm ~ P and Y1, :::, Yn ~ Q, the two-sample testing is to test whether P is significantly different from Q or not. Among the various statistical tests, the maximum mean discrepancy (MMD) has been proven to have excellent performance compared to traditional methods. In application of MMD, we are confronted with two challenges: (I) For large data set, MMD is computationally demanding with computational cost O((m + n)²); (II) The limiting distribution under the null hypothesis usually contains infinitely many unknown parameters. In this paper, we propose two testing methods that solve (I) and (II) respectively. To address (I), the data are independently and evenly divided into k groups, calculate the MMD on each group and then combine them by empirical likelihood to yield a test statistic Tk: It converges in distribution to the chi-square distribution with degree of freedom one (X1²) under the null hypothesis as k goes to infinity. To address (II), we propose to compute the two-sample jackknife pseudo-values on each group and combine them by empirical likelihood to get a new test statistic Rnm: Under the null hypothesis, Rnm converges in distribution to X1² if k = o(n): The simulation and real data application indicates that the proposed tests performs well. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
03610918
Volume :
52
Issue :
12
Database :
Academic Search Index
Journal :
Communications in Statistics: Simulation & Computation
Publication Type :
Academic Journal
Accession number :
174588689
Full Text :
https://doi.org/10.1080/03610918.2021.2005096