Next Article in Journal
Community Detection Method Based on Node Density, Degree Centrality, and K-Means Clustering in Complex Network
Next Article in Special Issue
Model Selection in a Composite Likelihood Framework Based on Density Power Divergence
Previous Article in Journal
OTEC Maximum Net Power Output Using Carnot Cycle and Application to Simplify Heat Exchanger Selection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Convergence Rates for Empirical Estimation of Binary Classification Bounds

1
School of Computing and Information Science, University of Maine, Orono, ME 04469, USA
2
Department of Electrical Engineering and Computer Science, University of Michigan, Ann Arbor, MI 48109, USA
3
Department of Mathematics and Statistics, Utah State University, Logan, UT 84322, USA
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(12), 1144; https://doi.org/10.3390/e21121144
Submission received: 5 November 2019 / Accepted: 15 November 2019 / Published: 23 November 2019

Abstract

:
Bounding the best achievable error probability for binary classification problems is relevant to many applications including machine learning, signal processing, and information theory. Many bounds on the Bayes binary classification error rate depend on information divergences between the pair of class distributions. Recently, the Henze–Penrose (HP) divergence has been proposed for bounding classification error probability. We consider the problem of empirically estimating the HP-divergence from random samples. We derive a bound on the convergence rate for the Friedman–Rafsky (FR) estimator of the HP-divergence, which is related to a multivariate runs statistic for testing between two distributions. The FR estimator is derived from a multicolored Euclidean minimal spanning tree (MST) that spans the merged samples. We obtain a concentration inequality for the Friedman–Rafsky estimator of the Henze–Penrose divergence. We validate our results experimentally and illustrate their application to real datasets.

1. Introduction

Divergence measures between probability density functions are used in many signal processing applications including classification, segmentation, source separation, and clustering (see [1,2,3]). For more applications of divergence measures, we refer to [4].
In classification problems, the Bayes error rate is the expected risk for the Bayes classifier, which assigns a given feature vector x to the class with the highest posterior probability. The Bayes error rate is the lowest possible error rate of any classifier for a particular joint distribution. Mathematically, let x 1 , x 2 , , x N R d be realizations of random vector X and class labels S { 0 , 1 } , with prior probabilities p = P ( S = 0 ) and q = P ( S = 1 ) , such that p + q = 1 . Given conditional probability densities f 0 ( x ) and f 1 ( x ) , the Bayes error rate is given by
ϵ = R d min p f 0 ( x ) , q f 1 ( x ) d x .
The Bayes error rate provides a measure of classification difficulty. Thus, when known, the Bayes error rate can be used to guide the user in the choice of classifier and tuning parameter selection. In practice, the Bayes error is rarely known and must be estimated from data. Estimation of the Bayes error rate is difficult due to the nonsmooth min function within the integral in (1). Thus, research has focused on deriving tight bounds on the Bayes error rate based on smooth relaxations of the min function. Many of these bounds can be expressed in terms of divergence measures such as the Bhattacharyya [5] and Jensen–Shannon [6]. Tighter bounds on the Bayes error rate can be obtained using an important divergence measure known as the Henze–Penrose (HP) divergence [7,8].
Many techniques have been developed for estimating divergence measures. These methods can be broadly classified into two categories: (i) plug-in estimators in which we estimate the probability densities and then plug them in the divergence function [9,10,11,12], (ii) entropic graph approaches, in which the relationship between the divergence function and a graph functional in Euclidean space is derived [8,13]. Examples of plug-in methods include k-nearest neighbor (K-NN) and Kernel density estimator (KDE) divergence estimators. Examples of entropic graph approaches include methods based on minimal spanning trees (MST), K-nearest neighbors graphs (K-NNG), minimal matching graphs (MMG), traveling salesman problem (TSP), and their power-weighted variants.
Disadvantages of plug-in estimators are that these methods often require assumptions on the support set boundary and are more computationally complex than direct graph-based approaches. Thus, for practical and computational reasons, the asymptotic behavior of entropic graph approaches has been of great interest. Asymptotic analysis has been used to justify graph based approaches. For instance, in [14], the authors showed that a cross match statistic based on optimal weighted matching converges to the the HP-divergence. In [15], a more complex approach based on the K-NNG was proposed that also converges to the HP-divergence.
The first contribution of our paper is that we obtain a bound on the convergence rates for the Friedman and Rafsky (FR) estimator of the HP-divergence, which is based on a multivariate extension of the non-parametric run length test of equality of distributions. This estimator is constructed using a multicolored MST on the labeled training set where MST edges connecting samples with dichotomous labels are colored differently from edges connecting identically labeled samples. While previous works have investigated the FR test statistic in the context of estimating the HP-divergence (see [8,16]), to the best of our knowledge, its minimax MSE convergence rate has not been previously derived. The bound on convergence rate is established by using the umbrella theorem of [17], for which we define a dual version of the multicolor MST. The proposed dual MST in this work is different than the standard dual MST introduced by Yukich in [17]. We show that the bias rate of the FR estimator is bounded by a function of N, η and d, as O ( N ) η 2 / ( d ( η + 1 ) ) , where N is the total sample size, d is the dimension of the data samples d 2 , and η is the Hölder smoothness parameter 0 < η 1 . We also obtain the variance rate bound as O ( N ) 1 .
The second contribution of our paper is a new concentration bound for the FR test statistic. The bound is obtained by establishing a growth bound and a smoothness condition for the multicolored MST. Since the FR test statistic is not a Euclidean functional, we cannot use the standard subadditivity and superadditivity approaches of [17,18,19]. Our concentration inequality is derived using a different Hamming distance approach and a dual graph to the multicolored MST.
We experimentally validate our theoretic results. We compare the MSE theory and simulation in three experiments with various dimensions d = 2 , 4 , 8 . We observe that, in all three experiments, as sample size increases, the MSE rate decreases and, for higher dimensions, the rate is slower. In all sets of experiments, our theory matches the experimental results. Furthermore, we illustrate the application of our results on estimation of the Bayes error rate on three real datasets.

1.1. Related Work

Much research on minimal graphs has focused on the use of Euclidean functionals for signal processing and statistics applications such as image registration [20,21], pattern matching [22], and non-parametric divergence estimation [23]. A K-NNG-based estimator of Rényi and f-divergence measures has been proposed in [13]. Additional examples of direct estimators of divergence measures include statistic based on the nonparametric two sample problem, the Smirnov maximum deviation test [24], and the Wald–Wolfowitz [25] runs test, which have been studied in [26].
Many entropic graph estimators such as MST, K-NNG, MMG, and TSP have been considered for multivariate data from a single probability density f. In particular, the normalized weight function of graph constructions all converge almost surely to the Rényi entropy of f [17,27]. For N uniformly distributed points, the MSE is O ( N 1 / d ) [28,29]. Later, Hero et al. [30,31] reported bounds on L γ -norm bias convergence rates of power-weighted Euclidean weight functionals of order γ for densities f belonging to the space of Hölder continuous functions Σ d ( η , K ) as O N α η / ( α η + 1 ) 1 / d , where 0 < η 1 , d 1 , γ ( 1 , d ) , and α = ( d γ ) / d . In this work, we derive a bound on convergence rate of FR estimator for the HP-divergence when the density functions belong to the Hölder class, Σ d ( η , K ) , for 0 < η 1 , d 2 [32]. Note that throughout the paper we assume the density functions are absolutely continuous and bounded with support on the unit cube [ 0 , 1 ] d .
In [28], Yukich introduced the general framework of continuous and quasi-additive Euclidean functionals. This has led to many convergence rate bounds of entropic graph divergence estimators.
The framework of [28] is as follows: Let F be finite subset of points in [ 0 , 1 ] d , d 2 , drawn from an underlying density. A real-valued function L γ defined on F is called a Euclidean functional of order γ if it is of the form L γ ( F ) = min E E e E | e ( F ) | γ , where E is a set of graphs, e is an edge in the graph E, | e | is the Euclidean length of e, and γ is called the edge exponent or power-weighting constant. The MST, TSP, and MMG are some examples for which γ = 1 .
Following this framework, we show that the FR test statistic satisfies the required continuity and quasi-additivity properties to obtain similar convergence rates to those predicted in [28]. What distinguishes our work from previous work is that the count of dichotomous edges in the multicolored MST is not Euclidean. Therefore, the results in [17,27,30,31] are not directly applicable.
Using the isoperimetric approach, Talagrand [33] showed that, when the Euclidean functional L γ is based on the MST or TSP, then the functional L γ for derived random vertices uniformly distributed in a hypercube [ 0 , 1 ] d is concentrated around its mean. Namely, with high probability, the functional L γ and its mean do not differ by more than C ( N log N ) ( d γ ) / 2 d . In this paper, we establish concentration bounds for the FR statistic: with high probability 1 δ , the FR statistic differs from its mean by not more than O ( N ) ( d 1 ) / d log ( C / δ ) ( d 1 ) / d , where C is a function of N and d.

1.2. Organization

This paper is organized as follows. In Section 2, we first introduce the HP-divergence and the FR multivariate test statistic. We then present the bias and variance rates of the FR-based estimator of HP-divergence followed by the concentration bounds and the minimax MSE convergence rate. Section 3 provides simulations that validate the theory. All proofs and relevant lemmas are given in the Appendix A, Appendix B, Appendix C, Appendix D and Appendix E.
Throughout the paper, we denote expectation by E and variance by abbreviation Var . Bold face type indicates random variables. In this paper, when we say number of samples we mean number of observations.

2. The Henze–Penrose Divergence Measure

Consider parameters p ( 0 , 1 ) and q = 1 p . We focus on estimating the HP-divergence measure between distributions f 0 and f 1 with domain R d defined by
D p ( f 0 , f 1 ) = 1 4 p q p f 0 ( x ) q f 1 ( x ) 2 p f 0 ( x ) + q f 1 ( x ) d x ( p q ) 2 .
It can be verified that this measure is bounded between 0 and 1 and, if f 0 ( x ) = f 1 ( x ) , then D p = 0 . In contrast with some other divergences such as the Kullback–Liebler [34] and Rényi divergences [35], the HP-divergence is symmetrical, i.e., D p ( f 0 , f 1 ) = D q ( f 1 , f 0 ) . By invoking relation (3) in [8],
p f 0 ( x ) q f 1 ( x ) 2 p f 0 ( x ) + q f 1 ( x ) d x = 1 4 p q A p ( f 0 , f 1 ) ,
where
A p ( f 0 , f 1 ) = f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x = E f 0 p f 0 ( X ) f 1 ( X ) + q 1 , u p ( f 0 , f 1 ) = 1 4 p q A p ( f 0 , f 1 ) ,
one can rewrite D p in the alternative form:
D p ( f 0 , f 1 ) = 1 A p ( f 0 , f 1 ) = u p ( f 0 , f 1 ) 4 p q ( p q ) 2 4 p q .
Throughout the paper, we refer to A p ( f 0 , f 1 ) as the HP-integral. The HP-divergence measure belongs to the class of ϕ -divergences [36]. For the special case p = 0.5 , the divergence (2) becomes the symmetric χ 2 -divergence and is similar to the Rukhin f-divergence. See [37,38].

2.1. The Multivariate Runs Test Statistic

The MST is a graph of minimum weight among all graphs E that span n vertices. The MST has many applications including pattern recognition [39], clustering [40], nonparametric regression [41], and testing of randomness [42]. In this section, we focus on the FR multivariate two sample test statistic constructed from the MST.
Assume that sample realizations from f 0 and f 1 , denoted by X m R m × d and Y n R n × d , respectively, are available. Construct an MST spanning the samples from both f 0 and f 1 and color the edges in the MST that connect dichotomous samples green and color the remaining edges black. The FR test statistic R m , n : = R m , n ( X m , Y n ) is the number of green edges in the MST. Note that the test assumes a unique MST, therefore all inter point distances between data points must be distinct. We recall the following theorem from [7,8]:
Theorem 1.
As m and n such that m n + m p and n n + m q ,
1 R m , n ( X m , Y n ) m + n 2 m n D p ( f 0 , f 1 ) , a . s .
In the next section, we obtain bounds on the MSE convergence rates of the FR approximation for HP-divergence between densities that belong to Σ d ( η , K ) , the class of Hölder continuous functions with Lipschitz constant K and smoothness parameter 0 < η 1 [32]:
Definition 1
(Hölder class). Let X R d be a compact space. The Hölder class Σ d ( η , K ) , with η-Hölder parameter, of functions with the L d -norm, consists of the functions g that satisfy
g : g ( z ) p x η ( z ) d K x z d η , x , z X ,
where p x k ( z ) is the Taylor polynomial (multinomial) of g of order k expanded about the point x and η is defined as the greatest integer strictly less than η.
In what follows, we will use both notations R m , n and R m , n ( X m , Y n ) for the FR statistic over the combined samples.

2.2. Convergence Rates

In this subsection, we obtain the mean convergence rate bounds for general non-uniform Lebesgue densities f 0 and f 1 belonging to the Hölder class Σ d ( η , K ) . Since the expectation of R m , n can be closely approximated by the sum of the expectation of the FR statistic constructed on a dense partition of [ 0 , 1 ] d , R m , n is a quasi-additive functional in mean. The family of bounds (A16) in Appendix B enables us to achieve the minimax convergence rate for the mean under the Hölder class assumption with smoothness parameter 0 < η 1 , d 2 :
Theorem 2
(Convergence Rate of the Mean). Let d 2 , and R m , n be the FR statistic for samples drawn from Hölder continuous and bounded density functions f 0 and f 1 in Σ d ( η , K ) . Then, for d 2 ,
E R m , n m + n 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x O ( m + n ) η 2 / ( d ( η + 1 ) ) .
This bound holds over the class of Lebesgue densities f 0 , f 1 Σ d ( η , K ) , 0 < η 1 . Note that this assumption can be relaxed to f 0 Σ d s ( η , K 0 ) and f 1 Σ d s ( η , K 1 ) that is Lebesgue densities f 0 and f 1 belong to the Strong Hölder class with the same Hölder parameter η and different constants K 0 and K 1 , respectively.
The following variance bound uses the Efron–Stein inequality [43]. Note that in Theorem 3 we do not impose any strict assumptions. We only assume that the density functions are absolutely continuous and bounded with support on the unit cube [ 0 , 1 ] d . Appendix C contains the proof.
Theorem 3.
The variance of the HP-integral estimator based on the FR statistic, R m , n / ( m + n ) is bounded by
V a r R m , n ( X m , Y n ) m + n 32 c d 2 q ( m + n ) ,
where the constant c d depends only on d.
By combining Theorems 2 and 3, we obtain the MSE rate of the form O m + n ) η 2 / ( d ( η + 1 ) ) + O ( m + n ) 1 . Figure 1 indicates a heat map showing the MSE rate as a function of d and N = m = n . The heat map shows that the MSE rate of the FR test statistic-based estimator given in (3) is small for large sample size N.

2.3. Proof Sketch of Theorem 2

In this subsection, we first establish subadditivity and superadditivity properties of the FR statistic, which will be employed to derive the MSE convergence rate bound. This will establish that the mean of the FR test statistic is a quasi-additive functional:
Theorem 4.
Let R m , n ( X m , Y n ) be the number of edges that link nodes from differently labeled samples X m = { X 1 , , X m } and Y n = { Y 1 , , Y n } in [ 0 , 1 ] d . Partition [ 0 , 1 ] d into l d equal volume subcubes Q i such that m i and n i are the number of samples from { X 1 , , X m } and { Y 1 , , Y n } , respectively, that fall into the partition Q i . Then, there exists a constant c 1 such that
E R m , n ( X m , Y n ) i = 1 l d E R m i , n i ( X m , Y n ) Q i + 2 c 1 l d 1 ( m + n ) 1 / d .
Here, R m i , n i is the number of dichotomous edges in partition Q i . Conversely, for the same conditions as above on partitions Q i , there exists a constant c 2 such that
E R m , n ( X m , Y n ) i = 1 l d E R m i , n i ( X m , Y n ) Q i 2 c 2 l d 1 ( m + n ) 1 / d .
The inequalities (7) and (8) are inspired by corresponding inequalities in [30,31]. The full proof is given in Appendix A. The key result in the proof is the inequality:
R m , n ( X m , Y n ) i = 1 l d R m i , n i ( X m , Y n ) Q i + 2 | D | ,
where | D | indicates the number of all edges of the MST which intersect two different partitions.
Furthermore, we adapt the theory developed in [17,30] to derive the MSE convergence rate of the FR statistic-based estimator by defining a dual MST and dual FR statistic, denoted by MST * and R m , n * respectively (see Figure 2):
Definition 2
(Dual MST, MST * and dual FR statistic R m , n * ). Let F i be the set of corner points of the subsection Q i for 1 i l d . Then, we define MST * ( X m Y n Q i ) as the boundary MST graph of partition Q i [17], which contains X m and Y n points falling inside the section Q i and those corner points in F i which minimize total MST length. Notice it is allowed to connect the MSTs in Q i and Q j through points strictly contained in Q i and Q j and corner points are taken into account under condition of minimizing total MST length. Another word, the dual MST can connect the points in Q i Q j by direct edges to pair to another point in Q i Q j or the corner the corner points (we assume that all corner points are connected) in order to minimize the total length. To clarify this, assume that there are two points in Q i Q j , then the dual MST consists of the two edges connecting these points to the corner if they are closed to a corner point; otherwise, dual MST consists of an edge connecting one to another. Furthermore, we define R m , n * ( X m , Y n Q i ) as the number of edges in an MST * graph connecting nodes from different samples and number of edges connecting to the corner points. Note that the edges connected to the corner nodes (regardless of the type of points) are always counted in dual FR test statistic R m , n * .
In Appendix B, we show that the dual FR test statistic is a quasi-additive functional in mean and R m , n * ( X m , Y n ) R m , n ( X m , Y n ) . This property holds true since MST ( X m , Y n ) and MST * ( X m , Y n ) graphs can only be different in the edges connected to the corner nodes, and in R * ( X m , Y n ) we take all of the edges between these nodes and corner nodes into account.
To prove Theorem 2, we partition [ 0 , 1 ] d into l d subcubes. Then, by applying Theorem 4 and the dual MST, we derive the bias rate in terms of partition parameter l (see (A16) in Theorem A1). See Appendix B and Appendix E for details. According to (A16), for d 2 , and l = 1 , 2 , , the slowest rates as a function of l are l d ( m + n ) η / d and l η d . Therefore, we obtain an l-independent bound by letting l be a function of m + n that minimizes the maximum of these rates i.e.,
l ( m + n ) = a r g min l max l d ( m + n ) η / d , l η d .
The full proof of the bound in (2) is given in Appendix B.

2.4. Concentration Bounds

Another main contribution of our work in this part is to provide an exponential inequality convergence bound derived for the FR estimator of the HP-divergence. The error of this estimator can be decomposed into a bias term and a variance-like term via the triangle inequality:
R m , n f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x R m , n E R m , n variance - like term + E R m , n f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x bias term .
The bias bound was given in Theorem 2. Therefore, we focus on an exponential concentration bound for the variance-like term. One application of concentration bounds is to employ these bounds to compare confidence intervals on the HP-divergence measure in terms of the FR estimator. In [44,45], the authors provided an exponential inequality convergence bound for an estimator of Rény divergence for a smooth Hölder class of densities on the d-dimensional unite cube [ 0 , 1 ] d . We show that if X m and Y n are the set of m and n points drawn from any two distributions f 0 and f 1 , respectively, the FR criteria R m , n is tightly concentrated. Namely, we establish that, with high probability, R m , n is within
1 O ( m + n ) 2 / d ϵ * 2
of its expected value, where ϵ * is the solution of the following convex optimization problem:
min ϵ 0 C m , n ( ϵ ) exp ( t / ( 2 ϵ ) ) d / ( d 1 ) ( m + n ) C ˜ subject to ϵ O 7 d + 1 ( m + n ) 1 / d ,
where C ˜ = 8 ( 4 ) d / ( d 1 ) and
C m , n ( ϵ ) = 8 1 O ( m + n ) 2 / d ϵ 2 2 .
Note that, under the assumption ( m + n ) 1 / d 1 , C m , n ( ϵ ) becomes a constant depending only on ϵ by 8 1 ( c ϵ 2 2 , where c is a constant. This is inferred from Theorems 5 and 6 below as ( m + n ) 1 / d 1 . See Appendix D, specifically Lemmas A8–A12 for more detail. Indeed, we first show the concentration around the median. A median is by definition any real number M e that satisfies the inequalities P ( X M e ) 1 / 2 and P ( X M e ) 1 / 2 . To derive the concentration results, the properties of growth bounds and smoothness for R m , n , given in Appendix D, are exploited.
Theorem 5
(Concentration around the median). Let M e be a median of R m , n which implies that P R m , n M e 1 / 2 . Recall ϵ * from (9) then we have
P | R m , n ( X m , Y n ) M e | t C m , n ( ϵ * ) exp ( t / ϵ * ) d / ( d 1 ) ( m + n ) C ˜ ,
where C ˜ = 8 ( 4 ) d / ( d 1 ) .
Theorem 6
(Concentration of R m , n around the mean). Let R m , n be the FR statistic. Then,
P | R m , n E [ R m , n ] | t C m , n ( ϵ * ) exp ( t / ( 2 ϵ * ) ) d / ( d 1 ) ( m + n ) C ˜ .
Here, C ˜ = 8 ( 4 ) d / ( d 1 ) and the explicit form for C m , n ( ϵ * ) is given by (10) when ϵ = ϵ * .
See Appendix D for full proofs of Theorems 5 and 6. Here, we sketch the proofs. The proof of the concentration inequality for R m , n , Theorem 6, requires involving the median M e , where P ( R m , n M e ) 1 / 2 , inside the probability term by using
| R m , n E [ R m , n ] | | R m , n M e | + | E [ R m , n ] M e | .
To prove the expressions for the concentration around the median, Theorem 5, we first consider the h d uniform partitions of [ 0 , 1 ] d , with edges parallel to the coordinate axes having edge lengths h 1 and volumes h d . Then, by applying the Markov inequality, we show that with at least probability 1 δ m , n h / ϵ , where δ m , n h = O h d 1 ( m + n ) 1 / d , the FR statistic R m , n is subadditive with 2 ϵ threshold. Afterward, owing to the induction method [17], the growth bound can be derived with at least probability 1 h δ m , n h / ϵ . The growth bound explains that with high probability there exists a constant depending on ϵ and h, C ϵ , h , such that R m , n C ϵ , h m n 1 1 / d . Applying the law of total probability and semi-isoperimetric inequality (A108) in Lemma A11 gives us (A35). By considering the solution to convex optimization problem (9), i.e., ϵ * and optimal h = 7 the claimed results (11) and (12) are derived. The only constraint here is that ϵ is lower bounded by a function of δ m , n h = O h d 1 ( m + n ) 1 / d .
Next, we provide a bound for the variance-like term with high probability at least 1 δ . According to the previous results, we expect that this bound depends on ϵ * , d, m and n. The proof is short and is given in Appendix D.
Theorem 7
(Variance-like bound for R m , n ). Let R m , n be the FR statistic. With at least probability 1 δ , we have
R m , n E [ R m , n ] O ϵ * ( m + n ) ( d 1 ) / d log C m , n ( ϵ * ) / δ ( d 1 ) / d .
or, equivalently,
R m , n m + n E [ R m , n ] m + n O ϵ * ( m + n ) 1 / d log C m , n ( ϵ * ) / δ ( d 1 ) / d ,
where C m , n ( ϵ * ) depends on m , n , and d is given in (10) when ϵ = ϵ * .

3. Numerical Experiments

3.1. Simulation Study

In this section, we apply the FR statistic estimate of the HP-divergence to both simulated and real data sets. We present results of a simulation study that evaluates the proposed bound on the MSE. We numerically validate the theory stated in Section 2.2 and Section 2.4 using multiple simulations. In the first set of simulations, we consider two multivariate Normal random vectors X , Y and perform three experiments d = 2 , 4 , 8 , to analyze the FR test statistic-based estimator performance as the sample sizes m, n increase. For the three dimensions d = 2 , 4 , 8 , we generate samples from two normal distributions with identity covariance and shifted means: μ 1 = [ 0 , 0 ] , μ 2 = [ 1 , 0 ] and μ 1 = [ 0 , 0 , 0 , 0 ] , μ 2 = [ 1 , 0 , 0 , 0 ] and μ 1 = [ 0 , 0 , , 0 ] , μ 2 = [ 1 , 0 , , 0 ] when d = 2 , d = 4 and d = 8 , respectively. For all of the following experiments, the sample sizes for each class are equal ( m = n ).
We vary N = m = n up to 800. From Figure 3, we deduce that, when the sample size increases, the MSE decreases such that for higher dimensions the rate is slower. Furthermore, we compare the experiments with the theory in Figure 3. Our theory generally matches the experimental results. However, the MSE for the experiments tends to decrease to zero faster than the theoretical bound. Since the Gaussian distribution has a smooth density, this suggests that a tighter bound on the MSE may be possible by imposing stricter assumptions on the density smoothness as in [12].
In our next simulation, we compare three bivariate cases: first, we generate samples from a standard Normal distribution. Second, we consider a distinct smooth class of distributions i.e., binomial Gamma density with standard parameters and dependency coefficient ρ = 0.5 . Third, we generate samples from Standard t-student distributions. Our goal in this experiment is to compare the MSE of the HP-divergence estimator between two identical distributions, f 0 = f 1 , when f 0 is one of the Gamma, Normal, and t-student density function. In Figure 4, we observe that the MSE decreases as N increases for all three distributions.

3.2. Real Datasets

We now show the results of applying the FR test statistic to estimate the HP-divergence using three different real datasets [46]:
  • Human Activity Recognition (HAR), Wearable Computing, Classification of Body Postures and Movements (PUC-Rio): This dataset contains five classes (sitting-down, standing-up, standing, walking, and sitting) collected on eight hours of activities of four healthy subjects.
  • Skin Segmentation dataset (SKIN): The skin dataset is collected by randomly sampling B,G,R values from face images of various age groups (young, middle, and old), race groups (white, black, and asian), and genders obtained from the FERET and PAL databases [47].
  • Sensorless Drive Diagnosis (ENGIN) dataset: In this dataset, features are extracted from electric current drive signals. The drive has intact and defective components. The dataset contains 11 different classes with different conditions. Each condition has been measured several times under 12 different operating conditions, e.g., different speeds, load moments, and load forces.
We focus on two classes from each of the HAR, SKIN, and ENGIN datasets, specifically, for HAR dataset two classes “sitting” and “standing” and for SKIN dataset the classes “Skin” and “Non-skin” are considered. In the ENGIN dataset, the drive has intact and defective components, which results in 11 different classes with different conditions. We choose conditions 1 and 2.
In the first experiment, we computed the HP-divergence using KDE plug-in estimator and then the MSE for the FR test statistic estimator is derived as the sample size N = m = n increases. We used 95% confidence interval as the error bars. We observe in Figure 5 that the estimated HP-divergence ranges in [ 0 , 1 ] , which is one of the HP-divergence properties [8]. Interestingly, when N increases the HP-divergence tends to 1 for all HAR, SKIN, and ENGIN datasets. Note that in this set of experiments we have repeated the experiments on independent parts of the datasets to obtain the error bars. Figure 6 shows that the MSE expectedly decreases as the sample size grows for all three datasets. Here, we have used the KDE plug-in estimator [12], implemented on the all available samples, to determine the true HP-divergence. Furthermore, according to Figure 6, the FR test statistic-based estimator suggests that the Bayes error rate is larger for the SKIN dataset compared to the HAR and ENGIN datasets.
In our next experiment, we add the first six features (dimensions) in order to our datasets and evaluate the FR test statistic’s performance as the HP-divergence estimator. Surprisingly, the estimated HP-divergence doesn’t change for the HAR sample; however, big changes are observed for the SKIN and ENGIN samples (see Figure 7).
Finally, we apply the concentration bounds on the FR test statistic (i.e., Theorems 6 and 7) and compute theoretical implicit variance-like bound for the FR criteria with δ = 0.05 error for the real datasets ENGIN, HAR, and SKIN. Since datasets ENGIN, HAR, and SKIN have the equal total sample size N = m + n = 1200 and different dimensions d = 14 , 12 , 4 , respectively; here, we first intend to compare the concentration bound (13) on the FR statistic in terms of dimension d when δ = 0.05 . For real datasets ENGIN, HAR, and SKIN, we obtain
P | R m , n E [ R m , n ] | ξ 0.95 ,
where ξ = ξ [ 0.257 , 0.005 , 0.6 × 10 11 ] , respectively, and ξ is a constant not dependent on d. One observes that as the dimension decreases the interval becomes significantly tighter. However, this could not be generally correct and computing bound (13) precisely requires the knowledge of distributions and unknown constants. In Table 1, we compute the standard variance-like bound by applying the percentiles technique and observe that the bound threshold is not monotonic in terms of dimension d. Table 1 shows the FR test statistic, HP-divergence estimate (denoted by R m , n , D ^ p , respectively), and standard variance-like interval for the FR statistic using the three real datasets HAR, SKIN, and ENGIN.

4. Conclusions

We derived a bound on the MSE convergence rate for the Friedman–Rafsky estimator of the Henze–Penrose divergence assuming the densities are sufficiently smooth. We employed a partitioning strategy to derive the bias rate which depends on the number of partitions, the sample size m + n , the Hölder smoothness parameter η , and the dimension d. However, by using the optimal partition number, we derived the MSE convergence rate only in terms of m + n , η , and d. We validated our proposed MSE convergence rate using simulations and illustrated the approach for the meta-learning problem of estimating the HP-divergence for three real-world data sets. We also provided concentration bounds around the median and mean of the estimator. These bounds explicitly provide the rate that the FR statistic approaches its median/mean with high probability, not only as a function of the number of samples, m, n, but also in terms of the dimension of the space d. By using these results, we explored the asymptotic behavior of a variance-like rate in terms of m, n, and d.

Author Contributions

Conceptualization, S.Y.S., M.N. and A.O.H.; methodology, S.Y.S. and M.N.; software, S.Y.S. and M.N.; validation, S.Y.S., M.N., K.R.M. and A.O.H.; formal analysis, S.Y.S., M.N. and K.R.M.; investigation, S.Y.S. and M.N.; resources, S.Y.S. and M.N.; data curation, M.N.; writing—original draft preparation, S.Y.S.; writing—review and editing, M.N., K.R.M. and A.O.H.; supervision, A.O.H.; project administration, A.O.H.; funding acquisition, A.O.H.

Funding

The work presented in this paper was partially supported by ARO grant W911NF-15-1-0479 and DOE grants DE-NA0002534 and DE-NA0003921.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

Abbreviations

HPHenze-Penrose
BERBayes error rate
MSTMinimal Spanning Tree
FRFriedman-Rafsky
MSEMean squared error

Appendix A. Proof of Theorem 4

In this section, we prove the subadditivity and superadditivity for the mean of FR test statistic. For this, first we need to illustrate the following lemma.
Lemma A1.
Let { Q i } i = 1 l d be a uniform partition of [ 0 , 1 ] d into l d subcubes Q i with edges parallel to the coordinate axes having edge lengths l 1 and volumes l d . Let D i j be the set of edges of MST graph between Q i and Q j with cardinality | D i j | , then for | D | defined as the sum of | D i j | for all i , j = 1 , , l d , i j , we have E | D | = O ( l d 1 n 1 / d ) , or more explicitly
E [ | D | ] C l d 1 n 1 / d + O ( l d 1 n ( 1 / d ) s ) ,
where η > 0 is the Hölder smoothness parameter and
s = ( 1 1 / d ) η d ( ( 1 1 / d ) η + 1 ) .
Here, and in what follows, denote Ξ M S T ( X n ) the length of the shortest spanning tree on X n = { X 1 , , X n } , namely
Ξ M S T ( X n ) : = min T e T | e | ,
where the minimum is over all spanning trees T of the vertex set X n . Using the subadditivity relation for Ξ M S T in [17], with the uniform partition of [ 0 , 1 ] d into l d subcubes Q i with edges parallel to the coordinate axes having edge lengths l 1 and volumes l d , we have
Ξ M S T ( X n ) i = 1 l d Ξ M S T ( X n Q i ) + C l d 1 ,
where C is constant. Denote D the set of all edges of M S T i = 1 M Q i that intersect two different subcubes Q i and Q j with cardinality | D | . Let | e i | be the length of i-th edge in set D. We can write
i | D | | e i | C l d 1 and E i | D | | e i | C l d 1 ,
also we know that
E i | D | | e i | = E D i | D | E | e i | | D .
Note that using the result from ([31], Proposition 3), for some constants C i 1 and C i 2 , we have
E | e i | C i 1 n 1 / d + C i 2 n ( 1 / d ) s , i | D | .
Now, let C 1 = max i { C i 1 } and C 2 = max i { C i 2 } , hence we can bound the expectation (A3) as
E | D | ( C 1 n 1 / d + C 2 ( n ( 1 / d ) s ) ) C l d 1 ,
which implies
E | D | ( C 1 n 1 / d + O ( n ( 1 / d ) s ) ) C l d 1 n 1 / d + O ( l d 1 n ( 1 / d ) s ) .
To aim toward the goal (7), we partition [ 0 , 1 ] d into M : = l d subcubes Q i of side 1 / l . Recalling Lemma 2.1 in [48], we therefore have the set inclusion:
M S T i = 1 M Q i i = 1 M M S T ( Q i ) D ,
where D is defined as in Lemma A1. Let m i and n i be the number of sample { X 1 , , X m } and { Y 1 , , Y n } respectively falling into the partition Q i , such that i m i = m and i n i = n . Introduce sets A and B as
A : = M S T i = 1 M Q i , B : = i = 1 M M S T ( Q i ) .
Since set B has fewer edges than set A, thus (A5) implies that the difference set of B and A contains at most 2 | D | edges, where | D | is the number of edges in D. On the other word,
| A Δ B | | A B | + | B A | = | D | + | B A | = | D | + ( | B | | B A | | D | + ( | A | | B A | ) = 2 | D | .
The number of edge linked nodes from different samples in set A is bounded by the number of edge linked nodes from different samples in set B plus 2 | D | :
R m , n ( X m , Y n ) i = 1 M R m i , n i ( X m , Y n ) Q i + 2 | D | .
Here, R m i , n i stands with the number edge linked nodes from different samples in partition Q i , M. Next, we address the reader to Lemma A1, where it has been shown that there is a constant c such that E | D | c l d 1 ( m + n ) 1 / d . This concludes the claimed assertion (7). Now, to accomplish the proof, the lower bound term in (8) is obtained with similar methodology and the set inclusion:
i = 1 M M S T ( Q i ) M S T i = 1 M Q i D .
This completes the proof.

Appendix B. Proof of Theorem 2

As many of continuous subadditive functionals on [ 0 , 1 ] d , in the case of the FR statistic, there exists a dual superadditive functional R m , n * based on dual MST, MST * , proposed in Definition 2. Note that, in the MST* graph, the degrees of the corner points are bounded by c d , where it only depends on dimension d, and is the bound for degree of every node in MST graph. The following properties hold true for dual FR test statistic, R m , n * :
Lemma A2.
Given samples X m = { X 1 , , X m } and Y n = { Y 1 , , Y n } , the following inequalities hold true:
(i)
For constant c d which depends on d:
R m , n * ( X m , Y n ) R m , n ( X m , Y n ) + c d 2 d , R m , n ( X m , Y n ) R m , n * ( X m , Y n ) .
(ii)
(Subadditivity on E [ R m , n * ] and Superadditivity) Partition [ 0 , 1 ] d into l d subcubes Q i such that m i , n i be the number of sample X m = { X 1 , , X m } and Y n = { Y 1 , , Y n } respectively falling into the partition Q i with dual R m i , n i * . Then, we have
E R m , n * ( X m , Y n ) i = 1 l d E R m i , n i * ( ( X m , Y n ) Q i ) + c l d 1 ( m + n ) 1 / d , R m , n * ( X m , Y n ) i = 1 l d R m i , n i * ( ( X m , Y n ) Q i ) 2 d c d l d ,
where c is a constant.
(i) Consider the nodes connected to the corner points. Since MST ( X m , Y n ) and MST * ( X m , Y n ) can only be different in the edges connected to these nodes, and in R * ( X m , Y n ) we take all of the edges between these nodes and corner nodes into account, so we obviously have the second relation in (A8). In addition, for the first inequality in (A8), it is enough to say that the total number of edges connected to the corner nodes is upper bounded by 2 d c d .
(ii) Let | D * | be the set of edges of the MST * graph which intersect two different partitions. Since MST and MST * are only different in edges of points connected to the corners and edges crossing different partitions. Therefore, | D * | | D | . By eliminating one edge in set D in the worse scenario we would face two possibilities: either the corresponding node is connected to the corner which is counted anyways or any other point in MST graph which wouldn’t change the FR test statistic. This implies the following subadditivity relation:
R m , n * ( X m , Y n ) | D | i = 1 l d R m i , n i * ( X m , Y n ) Q i .
Further from Lemma A1, we know that there is a constant c such that E | D | c l d 1 ( m + n ) 1 / d . Hence, the first inequality in (A9) is obtained. Next, consider | D c * | which represents the total number of edges from both samples only connected to the all corners points in MST * graph. Therefore, one can easily claim:
R m , n * ( X m , Y n ) i = 1 l d R m i , n i * ( X m , Y n ) Q i | D c * | .
In addition, we know that | D c * | 2 d l d c d where c d stands with the largest possible degree of any vertex. One can write
R m , n * ( X m , Y n ) i = 1 l d R m i , n i * ( X m , Y n ) Q i 2 d c d l d .
The following list of Lemmas A3, A4 and A6 are inspired from [49] and are required to prove Theorem A1. See Appendix E for their proofs.
Lemma A3.
Let g ( x ) be a density function with support [ 0 , 1 ] d and belong to the Hölder class Σ d ( η , L ) , 0 < η 1 , stated in Definition 1. In addition, assume that P ( x ) is a η-Hölder smooth function, such that its absolute value is bounded from above by a constant. Define the quantized density function with parameter l and constants ϕ i as
g ^ ( x ) = i = 1 M ϕ i 1 { x Q i } , w h e r e ϕ i = l d Q i g ( x ) d x .
Let M = l d and Q i = { x , x i : x x i < l d } . Then,
g ( x ) g ^ ( x ) P ( x ) d x O ( l d η ) .
Lemma A4.
Denote Δ ( x , S ) the degree of vertex x S in the M S T over set S with the n number of vertices. For given function P ( x , x ) , one obtains
P ( x , x ) g ( x ) E [ Δ ( x , S ) ] d x = 2 P ( x , x ) g ( x ) d x + ς η ( l , n ) ,
where, for constant η > 0 ,
ς η ( l , n ) = O l / n 2 l d / n g ( x ) P ( x , x ) d x + O ( l d η ) .
Lemma A5.
Assume that, for given k, g k ( x ) is a bounded function belong to Σ d ( η , L ) . Let P : R d × R d [ 0 , 1 ] be a symmetric, smooth, jointly measurable function, such that, given k, for almost every x R d , P ( x , . ) is measurable with x a Lebesgue point of the function g k ( . ) P ( x , . ) . Assume that the first derivative P is bounded. For each k, let Z 1 k , Z 2 k , , Z k k be an independent d-dimensional variable with common density function g k . Set Z k = { Z 1 k , Z 2 k , Z k k } and Z k x = { x , Z 2 k , Z 3 k , Z k k } . Then,
E j = 2 k P ( x , Z j k ) 1 ( x , Z j k ) M S T ( Z k x ) = P ( x , x ) E Δ ( x , Z k x ) + O k η / d + O k 1 / d .
Lemma A6.
Consider the notations and assumptions in Lemma A5. Then,
| k 1 1 i < j k P ( Z i k , Z j k ) 1 { ( Z i k , Z j k ) M S T ( Z k ) } R d P ( x , x ) g k ( x ) d x | ς η ( l , k ) + O ( k η / d ) + O ( k 1 / d ) .
Here, M S T ( S ) denotes the MST graph over nice and finite set S R d and η is the smoothness Hölder parameter. Note that ς η ( l , k ) is given as before in Lemma A4 (A13).
Theorem A1.
Assume R m , n : = R ( X m , Y n ) denotes the FR test statistic and densities f 0 and f 1 belong to the Hölder class Σ d ( η , L ) , 0 < η 1 . Then, the rate for the bias of the R m , n estimator for d 2 is of the form:
E R m , n m + n 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x O l d ( m + n ) η / d + O ( l d η ) .
The proof and a more explicit form for the bound (A16) are given in Appendix E.
Now, we are at the position to prove the assertion in (5). Without loss of generality, assume that ( m + n ) l d > 1 . In the range d 2 and 0 < η 1 , we select l as a function of m + n to be the sequence increasing in m + n which minimizes the maximum of these rates:
l ( m + n ) = a r g min l max l d ( m + n ) η / d , l η d .
The solution l = l ( m + n ) occurs when l d ( m + n ) η / d = l η d , or equivalently l = ( m + n ) η / ( d 2 ( η + 1 ) ) . Substitute this into l in the bound (A16), the RHS expression in (5) for d 2 is established.

Appendix C. Proof of Theorems 3

To bound the variance, we will apply one of the first concentration inequalities which was proved by Efron and Stein [43] and further was improved by Steele [18].
Lemma A7
(The Efron–Stein Inequality). Let X m = { X 1 , , X m } be a random vector on the space S . Let X = { X 1 , , X m } be the copy of random vector X m . Then, if f : S × × S R , we have
V f ( X m ) 1 2 i = 1 m E f ( X 1 , , X m ) f ( X 1 , , X i , , X m ) 2 .
Consider two set of nodes X i , 1 i m and Y j for 1 j n . Without loss of generality, assume that m < n . Then, consider the n m virtual random points X m + 1 , , X n with the same distribution as X i , and define Z i : = ( X i , Y i ) . Now, for using the Efron–Stein inequality on set Z n = { Z 1 , , Z n } , we involve another independent copy of Z n as Z n = { Z 1 , , Z n } , and define Z n ( i ) : = ( Z 1 , , Z i 1 , Z i , Z i + 1 , , Z n ) , then Z n ( 1 ) becomes ( Z 1 , Z 2 , , Z n ) = ( X 1 , Y 1 ) , ( X 2 , Y 2 ) , , ( X m , Y n ) = : ( X m ( 1 ) , Y n ( 1 ) ) where ( X 1 , Y 1 ) is independent copy of ( X 1 , Y 1 ) . Next, define the function r m , n ( Z n ) : = R m , n / ( m + n ) , which means that we discard the random samples X m + 1 , , X n , and find the previously defined R m , n function on the nodes X i , 1 i m and Y j for 1 j n , and multiply by some coefficient to normalize it. Then, according to the Efron–Stein inequality, we have
V a r ( r m , n ( Z n ) ) 1 2 i = 1 n E ( r m , n ( Z n ) r m , n ( Z n ( i ) ) ) 2 .
Now, we can divide the RHS as
1 2 i = 1 n E ( r m , n ( Z n ) r m , n ( Z n ( i ) ) ) 2 = 1 2 i = 1 m E ( r m , n ( Z n ) r m , n ( Z n ( i ) ) ) 2 + 1 2 i = m + 1 n E ( r m , n ( Z n ) r m , n ( Z n ( i ) ) ) 2 .
The first summand becomes
= 1 2 i = 1 m E ( r m , n ( Z n ) r m , n ( Z n ( i ) ) ) 2 = m 2 ( m + n ) 2 E ( R m , n ( X m , Y n ) R m , n ( X m ( 1 ) , Y n ( 1 ) ) ) 2 ,
which can also be upper bounded as follows:
R m , n ( X m , Y n ) R m , n ( X m ( 1 ) , Y n ( 1 ) ) R m , n ( X m , Y n ) R m , n ( X m ( 1 ) , Y n ) + R ( X m ( 1 ) , Y n ) R m , n ( X m ( 1 ) , Y n ( 1 ) ) .
For deriving an upper bound on the second line in (A19), we should observe how much changing a point’s position modifies the amount of R m , n ( X m , Y n ) . We consider two steps of changing X 1 ’s position: we first remove it from the graph, and then add it to the new position. Removing it would change R m , n ( X m , Y n ) at most by 2 c d because X 1 has a degree of at most c d , and c d edges will be removed from the MST graph, and c d edges will be added to it. Similarly, adding X 1 to the new position will affect R m , n ( X m , n , Y m , n ) at most by 2 c d . Thus, we have
R m , n ( X m , Y n ) R m , n ( X m ( 1 ) , Y n ) 4 c d ,
and we can also similarly reason that
R m , n ( X m ( 1 ) , Y n ) R m , n ( X m ( 1 ) , Y n ( 1 ) ) 4 c d .
Therefore, totally we would have
R m , n ( X m , Y n ) R m , n ( X m ( 1 ) , Y n ( 1 ) ) 8 c d .
Furthermore, the second summand in (A18) becomes
= 1 2 i = m + 1 n E ( r m , n ( Z n ) r m , n ( Z n ( i ) ) ) 2 = K m , n E ( R m , n ( X m , Y n ) R m , n ( X m ( m + 1 ) , Y n ( m + 1 ) ) ) 2 ,
where K m , n = n m 2 ( m + n ) 2 . Since, in ( X m ( m + 1 ) , Y n ( m + 1 ) ) , the point X m + 1 is a copy of virtual random point X m + 1 , therefore this point doesn’t change the FR test statistic R m , n . In addition, following the above arguments, we have
R m , n ( X m , Y n ) R m , n ( X m , Y n ( m + 1 ) ) 4 c d .
Hence, we can bound the variance as below:
V a r ( r m , n ( Z n ) ) 8 c d 2 ( n m ) ( m + n ) 2 + 32 c d 2 m ( m + n ) 2 .
Combining all results with the fact that n m + n q concludes the proof.

Appendix D. Proof of Theorems 5–7

We will need the following prominent results for the proofs.
Lemma A8.
For h = 1 , 2 , , let δ m , n h be the function c h d 1 ( m + n ) 1 / d , where c is a constant. Then, for ϵ > 0 , we have
P R m , n ( X m , Y n ) i = 1 h d R m i , n i ( X m i , Y n i ) + 2 ϵ ϵ δ m , n h ϵ .
Note that, in the case ϵ δ m , n h , the above claimed inequality becomes trivial.
The subadditivity property for FR test statistic R m , n in Lemma A8, as well as Euclidean functionals, leads to several non-trivial consequences. The growth bound was first explored by Rhee (1993b) [50], and as is illustrated in [17,27] has a wide range of applications. In this paper, we investigate the probabilistic growth bound for R m , n . This observation will lead us to our main goal in this appendix that is providing the proof of Theorem 6. For what follows, we will use δ m , n h notation for the expression O h d 1 ( m + n ) 1 / d .
Lemma A9
(Growth bounds for R m , n ). Let R m , n be the FR test statistic. Then, for given non-negative ϵ, such that ϵ h 2 δ m , n h , with at least probability g ( ϵ ) : = 1 h δ m , n h ϵ , h = 2 , 3 , , we have
R m , n ( X m , Y n ) c ϵ , h # X m # Y n 1 1 / d .
Here, c ϵ , h = O ϵ h d 1 1 depending only on ϵ and h.
The complexity of R m , n ’s behavior and the need to pursue the proof encouraged us to explore the smoothness condition for R m , n . In fact, this is where both subadditivity and superadditivity for the FR statistic are used together and become more important.
Lemma A10
(Smoothness for R m , n ). Given observations of
X m : = ( X m , X m ) = { X 1 , , X m , X m + 1 , , X m } ,
where m + m = m and Y n : = ( Y n , Y n ) = { Y 1 , , Y n , Y n + 1 , , Y n } , where n + n = n , denote R m , n ( X m , Y n ) as before, the number of edges of MST ( X m , Y n ) which connect a point of X m to a point of Y n . Then, for given integer h 2 , for all ( X n , Y m ) [ 0 , 1 ] d , ϵ h 2 δ m , n h where δ m , n h = O h d 1 ( m + n ) 1 / d , we have
P | R m , n ( X m , Y n ) R m , n ( X m , Y n ) | c ˜ ϵ , h # X m # Y n 1 1 / d 1 2 h δ m , n h ϵ ,
where c ˜ ϵ , h = O ϵ h d 1 1 .
Remark: Using Lemma A10, we can imply the continuty property, i.e., for all observations ( X m , Y n ) and ( X m , Y n ) , with at least probability 2 g ( ϵ ) 1 , one obtains
| R m , n ( X m , Y n ) R m , n ( X m , Y n ) | c ϵ , h * # ( X m Δ X m ) # ( Y n Δ Y n ) 1 1 / d ,
for given ϵ > 0 , c ϵ , h * = O ϵ h d 1 1 , h 2 . Here, X m Δ X m denotes symmetric difference of observations X m and X m .
The path to approach the assertions (11) and (12) proceeds via semi-isoperimentic inequality for the R m , n involving the Hamming distance.
Lemma A11
(Semi-Isoperimetry). Let μ be a measure on [ 0 , 1 ] d ; μ n denotes the product measure on space ( [ 0 , 1 ] d ) n . In addition, let M e denotes a median of R m , n . Set
A : = X m [ 0 , 1 ] d m , Y n [ 0 , 1 ] d n ; R m , n ( X m , Y n ) M e .
Following the notations in [17], H ( x , x ) = # { i , x i x i ) and ϕ A ( x ) + ϕ A ( y ) = min { H ( x , x ) + H ( y , y ) : x , y A } and ϕ A ( x ) ϕ A ( y ) = min { H ( x , x ) H ( y , y ) : x , y A } . Then,
μ m + n x ( [ 0 , 1 ] d ) m , y ( [ 0 , 1 ] d ) n : ϕ A ( x ) ϕ A ( y ) t 4 exp t 8 ( m + n ) .
Now, we continue by providing the proof of Theorem 5. Recall (A25) and denote
F x : = x i , i = 1 , , m , x i = x i , F y : = y j , j = 1 , , n , y j = y j , and G x : = x i , i = 1 , , m , x i x i , G y : = y j , j = 1 , , n , y j y j .
In addition, for given integer h, define events B , B by
B : = | R m , n ( X m , Y n ) R ( F x , F y ) | c ϵ , h # G x # G y 1 1 / d , B : = | R ( F x , F y ) R m , n ( X m , Y n ) | c ϵ , h # G x # G y 1 1 / d ,
where c ϵ , h is a constant. By virtue of smoothness property, Lemma A10, for ϵ h 2 δ m , n h , we know P ( B ) 2 g ( ϵ ) 1 and P ( B ) 2 g ( ϵ ) 1 . On the other hand, we have
R m , n ( X m , Y n ) | R m , n ( X m , Y n ) R ( F x , F y ) | + | R ( F x , F y ) R m , n ( X m , Y n ) | + R m , n ( X m , Y n ) . = | ϖ | + | ϖ | + R m , n ( X m , Y n ) ( say ) .
Moreover, P ( R m , n ( X m , Y n ) M e ) 1 / 2 . Therefore, we can write
1 / 2 P R m , n ( X m , Y n ) M e + | ϖ | + | ϖ | P R m , n ( X m , Y n ) M e + | ϖ | + | ϖ | | B B P ( B B ) + P ( B c B c ) .
Thus, we obtain
P R m , n ( X m , Y n ) M e + 4 ϵ # G x # G y 1 1 / d 1 / 2 1 + P ( B B ) / P ( B B ) = 1 2 P ( B B ) 1 .
Note that P ( B B ) = P ( B ) P ( B ) 2 g ( ϵ ) 1 2 . Now, we easily claim that
1 2 P ( B B ) 1 1 2 ( 2 g ( ϵ ) 1 ) 2 1 .
Thus,
P R m , n ( X m , Y n ) M e + 4 ϵ # G x # G y 1 1 / d 1 2 ( 2 g ( ϵ ) 1 ) 2 1 .
On the other word, calling ϕ A ( x ) and ϕ A ( y ) in Lemma A11, we get
P R m , n ( X m , Y n ) M e + 4 ϵ ϕ A ( x ) ϕ A ( y ) 1 1 / d 1 2 ( 2 g ( ϵ ) 1 ) 2 1 .
Furthermore, denote event
C : = R m , n ( X m , Y n ) M e + 4 ϵ ϕ A ( x ) ϕ A ( y ) 1 1 / d .
Then, we have
P R m , n ( X m , Y n ) M e + t = μ m + n R m , n ( X m , Y n ) M e + t = μ m + n ( R m , n ( X m , Y n ) M e + t | C ) P ( C ) + μ m + n ( R m , n ( X m , Y n ) M e + t | C c ) P ( C c ) μ m + n ϕ A ( x ) ϕ A ( y ) 1 1 / d t 4 ϵ P ( C ) + μ m + n ( R m , n ( X m , Y n ) M e + t | C c ) P ( C c ) . Using P ( C ) = 1 P ( C c ) = μ m + n ϕ A ( x ) ϕ A ( y ) 1 1 / d t 4 ϵ + P ( C c ) { μ m + n ( R m , n ( X m , Y n ) M e + t | C c ) μ m + n ϕ A ( x ) ϕ A ( y ) 1 1 / d t 4 ϵ } .
Define set K t = ϕ A ( x ) ϕ A ( y ) 1 1 / d t 4 ϵ , so
μ m + n R m , n ( X m , Y n ) M e + t | C c = μ m + n R m , n ( X m , Y n ) M e + t | C c , K t μ m + n ( K t ) + μ m + n ( R m , n ( X m , Y n ) M e + t | C c , K t c ) μ m + n ( K t c ) .
Since
μ m + n R m , n ( X m , Y n ) M e + t | C c , K t = 1 ,
and
μ m + n R m , n ( X m , Y n ) M e + t | C c , K t c = μ m + n R m , n ( X m , Y n ) M e + t .
Consequently, from (A30), one can write
P R m , n ( X m , Y n ) M e + t μ m + n ϕ A ( x ) ϕ A ( y ) 1 1 / d t 4 ϵ + P ( C c ) μ m + n R m , n ( X m , Y n ) M e + t μ m + n ( K t c ) μ m + n ϕ A ( x ) ϕ A ( y ) 1 1 / d t 4 ϵ + 2 ( 2 g ( ϵ ) 1 ) 2 1 P R m , n ( X m , Y n ) M e + t .
The last inequality implies by owing to (A29) and μ m + n ( K t c ) 1 . For g ( ϵ ) 1 / 2 + 1 / 2 2 , we have
1 2 ( 2 g ( ϵ ) 1 ) 2 1 0 ,
or equivalently this holds true when ϵ ( 2 h 2 δ m , n h ) / ( 2 1 ) . Furthermore, for h 7 , we have
h 2 δ m , n h ( 2 h 2 δ m , n h ) / ( 2 1 ) ,
therefore P R m , n ( X m , Y n ) M e + t is less than and equal to
1 2 ( 2 g ( ϵ ) 1 ) 2 1 1 μ m + n ϕ A ( x ) ϕ A ( y ) 1 1 / d t 4 ϵ .
By virtue of Lemma A11, finally we obtain
P R m , n ( X m , Y n ) M e + t 4 1 2 ( 2 g ( ϵ ) 1 ) 2 1 1 exp t d / ( d 1 ) 8 ( 4 ϵ ) d / d 1 ( m + n ) .
Similarly, we can derive the same bound on P R m , n ( X m , Y n ) M e t , so we obtain
P | R m , n M e | t C m , n ( ϵ , h ) exp t d / ( d 1 ) 8 ( 4 ϵ ) d / ( d 1 ) ( m + n ) ,
where
C m , n ( ϵ , h ) = 8 1 2 1 1 2 h O h d 1 ( m + n ) 1 / d ϵ 2 1 .
We will analyze (A35) together with Theorem 6. The next lemma will be employed in Theorem 6’s proof.
Lemma A12
(Deviation of the Mean and Median). Consider M e as a median of R m , n . Then, for ϵ h 2 δ m , n h and given h 7 , we have
| E R m , n ( X m , Y n ) M e | C m , n ( ϵ , h ) ( m + n ) ( d 1 ) / d ,
where C m , n ( ϵ , h ) is a constant depending on ϵ, h, m, and n by
C m , n ( ϵ , h ) = C 1 2 ( 2 g ( ϵ ) 1 ) 2 1 1 ,
where C is a constant and
δ m , n h = O h d 1 ( m + n ) 1 / d , and g ( ϵ ) = 1 h δ m , n h ϵ .
We conclude this part by pursuing our primary intension which has been the Theorem 6’s proof. Observe from Theorem 5, (11) that
P ( | R m , n E [ R m , n ] | t + C m , n ( ϵ , l ) m + n ) ( d 1 ) / d P ( | R m , n M e | + | E [ R m , n ] M e | t + C m , n ( ϵ , l ) m + n ) ( d 1 ) / d P | R m , n M e | t 8 1 2 ( 2 g ( ϵ ) 1 ) 2 1 1 exp t d / ( d 1 ) 8 ( 4 ϵ ) d / d 1 ( m + n ) .
Note that the last bound is derived by (11). The rest of the proof is as the following: When t 2 C m , n ( ϵ , h ) ( m + n ) ( d 1 ) / d , we use
( t C m , n ( ϵ , h ) m + n ) ( d 1 ) / d d / ( d 1 ) t / 2 d / ( d 1 ) .
Therefore, it turns out that
P | R m , n E [ R m , n ] | t 8 1 2 ( 2 g ( ϵ ) 1 ) 2 1 1 exp t d / ( d 1 ) 8 ( 8 ϵ ) d / ( d 1 ) ( m + n ) .
In other words, there exist constants C m , n ( ϵ , h ) depending on m , n , ϵ , and h such that
P | R m , n E [ R m , n ] | t C m , n ( ϵ , h ) exp ( t / ( 2 ϵ ) ) d / ( d 1 ) ( m + n ) C ˜ ,
where C ˜ = 8 ( 4 ) d / ( d 1 ) .
To verify the behavior of bound (A40) in terms of ϵ , observe (A35) first; it is not hard to see that this function is decreasing in ϵ . However, the function
exp ( t / ( 2 ϵ ) ) d / ( d 1 ) ( m + n ) C ˜
increases in ϵ . Therefore, one can not immediately infer that the bound in (12) is monotonic with respect to ϵ . For fixed N = n + m , d, and h, the first and second derivatives of the bound (12) with respect to ϵ are quite complicated functions. Thus, deriving an explicit optimal solution for the minimization problem with the objective function (12) is not feasible. However, in the sequel, we discuss that under conditions when t is not much larger than N = m + n , this bound becomes convex with respect to ϵ . Set
K ( ϵ ) = C m , n ( ϵ , h ) exp B ( t ) ϵ d / ( d 1 ) ,
where C m , n is given in (10) and
B ( t ) = t d / ( d 1 ) 8 ( 8 ) d / ( d 1 ) ( N ) .
By taking the derivative with respect to ϵ , we have
d K ( ϵ ) d ϵ = K ( ϵ ) d d ϵ log C m , n + B ( t ) d / ( d 1 ) ϵ ( 2 d 1 ) / ( d 1 ) ,
where
d d ϵ log C m , n = 4 a h ϵ ( ϵ 2 a h ) ( 8 a h 2 8 ϵ a h + ϵ 2 ) ,
where a h = h δ m , n h . The second derivative K ( ϵ ) with respect to ϵ after simplification is given as
d 2 d ϵ 2 K ( ϵ ) = 4 a h ϵ ( ϵ 2 a h ) ( 8 a h 2 8 ϵ a h + ϵ 2 ) + B ( t ) d ¯ ϵ d ¯ + 1 2 + K ( ϵ ) 8 a h ( 8 a h 3 + ϵ 2 ( ϵ 5 a h ) ) ( 8 a h 2 8 a h ϵ + ϵ 2 ) 2 ( ϵ 2 a h ) 2 B ( t ) d ¯ ( d ¯ + 1 ) ϵ d ¯ + 2 ,
where d ¯ = d / ( d 1 ) . The first term in (A44) and K ( ϵ ) are non-negative, so K ( ϵ ) is convex if the second term in the second line of (A44) is non-negative. We know that ϵ h 2 δ m , n h = h a h , when h = 7 , we can parameterize ϵ by setting it equal to γ a h , where γ 7 . After simplification, K ( ϵ ) is convex if
a h d ¯ 1 γ d ¯ 1 + 3 γ d ¯ 2 + B ( t ) d ¯ ( d ¯ + 1 ) × { a h 1 32 γ 6 + 64 γ 5 48 γ 4 + 8 γ 3 7 2 γ 2 + 2 γ 1 1 8 + a h 2 32 γ 6 64 γ 5 + 40 γ 4 + 8 γ 3 + 1 2 γ 2 } 0 .
This is implied if
0 B ( t ) d ¯ ( d ¯ + 1 ) a h 1 × 32 γ 6 + 64 γ 5 48 γ 4 + 8 γ 3 7 2 γ 2 + 2 γ 1 1 8 ,
such that γ 7 . One can easily check that, as γ , then (A46) tends to 1 8 B ( t ) d ¯ ( d ¯ + 1 ) a h 1 . This term can be negligible unless we have t that is much larger than N = m + n with the threshold depending on d. Here, by setting B ( t ) / a h = 1 , a rough threshold t = O 7 d 1 ( m + n ) 1 1 / d 2 depending on d, m + n is proposed. Therefore, minimizing (A35) and (A40) with respect to ϵ when optimal h = 7 is a convex optimization problem. Denote ϵ * the solution of the convex optimization problem (9). By plugging optimal h ( h = 7 ) and ϵ ( ϵ = ϵ * ) in (A35) and (A40), we derive (11) and (12), respectively.
In this appendix, we also analyze the bound numerically. By simulation, we observed that lower h i.e., h = 7 is the optimal value experimentally. Indeed, this can be verified by Theorem 11’s proof. We address the reader to Lemma A8 in Appendix D and Appendix E where, as h increases, the lower bound for the probability increases too. In other words, for fixed N = m + n and d, the lowest h implies the maximum bound in (A92). For this, we set h = 7 in our experiments. We vary the dimension d and sample size N = m + n in relatively large and small ranges. In Table A1, we solve (9) for various values of d and N = m + n . We also compute the lower bound for ϵ i.e., 7 d + 1 N 1 / d per experiment. In Table A1, we observe that as we have higher dimension the optimal value ϵ * equals the ϵ lower bound h d + 1 N 1 / d , but this is not true for smaller dimensions with even relatively large sample size.
Table A1. d, N, ϵ * are dimension, total sample size m + n , and optimal ϵ for the bound in (12). The column h d + 1 N 1 / d represents approximately the lower bound for ϵ which is our constraint in the minimization problem and our assumption in Theorems 5 and 6. Here, we set h = 7 .
Table A1. d, N, ϵ * are dimension, total sample size m + n , and optimal ϵ for the bound in (12). The column h d + 1 N 1 / d represents approximately the lower bound for ϵ which is our constraint in the minimization problem and our assumption in Theorems 5 and 6. Here, we set h = 7 .
Concentration Bound (11)
d N = m + n ϵ * t 0 h d + 1 N 1 / d Optimal (11)
2 10 3 1.1424 × 10 4 2 × 10 7 1.0847 × 10 4 0.3439
4 10 4 1.7746 × 10 5 3 × 10 10 168,0700.0895
5550 4.7236 × 10 5 10 10 4.1559 × 10 5 0.9929
6 10 4 3.8727 × 10 6 2 × 10 12 3.8225 × 10 6 0.1637
81200 9.7899 × 10 7 12 × 10 12 9.7899 × 10 7 0.7176
103500 4.4718 × 10 9 2 × 10 15 4.4718 × 10 9 0.4795
15 10 8 1.1348 × 10 14 10 24 1.1348 × 10 14 0.9042
To validate our proposed bound in (12), we again set h = 7 and for d = 4 , 5 , 7 we ran experiments with sample sizes N = m + n = 9000 , 1100 , 140 , respectively. Then, we solved the minimization problem to derive optimal bound for t in the range 10 10 [ 1 , 3 ] . Note that we chose this range to have a non-trivial bound for all three curves; otherwise, the bounds partly become one. Figure A1 shows that when t increases in the given range, the optimal curves approach zero.
Figure A1. Optimal bound for (12), when h = 7 versus t 10 10 [ 1 , 3 ] . The bound decreases as t grows.
Figure A1. Optimal bound for (12), when h = 7 versus t 10 10 [ 1 , 3 ] . The bound decreases as t grows.
Entropy 21 01144 g0a1
To prove the Theorem 7 in the concentration of R m , n , Theorem 6, let
δ = C m , n ( ϵ * ) exp ( t / ( 2 ϵ * ) ) d / ( d 1 ) ( m + n ) C ˜ ,
this implies t = O ( ϵ * ( m + n ) ( d 1 ) / d log C m , n ( ϵ * ) / δ ) ( d 1 ) / d and the proofs are completed.

Appendix E. Additional Proofs

Lemma A3: Let g ( x ) be a density function with support [ 0 , 1 ] d and belong to the Hölder class Σ d ( η , L ) , 0 < η 1 , expressed in Definition 1. In addition, assume that P ( x ) is a η -Hölder smooth function, such that its absolute value is bounded from above by some constants c. Define the quantized density function with parameter l and constants ϕ i as
g ^ ( x ) = i = 1 M ϕ i 1 { x Q i } , where ϕ i = l d Q i g ( x ) d x ,
and M = l d and Q i = { x , x i : x x i < l d } . Then,
g ( x ) g ^ ( x ) P ( x ) d x O ( l d η ) .
Proof. 
By the mean value theorem, there exist points ϵ i Q i such that
ϕ i = l d Q i g ( x ) d x = g ( ϵ i ) .
Using the fact that g Σ d ( η , L ) and P ( x ) is a bounded function, we have
g ( x ) g ^ ( x ) ) P ( x ) d x = i = 1 M Q i ( g ( x ) Φ i ) P ( x ) d x = i = 1 M Q i ( g ( x ) g ( ϵ i ) ) P ( x ) d x c L i = 1 M Q i x ϵ i η d x .
Here, L is the Hölder constant. As x , ϵ i Q i , a sub-cube with edge length l 1 , then x ϵ i η = O ( l d η ) and i = 1 M Q i d x = 1 . This concludes the proof. □
Lemma A4: Let Δ ( x , S ) denote the degree of vertex x S in the M S T over set S R d with the n number of vertices. For given function P ( x , x ) , one yields
P ( x , x ) g ( x ) E [ Δ ( x , S ) ] d x = 2 P ( x , x ) g ( x ) d x + ς η ( l , n ) ,
where for constant η > 0 ,
ς η ( l , n ) = O l / n 2 l d / n g ( x ) P ( x , x ) d x + O ( l d η ) .
Proof. 
Recall notations in Lemma A3 and
| g ( x ) P ( x ) d x g ^ ( x ) P ( x ) d x | | g ( x ) g ^ ( x ) P ( x ) | d x .
Therefore, by substituting g ^ , defined in (A47), into g with considering its error, we have
P ( x , x ) g ( x ) E [ Δ ( x , S ) ] d x = P ( x , x ) E [ Δ ( x , S ) ] i = 1 M ϕ i 1 { x Q i } d x + O ( l d η ) = i = 1 M ϕ i Q i P ( x , x ) E [ Δ ( x , S ) ] d x + O ( l d η ) .
Here, Q i represents as before in Lemma A3, so the RHS of (A51) becomes
i = 1 M ϕ i Q i P ( x , x ) E [ Δ ( x , S Q i ) ] d x + i = 1 M ϕ i Q i P ( x , x ) O ( l 1 d / n ) + O ( l d η ) = i = 1 M ϕ i P ( x i , x i ) 1 M Q i M E [ Δ ( x , S Q i ) ] d x + i = 1 M ϕ i Q i P ( x , x ) O ( l 1 d / n ) + 2 O ( l d η ) .
Now, note that Q i M E [ Δ ( x , S Q i ) ] d x is the expectation of E [ Δ ( x , S Q i ) ] over the nodes in Q i , which is equal to 2 2 k i , where k i = n M . Consequently, we have
P ( x , x ) g ( x ) E [ Δ ( x , S ) ] d x = 2 2 M n i = 1 M ϕ i P ( x i , x i ) 1 M + O l 1 d n i = 1 M ϕ i P ( x i , x i ) + 3 O ( l d η ) = 2 g ( x ) P ( x , x ) d x + 5 O ( l d η ) ) + M O l 1 d n 2 n g ( x ) P ( x , x ) d x .
This gives the assertion (A49). □
Lemma A5: Assume that, for given k, g k ( x ) is a bounded function belong to Σ d ( η , L ) . Let P : R d × R d [ 0 , 1 ] be a symmetric, smooth, jointly measurable function, such that, given k, for almost every x R d , P ( x , . ) is measurable with x a Lebesgue point of the function g k ( . ) P ( x , . ) . Assume that the first derivative P is bounded. For each k, let Z 1 k , Z 2 k , , Z k k be independent d-dimensional variable with common density function g k . Set Z k = { Z 1 k , Z 2 k , Z k k } and Z k x = { x , Z 2 k , Z 3 k , Z k k } . Then,
E j = 2 k P ( x , Z j k ) 1 ( x , Z j k ) M S T ( Z k x ) = P ( x , x ) E Δ ( x , Z k x ) + O k η / d + O k 1 / d .
Proof. 
Let B ( x , r ) = { y : y x d r } . For any positive K, we can obtain:
E j = 2 k | P ( x , Z j k ) P ( x , x ) | 1 Z j k B x , K k 1 / d = ( k 1 ) B x ; K k 1 / d | P ( x , y ) g k ( y ) P ( x , x ) g k ( x ) + P ( x , x ) g k ( x ) g k ( y ) | d y ( k 1 ) [ B x ; K k 1 / d | P ( x , y ) g k ( y ) P ( x , x ) g k ( x ) | d y + O k η / d V B x , K k 1 / d ,
where V is the volume of space B which equals O ( k 1 ) . Note that the above inequality appears because g k ( x ) Σ d ( η , L ) and P ( x , x ) [ 0 , 1 ] . The first order Taylor series expansion of P ( x , y ) around x is
P ( x , y ) = P ( x , x ) + P ( 1 ) ( x , x ) y x + o y x 2 = P ( x , x ) + O k 1 / d + o k 2 / d .
Then, by recalling the Hölder class, we have
| P ( x , y ) g k ( y ) P ( x , x ) g k ( x ) | = | P ( x , x ) + O ( k 1 / d ) g k ( x ) + O ( k η / d ) P ( x , x ) g k ( x ) | = O ( k η / d ) + O ( k 1 / d ) .
Hence, the RHS of (A55) becomes
( k 1 ) O ( k η / d ) + O ( k 1 / d ) V B x , K k 1 / d + O k η / d V B x , K k 1 / d = ( k 1 ) O k 1 η / d + O k 1 1 / d .
The expression in (A54) can be obtained by choice of K. □
Lemma A6: Consider the notations and assumptions in Lemma A5. Then,
| k 1 1 i < j k P ( Z i k , Z j k ) 1 { ( Z i k , Z j k ) M S T ( Z k ) R d P ( x , x ) g k ( x ) d x | ς η ( l , k ) + O ( k η / d ) + O ( k 1 / d ) .
Here, M S T ( S ) denotes the MST graph over nice and finite set S R d and η is the smoothness Hölder parameter. Note that ς η ( l , k ) is given as before in (A50).
Proof. 
Following notations in [49], let Δ ( x , S ) denote the degree of vertex x in the M S T ( S ) graph. Moreover, let x be a Lebesgue point of g k with g k ( x ) > 0 . In addition, let Z k x be the point process { x , Z 2 k , Z 3 k , , Z k k } . Now, by virtue of (A55) in Lemma A5, we can write
E j = 2 k P ( x , Z j k ) 1 { ( x , Z j k ) M S T ( Z k x ) } = P ( x , x ) E Δ ( x , Z k x ) + O k η / d + O k 1 / d .
On the other hand, it can be seen that
k 1 E 1 i < j k P ( Z i k , Z j k ) 1 { ( Z i k , Z j k ) M S T ( Z k ) } = 1 2 E j = 2 k P ( Z 1 k , Z j k ) 1 { ( Z i k , Z j k ) M S T ( Z k ) } = 1 2 g k ( x ) d x E j = 2 k P ( x , Z j k ) 1 { ( x , Z j k ) M S T ( Z k ) } .
Recalling (A57),
= 1 2 g k ( x ) P ( x , x ) E Δ ( x , Z k x ) d x + O k η / d + O k 1 / d .
By virtue of Lemma A4, (A49) can be substituted into expression (A59) to obtain (A56). □
Theorem A1: Assume R m , n : = R ( X m , Y n ) denotes the FR test statistic as before. Then, the rate for the bias of the R m , n estimator for 0 < η 1 , d 2 is of the form:
| E R m , n m + n 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x | O l d ( m + n ) η / d + O ( l d η ) .
Here, η is the Holder smoothness parameter. A more explicit form for the bound on the RHS is given in (A61) below:
| E R m , n ( X m , Y n ) m + n 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x | O l d ( m + n ) η / d + O l d ( m + n ) 1 / 2 + 2 c 1 l d 1 ( m + n ) ( 1 / d ) 1 + c d 2 d ( m + n ) 1 2 l d ( m + n ) 1 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x + c 2 ( m + n ) 1 l d + O ( l ) ( m + n ) 1 i = 1 M l d ( a i ) 1 2 f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x + O ( l d η ) + O ( l ) i = 1 M l d / 2 b i a i 2 2 f 0 ( x ) f 1 ( x ) f 0 ( x ) m + f 1 ( x ) n m f 0 ( x ) + n f 1 ( x ) 2 d x + i = 1 M 2 l d / 2 b i a i 2 f 0 ( x ) f 1 ( x ) α i β i m a i f 0 2 ( x ) + n b i f 1 2 ( x ) 1 / 2 m f 0 ( x ) + n f 1 ( x ) 2 ( m + n ) d x .
Proof. 
Assume M m and N n be Poisson variables with mean m and n, respectively, one independent of another and of { X i } and { Y j } . Let also X m and Y n be the Poisson processes { X 1 , , X M n } and { Y 1 , , Y N n } . Set R m , n : = R m , n ( X m , Y n ) . Applying Lemma 1, and (12) cf. [49], we can write
| R m , n R m , n | K d ( | M m m | + | N n n | ) .
Here, K d denotes the largest possible degree of any vertex of the MST graph in R d . Moreover, by the matter of Poisson variable fact and using Stirling approximation [51], we have
E | M m m | = e m m m + 1 m ! e m m m + 1 2 π m m + 1 / 2 e m = O m 1 / 2 .
Similarly, E | N n n | = O ( n 1 / 2 ) . Therefore, by (A62), one yields
E [ R m , n ] = E R m , n R m , n + E R m , n = O ( m + n ) 1 / 2 + E R m , n .
Therefore,
E [ R m , n ] m + n = E R m , n m + n + O ( m + n ) 1 / 2 .
Hence, it will suffice to obtain the rate of convergence of E R m , n / ( m + n ) in the RHS of (A65). For this, let m i , n i denote the number of Poisson process samples X m and Y n with the FR statistic R m , n , falling into partitions Q i with FR statistic R m i , n i . Then, by virtue of Lemma 4, we can write
E R m , n i = 1 M E R m i , n i + 2 c 1 l d 1 ( m + n ) 1 / d .
Note that the Binomial RVs m i , n i are independent with marginal distributions m i B ( m , a i l d ) , n i B ( n , b i l d ) , where a i , b i are non-negative constants satisfying, i , a i b i and i = 1 l d a i l d = i = 1 l d b i l d = 1 . Therefore,
E R m , n i = 1 M E E R m i , n i | m i , n i + 2 c 1 l d 1 ( m + n ) 1 / d .
Let us first compute the internal expectation given m i , n i . For this reason, given m i , n i , let Z 1 m i , n i , Z 2 m i , n i , be independent variables with common densities g m i , n i ( x ) = m i f 0 ( x ) + n i f 1 ( x ) / ( m i + n i ) , x R d . Moreover, let L m i , n i be an independent Poisson variable with mean m i + n i . Denote F m i , n i = { Z 1 m i , n i , , Z L m i . n i m i , n i } a non-homogeneous Poisson of rate m i f 0 + n i f 1 . Let F m i , n i be the non-Poisson point process { Z 1 m i , n i , Z m i + n i m i , n i } . Assign a mark from the set { 1 , 2 } to each points of F m i , n i . Let X ˜ m i be the sets of points marked 1 with each probability m i f 0 ( x ) / m i f 0 ( x ) + n i f i ( x ) and let Y ˜ n i be the set points with mark 2. Note that owing to the marking theorem [52], X ˜ m i and Y ˜ n i are independent Poisson processes with the same distribution as X m i and Y n i , respectively. Considering R ˜ m i . n i as FR statistic over nodes in X ˜ m i Y ˜ n i we have
E R m i , n i | m i , n i = E R ˜ m i , n i | m i , n i .
Again using Lemma 1 and analogous arguments in [49] along with the fact that E | M m + N n m n | = O ( ( m + n ) 1 / 2 ) , we have
E R ˜ m i , n i | m i , n i = E E R ˜ m i , n i | F m i , n i = E s < j < m i + n i P m i , n i ( Z s m i , n i , Z j m i , n i ) 1 ( Z s m i , n i , Z j m i , n i ) F m i , n i + O ( ( m i + n i ) 1 / 2 ) ) .
Here,
P m i , n i ( x , y ) : = P r { mark x mark y , ( x , y ) F m i , n i } = m i f 0 ( x ) n i f 1 ( y ) + n i f 1 ( x ) m i f 0 ( y ) m i f 0 ( x ) + n i f 1 ( x ) m i f 0 ( y ) + n i f 1 ( y ) .
By owing to Lemma A6, we obtain
i = 1 M E m i , n i E s < j < m i + n i P m i , n i ( Z s m i , n i , Z j m i , n i ) 1 ( Z s m i , n i , Z j m i , n i ) F m i , n i + i = 1 M E m i , n i O ( m i + n i ) 1 / 2 = i = 1 M E m i , n i [ ( m i + n i ) g m i , n i ( x , x ) P m i , n i ( x , x ) d x + ( ς η ( l , m i , n i ) + O ( m i + n i ) η / d + O ( m i + n i ) 1 / d ) ( m i + n i ) ] + i = 1 M E m i , n i O ( m i + n i ) 1 / 2 ,
where
ς η ( l , m i , n i ) = O l / ( m i + n i ) 2 l d / ( m i + n i ) g m i , n i ( x ) P m i , n i ( x , x ) d x + O ( l d η ) .
The expression in (A67) equals
i = 1 M E m i , n i 2 m i n i f 0 ( x ) f 1 ( x ) m i f 0 ( x ) + n i f 1 ( x ) d x + i = 1 M E m i , n i ( m i + n i ) ς η ( l , m i , n i ) + O l d ( m + n ) 1 η / d + O l d ( m + n ) 1 / 2 .
Because of Jensen inequality for concave function:
i = 1 M E m i , n i O ( m i + n i ) 1 / 2 = i = 1 M O E [ m i ] + E [ n i ] 1 / 2 = i = 1 M O ( m a i l d + n b i l d ) 1 / 2 = O l d ( m + n ) 1 / 2 .
In addition, similarly since η < d , we have
i = 1 M E m i , n i O ( m i + n i ) 1 η / d = O l d ( m + n ) 1 η / d ,
and, for d 2 , one yields
i = 1 M E m i , n i O ( m i + n i ) 1 1 / d = O l d ( m + n ) 1 1 / d = O l d ( m + n ) 1 / 2 .
Next, we state the following lemma (Lemma 1 from [30,31]), which will be used in the sequel:
Lemma A13.
Let k ( x ) be a continuously differential function of x R which is convex and monotone decreasing over x 0 . Set k ( x ) = d k ( x ) d x . Then, for any x 0 > 0 , we have
k ( x 0 ) + k ( x 0 ) x 0 | x x 0 | k ( x ) k ( x 0 ) k ( x 0 ) | x x 0 | .
Next, continuing the proof of (A60), we attend to find an upper bound for
E m i , n i m i n i m i f 0 ( x ) + n i f 1 ( x ) .
In order to pursue this aim, in Lemma A13, consider k ( x ) = 1 x and x 0 = E m i , n i m i f 0 ( x ) + n i f 1 ( x ) , therefore as the function k ( x ) is decreasing and convex, one can write
1 m i f 0 ( x ) + n i f 1 ( x ) 1 E m i , n i m i f 0 ( x ) + n i f 1 ( x ) + | m i f 0 ( x ) + n i f 1 ( x ) E m i , n i m i f 0 ( x ) + n i f 1 ( x ) | E m i , n i 2 m i f 0 ( x ) + n i f 1 ( x ) .
Using the Hölder inequality implies the following inequality:
E m i , n i m i n i m i f 0 ( x ) + n i f 1 ( x ) E m i , n i [ m i n i ] E m i , n i m i f 0 ( x ) + n i f 1 ( x ) + E m i , n i m i 2 n i 2 1 / 2 E m i , n i 2 m i f 0 ( x ) + n i f 1 ( x ) × E m i , n i m i f 0 ( x ) + n i f 1 ( x ) E m i , n i m i f 0 ( x ) + n i f 1 ( x ) 2 1 / 2 .
As random variables m i , n i are independent, and because of V [ m i ] m a i l d , V [ n i ] n b i l d , we can claim that the RHS of (A74) becomes less than and equal to
m n a i b i l 2 d m a i l d f 0 ( x ) + n b i l d f 1 ( x ) + α i β i m a i l d f 0 2 ( x ) + n b i l d f 1 2 ( x ) 1 / 2 m a i f 0 ( x ) + n b i f 1 ( x ) 2 ,
where
α i = m a i l d ( 1 a i l d ) + m 2 a i 2 , β i = n b i l d ( 1 b i l d ) + n 2 b i 2 .
Going back to (A66), we have
E R m , n ( X m , Y n ) i = 1 M a i b i l d 2 m n f 0 ( x ) f 1 ( x ) m a i f 0 ( x ) + n b i f 1 ( x ) d x + i = 1 M 2 f 0 ( x ) f 1 ( x ) α i β i m a i l d f 0 2 ( x ) + n b i l d f 1 2 ( x ) 1 / 2 m a i f 0 ( x ) + n b i f 1 ( x ) 2 d x + i = 1 M E m i , n i ( m i + n i ) ς η ( l , m i , n i ) + O l d ( m + n ) 1 η / d + O l d ( m + n ) 1 / 2 + 2 c 1 l d 1 ( m + n ) 1 / d .
Finally, owing to a i b i and i = 1 M b i l d = 1 , when m m + n p , we have
E R m , n ( X m , Y n ) m + n 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x + i = 1 M 2 f 0 ( x ) f 1 ( x ) α i β i m a i l d f 0 2 ( x ) + n b i l d f 1 2 ( x ) 1 / 2 m a i f 0 ( x ) + n b i f 1 ( x ) 2 ( m + n ) d x + 1 m + n i = 1 M E m i , n i ( m i + n i ) ς η ( l , m i , n i ) + O l d ( m + n ) η / d + O l d ( m + n ) 1 / 2 + 2 c 1 l d 1 ( m + n ) ( 1 / d ) 1 .
Passing to Definition 2, MST * , and Lemma A2, a similar discussion as above, consider the Poisson processes samples and the FR statistic under the union of samples, denoted by R m , n * , and superadditivity of dual R m , n * , we have
E R m , n * ( X m , Y n ) i = 1 M E R m i , n i * ( X m , Y n ) Q i c 2 l d = i = 1 M E m i , n i E R m i , n i * ( X m , Y n ) Q i | m i , n i c 2 l d i = 1 M E m i , n i E R m i , n i ( X m , Y n ) Q i | m i , n i c 2 l d ,
the last line is derived from Lemma A2, (ii), inequality (A8). Owing to the Lemma A6, (A69), and (A70), one obtains
E R m , n * ( X m , Y n ) i = 1 M E m i , n i 2 m i n i f 0 ( x ) f 1 ( x ) m i f 0 ( x ) + n i f 1 ( x ) d x i = 1 M E m i , n i ( m i + n i ) ς η ( l , m i , n i ) O l d ( m + n ) 1 η / d O l d ( m + n ) 1 / 2 c 2 l d .
Furthermore, by using the Jenson’s inequality, we get
E m i , n i m i n i m i f 0 ( x ) + n i f 1 ( x ) E [ m i ] E [ n i ] E [ m i ] f 0 ( x ) + E [ n i ] f 1 ( x ) = l d m a i n b i m a i f 0 ( x ) + n b i f 1 ( x ) .
Therefore, since a i b i , we can write
E m i , n i m i n i m i f 0 ( x ) + n i f 1 ( x ) l d m n a i b i b i m f 0 ( x ) + n f 1 ( x ) = l d m n a i m f 0 ( x ) + n f 1 ( x ) .
Consequently, the RHS of (A79) becomes greater than or equal to
i = 1 M a i l d 2 m n f 0 ( x ) f 1 ( x ) m f 0 ( x ) + n f 1 ( x ) d x i = 1 M E m i , n i ( m i + n i ) ς η ( l , m i , n i ) O l d ( m + n ) 1 η / d O l d ( m + n ) 1 / 2 c 2 l d .
Finally, since i = 1 M a i l d = 1 and m m + n p , we have
E R m , n * ( X m , Y n ) m + n 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x ( m + n ) 1 i = 1 M E m i , n i ( m i + n i ) ς ( l , m i , n i ) O l d ( m + n ) η / d O l d ( m + n ) 1 / 2 c 2 l d ( m + n ) 1 .
By definition of the dual R m , n * and (i) in Lemma A2,
E R m , n ( X m , Y n ) m + n + c d 2 d m + n E R m , n * ( X m , Y n ) m + n ,
we can imply
E R m , n ( X m , Y n ) m + n 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x ( m + n ) 1 i = 1 M E m i , n i ( m i + n i ) ς η ( l , m i , n i ) O l d ( m + n ) η / d O l d ( m + n ) 1 / 2 c 2 l d ( m + n ) 1 c d 2 d ( m + n ) 1 .
The combination of two lower and upper bounds (A84) and (A77) yields the following result
| E R m , n ( X m , Y n ) m + n 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x | O l d ( m + n ) η / d + O l d ( m + n ) 1 / 2 + 2 c 1 l d 1 ( m + n ) ( 1 / d ) 1 + c d 2 d ( m + n ) 1 + c 2 ( m + n ) 1 l d + 1 m + n i = 1 M E m i , n i ( m i + n i ) ς η ( l , m i , n i ) + i = 1 M 2 f 0 ( x ) f 1 ( x ) α i β i m a i l d f 0 2 ( x ) + n b i l d f 1 2 ( x ) 1 / 2 m a i f 0 ( x ) + n b i f 1 ( x ) 2 ( m + n ) d x .
Recall ς η ( l , m i , n i ) , then we obtain
i = 1 M E m i , n i ( m i + n i ) ς η ( l , m i , n i ) = i = 1 M O ( l ) E 2 m i n i f 0 ( x ) f 1 ( x ) ( m i + n i ) ( m i f 0 ( x ) + n i f 1 ( x ) ) d x 2 l d i = 1 M E 2 m i n i f 0 ( x ) f 1 ( x ) ( m i + n i ) ( m i f 0 ( x ) + n i f 1 ( x ) ) d x + O ( l η ) i = 1 M E m i , n i [ m i + n i ] .
In addition, we have
E m i , n i 2 m i n i f 0 ( x ) f 1 ( x ) ( m i + n i ) ( m i f 0 ( x ) + n i f 1 ( x ) ) 1 m + n E m i , n i 2 m i n i f 0 ( x ) f 1 ( x ) ( m i f 0 ( x ) + n i f 1 ( x ) ) .
This implies
i = 1 M E 2 m i n i f 0 ( x ) f 1 ( x ) ( m i + n i ) ( m i f 0 ( x ) + n i f 1 ( x ) ) d x 2 p q f 0 ( x ) f 1 ( x ) p f 0 ( x ) + q f 1 ( x ) d x .
Note that the above inequality is derived from (A80) and m m + n p . Furthermore,
1 m + n i = 1 M O ( l ) E m i , n i 2 m i n i f 0 ( x ) f 1 ( x ) ( m i + n i ) ( m i f 0 ( x ) + n i f 1 ( x ) ) d x i = 1 M O ( l ) E m i , n i 2 m i n i f 0 ( x ) f 1 ( x ) ( m i + n i ) 2 ( m i f 0 ( x ) + n i f 1 ( x ) ) d x i = 1 M O ( l ) E m i , n i 2 f 0 ( x ) f 1 ( x ) ( m i f 0 ( x ) + n i f 1 ( x ) ) d x .
The last line holds because of m i n i ( m i + n i ) 2 . Going back to (A73), we can give an upper bound for the RHS of above inequality as
E m i , n i m i f 0 ( x ) + n i f 1 ( x ) 1 m a i l d f 0 ( x ) + n b i l d f 1 ( x ) 1 + ( E m i , n i | m i f 0 ( x ) + n i f 1 ( x ) E [ m i ] f 0 ( x ) + E [ n i ] f 1 ( x ) | / m a i l d f 0 ( x ) + n b i l d f 1 ( x ) 2 .
Note that we have assumed a i b i and by using Hölder inequality we write
E m i , n i m i f 0 ( x ) + n i f 1 ( x ) 1 l d ( a i ) 1 m f 0 ( x ) + n f 1 ( x ) 1 + f 0 ( x ) V ( m i ) + f 1 ( x ) V ( n i ) / a i 2 l d ( m f 0 ( x ) + n f 1 ( x ) ) 2 l d ( a i ) 1 m f 0 ( x ) + n f 1 ( x ) 1 + l d / 2 b i f 0 ( x ) m + f 1 ( x ) n / a i 2 l d ( m f 0 ( x ) + n f 1 ( x ) ) 2 .
As result, we have
i = 1 M O ( l ) E m i , n i 2 f 0 ( x ) f 1 ( x ) ( m i f 0 ( x ) + n i f 1 ( x ) ) d x i = 1 M O ( l ) l d ( a i ) 1 2 f 0 ( x ) f 1 ( x ) m f 0 ( x ) + n f 1 ( x ) d x + i = 1 M O ( l ) l d / 2 b i 2 f 0 ( x ) f 1 ( x ) f 0 ( x ) m + f 1 ( x ) n a i 2 l d m f 0 ( x ) + n f 1 ( x ) 2 d x .
As a consequence, owing to (A85), for 0 < η 1 , d 2 , which implies η d 1 , we can derive (A61). Thus, the proof can be concluded by giving the summarized bound in (A60). □
Lemma A8: For h = 1 , 2 , , let δ m , n h be the function c h d 1 ( m + n ) 1 / d . Then, for ϵ > 0 , we have
P R m , n ( X m , Y n ) i = 1 h d R m i , n i ( X m i , Y n i ) + 2 ϵ ϵ δ m , n h ϵ .
Note that in case ϵ δ m , n h the above claimed inequality is trivial.
Proof. 
Consider the cardinality of the set of all edges of MST i = 1 h d Q i which intersect two different subcubes Q i and Q j , | D | . Using the Markov inequality, we can write
P | D | ϵ E ( | D | ) ϵ ,
where ϵ > 0 . Since E | D | c h d 1 ( m + n ) 1 / d : = δ m , n h , therefore for ϵ > δ m , n h and h = 1 , 2 , :
P | D | ϵ δ m , n h ϵ .
In addition, if Q i , i = 1 , h d is a partition of [ 0 , 1 ] d into congruent subcubes of edge length 1 / h , then
P i = 1 h d R m i , n i ( X m , Y n Q i ) + 2 | D | i = 1 h d R m i , n i ( X m , Y n Q i ) + 2 ϵ δ m , n h ϵ .
This implies
P i = 1 h d R m i , n i ( X m , Y n Q i ) + 2 | D | i = 1 h d R m i , n i ( X m , Y n Q i ) + 2 ϵ 1 δ m , n h ϵ .
By subadditivity (A6), we can write
R m , n ( X m , Y n ) i = 1 h d R m i , n i ( X m , Y n Q i ) + 2 | D | ,
and this along with (A94) establishes (A92). □
Lemma A9: (Growth bounds for R m , n ) Let R m , n be the FR statistic. Then, for given non-negative ϵ , such that ϵ h 2 δ m , n h , with at least probability g ( ϵ ) : = 1 h δ m , n h ϵ , h = 2 , 3 , , we have
R m , n ( X m , Y n ) c ϵ , h # X m # Y n 1 1 / d .
Here, c ϵ , h = O ϵ h d 1 1 depending only on ϵ , h. Note that, for ϵ < h 2 δ m , n h , the claim is trivial.
Proof. 
Without loss of generality, consider the unit cube [ 0 , 1 ] d . For given h, if Q i , i = 1 , h d is a partition of [ 0 , 1 ] d into congruent subcubes of edge length 1 / h , then, by Lemma A8, we have
P R m , n ( X m , Y n ) i = 1 h d R m i , n i ( X m i , Y n i ) + 2 ϵ ϵ δ m , n h ϵ .
We apply the induction methodology on # X m and # Y n . Set c : = sup x , y [ 0 , 1 ] d R m , n ( { x , y } ) which is finite according to assumption. Moreover, set c 2 : = 2 ϵ h d 1 1 and c 1 : = c + d h d 1 c 2 . Therefore, it is sufficient to show that for all ( X m , Y n ) [ 0 , 1 ] d with at least probability g ( ϵ )
R m , n ( X m , Y n ) c 1 # X m # Y n ( d 1 ) / d .
Alternatively, as for the induction hypothesis, we assume the stronger bound
R m , n ( X m , Y n ) c 1 # X m # Y n ( d 1 ) / d c 2
holds whenever # X m < m and # Y n < n with at least probability g ( ϵ ) . Note that d 2 , ϵ > 0 and c 1 , c 2 both depend on ϵ , h. Hence,
c 1 c 2 = c + c 2 d h d 1 1 c + c 2 h d 1 1 = c + 2 ϵ c ,
which implies P ( R m , n c 1 c 2 ) P ( R m , n c ) . In addition, we know that P ( R m , n c ) = 1 g ( ϵ ) ; therefore, the induction hypothesis holds particularly # X m = 1 and # Y n = 1 . Now, consider the partition Q i of [ 0 , 1 ] d ; therefore, for all 1 i h d , we have m i : = # ( X m Q i ) < m and n i : = # ( Y n Q i ) < n and thus, by induction hypothesis, one yields with at least probability g ( ϵ )
R m i , n i ( X m , Y n Q i ) c 1 ( m i n i ) 1 1 / d c 2 .
Set B the event all i : R m i , n i c 1 ( m i n i ) 1 1 / d c 2 and B i stands with the event R m i , n i c 1 ( m i n i ) 1 1 / d c 2 . From (A96) and since Q i ’s are partitions, which implies
P ( B ) = P ( B i ) h d P ( B i ) , P ( B c ) = P ( i = 1 l d B i c ) i = 1 h d P ( B i c ) h d 1 g ( ϵ ) , and P ( B ) = i = 1 h d P ( B i ) g ( ϵ ) h d ,
we thus obtain
ϵ δ m , n h ϵ P R m , n i = 1 h d R m i , n i ( X m i , Y n i ) + 2 ϵ | B P ( B ) + P R m , n i = 1 h d R m i , n i ( X m i , Y n i ) + 2 ϵ | B c P ( B c ) P R m , n i = 1 l d R m i , n i ( X m i , Y n i ) + 2 ϵ | B P ( B ) + P ( B c ) .
Equivalently,
P R m , n i = 1 h d R m i , n i ( X m i , Y n i ) + 2 ϵ | B 1 δ m , n h ϵ 1 + P ( B ) / P ( B ) = 1 δ m , n h ϵ P ( B ) .
In fact, in this stage, we want to show that
1 δ m , n h ϵ P ( B ) g ( ϵ ) or P ( B ) δ m , n h ϵ ( 1 g ( ϵ ) ) .
Since P ( B ) g ( ϵ ) h d , therefore it is sufficient to derive that g ( ϵ ) h d δ m , n h ϵ ( 1 g ( ϵ ) ) . Indeed, for given g ( ϵ ) = ϵ h δ m , n h ϵ , we have g ( ϵ ) ϵ δ m , n h ϵ hence δ m , n h ϵ ( 1 g ( ϵ ) ) = 1 h 1 . Furthermore, we know 1 h 1 1 h ( 1 / h d ) and since ϵ h 2 δ m , n h this implies h δ m , n h ϵ 1 h and consequently
h δ m , n h ϵ 1 1 h h d
or
g ( ϵ ) h d = ϵ h δ m , n h ϵ h d 1 h = δ m , n h ϵ ( 1 g ( ϵ ) ) .
This implies the fact that for ϵ h 2 δ m , n h
P R m , n i = 1 h d c 1 ( m i n i ) 1 1 / d c 2 + 2 ϵ g ( ϵ ) , where g ( ϵ ) = ϵ h δ m , n h ϵ .
Now, let γ : = # { i : m i , n i > 0 } and using Hölder inequality gives
P R m , n ( X m , Y n ) c 1 γ 1 / d ( m n ) 1 1 / d γ c 2 + c 2 ( h d 1 1 ) g ( ϵ ) .
Next, we just need to show that c 1 γ 1 / d ( m n ) 1 1 / d γ c 2 + c 2 ( h d 1 1 ) in (A100) is less than or equal to c 1 ( m n ) 1 1 / d c 2 , which is equivalent to show
c 2 h d 1 γ c 1 ( m n ) 1 1 / d ( 1 γ 1 / d ) .
We know that m , n 1 and c 1 d h d 1 c 2 , so it is sufficient to get
c 2 h d 1 γ d h d 1 c 2 ( 1 γ 1 / d ) ,
choose t as γ = t h d , then 0 < t 1 , so (A101) becomes
( h 1 t ) d h 1 ( 1 h t 1 / d ) .
Note that the function d h 1 ( 1 h t 1 / d ) + t h 1 has a minimum at t = 1 which implies (A101) and subsequently (A95). Hence, the proof is completed. □
Lemma A10: (Smoothness for R m , n ) Given observations of
X m : = ( X m , X m ) = { X 1 , , X m , X m + 1 , , X m } ,
such that m + m = m and Y n : = ( Y n , Y n ) = { Y 1 , , Y n , Y n + 1 , , Y n } , where n + n = n , denote R m , n ( X m , Y n ) as before, the number of edges of MST ( X m , Y n ) which connect a point of X m to a point of Y n . Then, for integer h 2 , for all ( X n , Y m ) [ 0 , 1 ] d , ϵ h 2 δ m , n h , where δ m , n h = O h d 1 ( m + n ) 1 / d , we have
P | R m , n ( X m , Y n ) R m , n ( X m , Y n ) | c ˜ ϵ , h # X m # Y n 1 1 / d 1 2 h δ m , n h ϵ ,
where c ˜ ϵ , h = O ϵ h d 1 1 . For the case ϵ < h 2 δ m , n h , this holds trivially.
Proof. 
We begin with removing the edges which contain a vertex in X m and Y n in minimal spanning tree on ( X m , Y n ) . Now, since each vertex has bounded degree, say c d , we can generate a subgraph in which has at most c d ( # X m + # Y n ) components. Next, choose one vertex from each component and form the minimal spanning tree on these vertices, assuming all of them can be considered in FR test statistic, we can write
R m , n ( X m , Y n ) R m , n ( X m , Y n ) + c ϵ , h c d 2 # X m # Y n 1 1 / d , or equivalently R m , n ( X m , Y n ) + c ϵ 1 h # X m # Y n 1 1 / d ,
with probability at least g ( ϵ ) , where g ( ϵ ) is as in Lemma A9. Note that this expression is obtained from Lemma A9. In this stage, it remains to show that with at least probability g ( ϵ )
R m , n ( X m , Y n ) R m , n ( X m , Y n ) c ˜ ϵ , h # X m # Y n 1 1 / d ,
which, again by using the method before, with at least probability g ( ϵ ) , one derives
R m , n ( X m , Y n ) R m , n ( X m , Y n ) + c ^ ϵ , h c d 2 ( # X m # Y n ) 1 1 / d , o r e q u i v a l e n t l y R m , n ( X m , Y n ) + c ϵ 2 h # X m # Y n 1 1 / d .
Letting c ˜ ϵ , h = max { c ϵ 1 h , c ϵ 2 h } implies (A105). Thus,
P | R m , n ( X m , Y n ) R m , n ( X m , Y n ) | c ˜ ϵ , h # X m # Y n 1 1 / d 2 2 g ( ϵ ) ,
Hence, the smoothness is given with at least probability 2 g ( ϵ ) 1 as in the statement of Lemma A10. □
Lemma A11: (Semi-Isoperimetry) Let μ be a measure on [ 0 , 1 ] d ; μ n denotes the product measure on space ( [ 0 , 1 ] d ) n . In addition, let M e denotes a median of R m , n . Set
A : = X m [ 0 , 1 ] d m , Y n [ 0 , 1 ] d n ; R m , n ( X m , Y n ) M e .
Then,
μ m + n x ( [ 0 , 1 ] d ) m , y ( [ 0 , 1 ] n ) : ϕ A ( x ) ϕ A ( y ) t 4 exp t 8 ( m + n ) .
Proof. 
Let ϕ A ( z ) = min { H ( z , z ) , z A } . Using Proposition 6.5 in [17], isoperimetric inequality, we have
μ m + n z ( [ 0 , 1 ] d ) m + n : ϕ A ( z ) t 4 exp t 2 8 ( m + n ) .
Furthermore, we know that
ϕ A ( x ) + ϕ A ( y ) 2 ϕ A ( x ) ϕ A ( y ) ,
hence
μ m + n ( x ( [ 0 , 1 ] d ) m , y ( [ 0 , 1 ] n ) : ϕ A ( x ) ϕ A ( y ) t μ m + n ( x ( [ 0 , 1 ] d ) m , y ( [ 0 , 1 ] n ) : ϕ A ( x ) + ϕ A ( y ) 2 t = μ m + n ( x ( [ 0 , 1 ] d ) m , y ( [ 0 , 1 ] n ) : ϕ A ( x ) + ϕ A ( y ) t .
The last equality in (A110) achieves because of ϕ A ( x ) , ϕ A ( y ) 0 and note that ϕ A ( z ) ϕ A ( x ) + ϕ A ( y ) . Therefore,
μ m + n ( x ( [ 0 , 1 ] d ) m , y ( [ 0 , 1 ] n ) : ϕ A ( x ) + ϕ A ( y ) t μ m + n ( z ( [ 0 , 1 ] d ) m + n : ϕ A ( z ) t .
By recalling (A109), we derive the bound (A108). □
Lemma A12: (Deviation of the Mean and Median) Consider M e as a median of R m , n . Then, for given g ( ϵ ) = 1 h δ m , n h ϵ , and δ m , n h = O h d 1 ( m + n ) 1 / d such that for h 7 , ϵ h 2 δ m , n h , we have
| E R m , n ( X m , Y n ) M e | C m , n ( ϵ , h ) ( m + n ) ( d 1 ) / d ,
where C m , n ( ϵ , h ) stands with a form depends on ϵ , h, m, n as
C m , n ( ϵ , h ) = C 1 2 ( 2 g ( ϵ ) 1 ) 2 1 1 ,
where C is a constant.
Proof. 
Following the analogous arguments in [17,53], we have
| E R m , n ( X m , Y n ) M e | E | R m , n ( X m , Y n ) M e | = 0 P | R m , n ( X m , Y n ) M e | t d t 8 1 1 / 2 ( 2 g ( ϵ ) 1 ) 2 1 0 exp t d / ( d 1 ) 8 ( 4 ϵ ) d / d 1 ( m + n ) d t = C 1 2 ( 2 g ( ϵ ) 1 ) 2 1 1 ( m + n ) ( d 1 ) / d ,
where g ( ϵ ) = 1 h O h d 1 ( m + n ) 1 / d / ϵ . The inequality in (A113) is implied from Theorem 5. Hence, the proof is completed. □

References

  1. Xuan, G.; Chia, P.; Wu, M. Bhattacharyya distance feature selection. In Proceedings of the 13th International Conference on Pattern Recognition, Vienna, Austria, 25–29 August 1996; Volume 2, pp. 195–199. [Google Scholar]
  2. Hamza, A.; Krim, H. Image registration and segmentation by maximizing the Jensen-Renyi divergence. In Energy Minimization Methods in Computer Vision and Pattern Recognition. EMMCVPR 2003; Springer: Berlin/Heidelberg, Germany, 2003; pp. 147–163. [Google Scholar]
  3. Hild, K.E.; Erdogmus, D.; Principe, J. Blind source separation using Renyi’s mutual information. IEEE Signal Process. Lett. 2001, 8, 174–176. [Google Scholar] [CrossRef]
  4. Basseville, M. Divergence measures for statistical data processing–An annotated bibliography. Signal Process. 2013, 93, 621–633. [Google Scholar] [CrossRef]
  5. Battacharyya, A. On a measure of divergence between two multinomial populations. Sankhy ā Indian J. Stat. 1946, 7, 401–406. [Google Scholar]
  6. Lin, J. Divergence Measures Based on the Shannon Entropy. IEEE Trans. Inf. Theory 1991, 37, 145–151. [Google Scholar] [CrossRef]
  7. Berisha, V.; Hero, A. Empirical non-parametric estimation of the Fisher information. IEEE Signal Process. Lett. 2015, 22, 988–992. [Google Scholar] [CrossRef]
  8. Berisha, V.; Wisler, A.; Hero, A.; Spanias, A. Empirically estimable classification bounds based on a nonparametric divergence measure. IEEE Trans. Signal Process. 2016, 64, 580–591. [Google Scholar] [CrossRef]
  9. Moon, K.; Hero, A. Multivariate f-divergence estimation with confidence. In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014; pp. 2420–2428. [Google Scholar]
  10. Moon, K.; Hero, A. Ensemble estimation of multivariate f-divergence. In Proceedings of the IEEE International Symposium on Information Theory (ISIT), Honolulu, HI, USA, 29 June–4 July 2014; pp. 356–360. [Google Scholar]
  11. Moon, K.; Sricharan, K.; Greenewald, K.; Hero, A. Improving convergence of divergence functional ensemble estimators. In Proceedings of the IEEE International Symposium on Information Theory (ISIT), Barcelona, Spain, 10–15 July 2016; pp. 1133–1137. [Google Scholar]
  12. Moon, K.; Sricharan, K.; Greenewald, K.; Hero, A. Nonparametric ensemble estimation of distributional functionals. arXiv 2016, arXiv:1601.06884v2. [Google Scholar]
  13. Noshad, M.; Moon, K.; Yasaei Sekeh, S.; Hero, A. Direct Estimation of Information Divergence Using Nearest Neighbor Ratios. In Proceedings of the IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017. [Google Scholar]
  14. Yasaei Sekeh, S.; Oselio, B.; Hero, A. A Dimension-Independent discriminant between distributions. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018. [Google Scholar]
  15. Noshad, M.; Hero, A. Rate-optimal Meta Learning of Classification Error. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018. [Google Scholar]
  16. Wisler, A.; Berisha, V.; Wei, D.; Ramamurthy, K.; Spanias, A. Empirically-estimable multi-class classification bounds. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016. [Google Scholar]
  17. Yukich, J. Probability Theory of Classical Euclidean Optimization; Lecture Notes in Mathematics; Springer: Berlin, Germany, 1998; Volume 1675. [Google Scholar]
  18. Steele, J. An Efron–Stein inequality for nonsymmetric statistics. Ann. Stat. 1986, 14, 753–758. [Google Scholar] [CrossRef]
  19. Aldous, D.; Steele, J.M. Asymptotic for Euclidean minimal spanning trees on random points. Probab. Theory Relat. Fields 1992, 92, 247–258. [Google Scholar] [CrossRef]
  20. Ma, B.; Hero, A.; Gorman, J.; Michel, O. Image registration with minimal spanning tree algorithm. In Proceedings of the IEEE International Conference on Image Processing, Vancouver, BC, Canada, 10–13 September 2000; pp. 481–484. [Google Scholar]
  21. Neemuchwala, H.; Hero, A.; Carson, P. Image registration using entropy measures and entropic graphs. Eur. J. Signal Process. 2005, 85, 277–296. [Google Scholar] [CrossRef]
  22. Hero, A.; Ma, B.; Michel , O.J.; Gorman, J. Applications of entropic spanning graphs. IEEE Signal Process. Mag. 2002, 19, 85–95. [Google Scholar] [CrossRef]
  23. Hero, A.; Michel, O. Estimation of Rényi information divergence via pruned minimal spanning trees. In Proceedings of the IEEE Workshop on Higher Order Statistics, Caesarea, Isreal, 16 June 1999. [Google Scholar]
  24. Smirnov, N. On the estimation of the discrepancy between empirical curves of distribution for two independent samples. Bull. Mosc. Univ. 1939, 2, 3–6. [Google Scholar]
  25. Wald, A.; Wolfowitz, J. On a test whether two samples are from the same population. Ann. Math. Stat. 1940, 11, 147–162. [Google Scholar] [CrossRef]
  26. Gibbons, J. Nonparametric Statistical Inference; McGraw-Hill: New York, NY, USA, 1971. [Google Scholar]
  27. Singh, S.; Póczos, B. Probability Theory and Combinatorial Optimization; CBMF-NSF Regional Conference in Applied Mathematics; Society for Industrial and Applied Mathematics (SIAM): Philadelphia, PA, USA, 1997; Volume 69. [Google Scholar]
  28. Redmond, C.; Yukich, J. Limit theorems and rates of convergence for Euclidean functionals. Ann. Appl. Probab. 1994, 4, 1057–1073. [Google Scholar] [CrossRef]
  29. Redmond, C.; Yukich, J. Asymptotics for Euclidean functionals with power weighted edges. Stoch. Process. Their Appl. 1996, 6, 289–304. [Google Scholar] [CrossRef]
  30. Hero, A.; Costa, J.; Ma, B. Convergence Rates of Minimal Graphs with Random Vertices. Available online: https://pdfs.semanticscholar.org/7817/308a5065aa0dd44098319eb66f81d4fa7a14.pdf (accessed on 18 November 2019).
  31. Hero, A.; Costa, J.; Ma, B. Asymptotic Relations between Minimal Graphs and Alpha-Entropy; Tech. Rep.; Communication and Signal Processing Laboratory (CSPL), Department EECS, University of Michigan: Ann Arbor, MI, USA, 2003. [Google Scholar]
  32. Lorentz, G. Approximation of Functions; Holt, Rinehart and Winston: New York, NY, USA, 1996. [Google Scholar]
  33. Talagrand, M. Concentration of measure and isoperimetric inequalities in product spaces. Publications Mathématiques de i’I. H. E. S. 1995, 81, 73–205. [Google Scholar] [CrossRef]
  34. Kullback, S.; Leibler, R. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  35. Rényi, A. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, USA, 20 June–30 July 1961; pp. 547–561. [Google Scholar]
  36. Ali, S.; Silvey, S.D. A general class of coefficients of divergence of one distribution from another. J. R. Stat. Soc. Ser. B (Methodol.) 1996, 28, 131–142. [Google Scholar] [CrossRef]
  37. Cha, S. Comprehensive survey on distance/similarity measures between probability density functions. Int. J. Math. Models Methods Appl. Sci. 2007, 1, 300–307. [Google Scholar]
  38. Rukhin, A. Optimal estimator for the mixture parameter by the method of moments and information affinity. In Proceedings of the 12th Prague Conference on Information Theory, Prague, Czech Republic, 29 August–2 September 1994; pp. 214–219. [Google Scholar]
  39. Toussaint, G. The relative neighborhood graph of a finite planar set. Pattern Recognit. 1980, 12, 261–268. [Google Scholar] [CrossRef]
  40. Zahn, C. Graph-theoretical methods for detecting and describing Gestalt clusters. IEEE Trans. Comput. 1971, 100, 68–86. [Google Scholar] [CrossRef]
  41. Banks, D.; Lavine, M.; Newton, H. The minimal spanning tree for nonparametric regression and structure discovery. In Computing Science and Statistics, Proceedings of the 24th Symposium on the Interface; Joseph Newton, H., Ed.; Interface Foundation of North America: Fairfax Station, FA, USA, 1992; pp. 370–374. [Google Scholar]
  42. Hoffman, R.; Jain, A. A test of randomness based on the minimal spanning tree. Pattern Recognit. Lett. 1983, 1, 175–180. [Google Scholar] [CrossRef]
  43. Efron, B.; Stein, C. The jackknife estimate of variance. Ann. Stat. 1981, 9, 586–596. [Google Scholar] [CrossRef]
  44. Singh, S.; Póczos, B. Generalized exponential concentration inequality for Rényi divergence estimation. In Proceedings of the 31st International Conference on Machine Learning (ICML-14), Bejing, China, 22–24 June 2014; pp. 333–341. [Google Scholar]
  45. Singh, S.; Póczos, B. Exponential concentration of a density functional estimator. In Proceedings of the 27th International Conference on Neural Information Processing Systems (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014; pp. 3032–3040. [Google Scholar]
  46. Lichman, M. UCI Machine Learning Repository. 2013. Available online: https://www.re3data.org/repository/r3d100010960 (accessed on 18 November 2019).
  47. Bhatt, R.B.; Sharma, G.; Dhall, A.; Chaudhury, S. Efficient skin region segmentation using low complexity fuzzy decision tree model. In Proceedings of the IEEE-INDICON, Ahmedabad, India, 16–18 December 2009; pp. 1–4. [Google Scholar]
  48. Steele, J.; Shepp, L.; Eddy, W. On the number of leaves of a euclidean minimal spanning tree. J. Appl. Prob. 1987, 24, 809–826. [Google Scholar] [CrossRef]
  49. Henze, N.; Penrose, M. On the multivarite runs test. Ann. Stat. 1999, 27, 290–298. [Google Scholar]
  50. Rhee, W. A matching problem and subadditive Euclidean funetionals. Ann. Appl. Prob. 1993, 3, 794–801. [Google Scholar] [CrossRef]
  51. Whittaker, E.; Watson, G. A Course in Modern Analysis, 4th ed.; Cambridge University Press: New York, NY, USA, 1996. [Google Scholar]
  52. Kingman, J. Poisson Processes; Oxford Univ. Press: Oxford, UK, 1993. [Google Scholar]
  53. Pál, D.; Póczos, B.; Szapesvári, C. Estimation of Renyi entropy andmutual information based on generalized nearest-neighbor graphs. In Proceedings of the 23th International Conference on Neural Information Processing Systems (NIPS 2010), Vancouver, BC, Canada, 6–9 December 2010. [Google Scholar]
Figure 1. Heat map of the theoretical MSE rate of the FR estimator of the HP-divergence based on Theorems 2 and 3 as a function of dimension and sample size when N = m = n . Note the color transition (MSE) as sample size increases for high dimension. For fixed sample size N, the MSE rate degrades in higher dimensions.
Figure 1. Heat map of the theoretical MSE rate of the FR estimator of the HP-divergence based on Theorems 2 and 3 as a function of dimension and sample size when N = m = n . Note the color transition (MSE) as sample size increases for high dimension. For fixed sample size N, the MSE rate degrades in higher dimensions.
Entropy 21 01144 g001
Figure 2. The dual MST spanning the merged set X m (blue points) and Y n (red points) drawn from two Gaussian distributions. The dual FR statistic ( R m , n * ) is the number of edges in the MST * (contains nodes in X m Y n { 2 corner points } ) that connect samples from different color nodes and corners (denoted in green). Black edges are the non-dichotomous edges in the MST * .
Figure 2. The dual MST spanning the merged set X m (blue points) and Y n (red points) drawn from two Gaussian distributions. The dual FR statistic ( R m , n * ) is the number of edges in the MST * (contains nodes in X m Y n { 2 corner points } ) that connect samples from different color nodes and corners (denoted in green). Black edges are the non-dichotomous edges in the MST * .
Entropy 21 01144 g002
Figure 3. Comparison of the bound on the MSE theory and experiments for d = 2 , 4 , 8 standard Gaussian random vectors versus sample size from 100 trials.
Figure 3. Comparison of the bound on the MSE theory and experiments for d = 2 , 4 , 8 standard Gaussian random vectors versus sample size from 100 trials.
Entropy 21 01144 g003
Figure 4. Comparison of experimentally predicted MSE of the FR-statistic as a function of sample size m = n in various distributions Standard Normal, Gamma ( α 1 = α 2 = 1 , β 1 = β 2 = 1 , ρ = 0.5 ) and Standard t-Student.
Figure 4. Comparison of experimentally predicted MSE of the FR-statistic as a function of sample size m = n in various distributions Standard Normal, Gamma ( α 1 = α 2 = 1 , β 1 = β 2 = 1 , ρ = 0.5 ) and Standard t-Student.
Entropy 21 01144 g004
Figure 5. HP-divergence vs. sample size for three real datasets HAR, SKIN, and ENGIN.
Figure 5. HP-divergence vs. sample size for three real datasets HAR, SKIN, and ENGIN.
Entropy 21 01144 g005
Figure 6. The empirical MSE vs. sample size. The empirical MSE of the FR estimator for all three datasets HAR, SKIN, and ENGIN decreases for larger sample size N.
Figure 6. The empirical MSE vs. sample size. The empirical MSE of the FR estimator for all three datasets HAR, SKIN, and ENGIN decreases for larger sample size N.
Entropy 21 01144 g006
Figure 7. HP-divergence vs. dimension for three datasets HAR, SKIN, and ENGIN.
Figure 7. HP-divergence vs. dimension for three datasets HAR, SKIN, and ENGIN.
Entropy 21 01144 g007
Table 1. R m , n , D ^ p , m, and n are the FR test statistic, HP-divergence estimates using R m , n , and sample sizes for two classes, respectively.
Table 1. R m , n , D ^ p , m, and n are the FR test statistic, HP-divergence estimates using R m , n , and sample sizes for two classes, respectively.
FR Test Statistic
Dataset E [ R m , n ] D ^ p m n Variance-Like Interval
HAR30.995600600(2.994,3.006)
SKIN4.20.993600600(4.196,4.204)
ENGIN1.80.997600600(1.798,1.802)

Share and Cite

MDPI and ACS Style

Sekeh, S.Y.; Noshad, M.; Moon, K.R.; Hero, A.O. Convergence Rates for Empirical Estimation of Binary Classification Bounds. Entropy 2019, 21, 1144. https://doi.org/10.3390/e21121144

AMA Style

Sekeh SY, Noshad M, Moon KR, Hero AO. Convergence Rates for Empirical Estimation of Binary Classification Bounds. Entropy. 2019; 21(12):1144. https://doi.org/10.3390/e21121144

Chicago/Turabian Style

Sekeh, Salimeh Yasaei, Morteza Noshad, Kevin R. Moon, and Alfred O. Hero. 2019. "Convergence Rates for Empirical Estimation of Binary Classification Bounds" Entropy 21, no. 12: 1144. https://doi.org/10.3390/e21121144

APA Style

Sekeh, S. Y., Noshad, M., Moon, K. R., & Hero, A. O. (2019). Convergence Rates for Empirical Estimation of Binary Classification Bounds. Entropy, 21(12), 1144. https://doi.org/10.3390/e21121144

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop