1. Introduction
Probability measures on treated in this paper are absolutely continuous with respect to the standard Lebesgue measure and we shall identify them with their densities.
For a probability measure
f, the entropy
and the Fisher information
can be introduced, which play important roles in information theory, probability, and statistics. For more details on these subjects see the famous book [
1].
Hereafter, for an
n-variables function
on
, the integral of
ϕ over the whole
by the standard Lebesgue measure
is abbreviated as
that is, we shall leave out
in the integrand in order to simplify the expressions.
Definition 1.1. Let
f be a probability measure on
. Then
the (
differential)
entropy of f is defined by
For a random variable
on
with the density
f, we write the entropy of
by
.
The Fisher information for a differentiable density
f is defined by
When the random variable
on
has the differentiable density
f, we also write as
.
The important result for a behavior of the Fisher information on convolution (sum of independent random variables) is the Stam inequality, which was first stated by Stam in [
2] and subsequently proved by Blachman [
3],
where we have the equality if and only if
f and
g are Gaussian.
The importance of the Stam inequality can be found in its applications, for instance, the entropy power inequality [
2]; the logarithmic Sobolev inequality [
4]; Cercignani conjecture [
5]; the Shannon conjecture on entropy and the central limit theorem [
6,
7].
For
, we denote by
the convolution of
f with the
n-dimensional Gaussian density with mean vector
and covariance matrix
, where
is the identity matrix. Namely,
is the heat semigroup acting on
f and satisfies the partial differential equation
which is called
the heat equation. In this paper, we simply denote
by
and call it
the Gaussian perturbation of
f. Namely, letting
be the random variable on
with the density
f and
be an
n-dimensional Gaussian random variable independent of
with mean vector
and covariance matrix
, the Gaussian perturbation
stands the density function
of the independent sum
.
The remarkable relation between the entropy and the Fisher information can be established by a Gaussian perturbation (see, for instance, [
1], [
2] or [
8]);
which is known as
the de Bruijn identity.
Let f and g be probability measures on such that (f is absolutely continuous with respect to g). Setting the probability measure g as a reference, the relative entropy and the relative Fisher information can be introduced as follows:
Definition 1.2. The relative entropy of f with respect to g,
is defined by
which takes always a non-negative value.
We also define
the relative Fisher information of f with respect to g by
which is also non-negative. When random variables
and
have the densities
f and
g, respectively, the relative entropy and the relative Fisher information of
with respect to
are defined by
and
, respectively.
In view of the de Bruijn identity, one might expect that there is a similar connection between the relative entropy and the relative Fisher information. Indeed, the gradient formulas for the relative entropy functionals were obtained in [
9,
10,
11], where the reference measures would not be changed in their cases.
Recently Verdú in [
12], however, investigated the derivative in
t of
for two Gaussian perturbations
and
. Here we should note that the reference measure does move by the same time parameter in this case. The following identity of de Bruijn type
has been derived via MMSE in estimation theory (see also [
13], for general perturbations).
The main aim in this paper is that we shall give an alternative proof of this identity by direct calculation with integrations by part, the method of which is similar to ones in [
11,
14]. Moreover, it will be easily found that the above identity yields an integral representation of the relative entropy. We shall also see the simple proof of the logarithmic Sobolev inequality for centered Gaussian in univariate (
) case as an application of the integral representation.
2. An Integral Representation of the Relative Entropy
We shall make the Gaussian perturbations and , respectively, and consider the relative entropy , where the absolute continuity remains true for .
Here, we regard
as a function of
t and calculate the derivative,
by integrations by part with help of the heat equation.
Proposition 2.1. Let be probability measures on with finite Fisher informations and , and finite relative entropy . Then we obtainProof. First we should notice that the Fisher informations
and
are finite at any
. Because, for instance, if an
n-dimensional random variable
has the density
f and
is an
n-dimensional Gaussian random variable independent of
with mean vector
and covariance matrix
, then by applying the Stam inequality (1) to independent random variables
and
, we have that
where
is by simple calculation. We shall also notice that the function
is non-increasing in
t, that is, for
,
which can be found in [
15] (p. 101). Therefore,
is finite for
. But by a nonlinear approximation argument in [
11], we can impose a stronger assumption without loss of generality that
Concerning the first term in the most right hand side of (4), it follows immediately that
by the de Bruijn identity (3), hence, we shall concentrate our attention upon the second term.
Since the densities
and
satisfy the heat equation (2), the second term can be reformulated as follows:
In this reformulation, we have changed integration and differentiation at the first equality, which is justified by a routine argument with the bounded convergence theorem (see, for instance, [
16]).
Applying integration by part to the first term in the last expression of (8), it becomes
which can be asserted by the observation below. As
has finite Fisher information
,
has finite 2-norm in
and must be bounded at infinity. Furthermore, from our technical assumption (6),
is also bounded. Hence if we factorize as
then it can be found that
will vanish at infinity.
Applying integration by part to the second term in the last expression of (8), it becomes
Here it should be noted that
will vanish at infinity by the following observation. Similarly, we factorize it as
Then the boundedness of
comes from that
, and one of
is by the assumption (6) same as before. Furthermore, the limit formula
ensures that
will vanish at infinity.
Substitute the Equation (
9) and Equation (
10) into (
8), it follows that
Combining the Equation (
7) and Equation (
11), we have that
which means
Let and be n-dimensional random variables with the densities f and g, respectively, and be an n-dimensional Gaussian random variable independent of and with mean vector and covariance matrix .
Since the relative entropy is scale invariant, it follows that
We know that both of
and
, as
converge to
Z in distribution. Thus, we have
and the following integral representation for the relative entropy can be obtained:
Theorem 2.2. Let be probability measures with finite Fisher informations and finite relative entropy . Then we have the integral representation, 3. An Application to the Logarithmic Sobolev Inequality
In this section, we shall give a proof of the logarithmic Sobolev inequality for a centered Gaussian measure in case of
. Although several proofs of the logarithmic Sobolev inequality have already been given in many literatures (see, for instance, [
10,
17]), we shall give it here again as an application of the integral representation in Theorem 2.2.
Theorem 3.1.
Let g be the centered Gaussian measure of variance . Then for any probability measure f on of finite moment of order 2 with finite Fisher information , the following inequality holds:Proof. It is clear that the perturbed measure
is the centered Gaussian of variance
and the score of which is given by
Then using the Stein relation (see, for instance, [
15]), the relative Fisher information
can be expanded as follows:
As it was seen in (5), by Stam inequality, we have that
where we put
.
Since
f has finite moment of order 2, if we put the second moment of
f as
, then it is easy to see that the second moment of
is given by
Substitute (13) and (14) into (12) and we obtain that
Integrating for
, we have
Since
is dominated as
for
, it follows that
On the other hand, the relative Fisher information
can be given as
Combining (
15) and (
16), we have
which means our desired inequality by Theorem 2.2.
Remark 3.2. Similar way to the proof of Theorem 3.1 can be found in the paper by Stam [
2], where it is not for relative case. Namely, based on convolution inequalities and the de Bruijn identity, the isoperimetric inequality on entropy for a standardized random variable
X on
,
was shown. This inequality is essentially the same as the logarithmic Sobolev inequality for the standard Gaussian measure, where the left hand side in (
17) is the reciprocal of the entropy power.