1. Introduction
Probability measures on  treated in this paper are absolutely continuous with respect to the standard Lebesgue measure and we shall identify them with their densities.
For a probability measure 
f, the entropy 
 and the Fisher information 
 can be introduced, which play important roles in information theory, probability, and statistics. For more details on these subjects see the famous book [
1].
Hereafter, for an 
n-variables function 
 on 
, the integral of 
ϕ over the whole 
 by the standard Lebesgue measure 
 is abbreviated as
      
      that is, we shall leave out 
 in the integrand in order to simplify the expressions.
Definition 1.1. Let 
f be a probability measure on 
. Then 
the (
differential) 
entropy of f is defined by
      
      For a random variable 
 on 
 with the density 
f, we write the entropy of 
 by 
.
 The Fisher information for a differentiable density 
f is defined by
      
      When the random variable 
 on 
 has the differentiable density 
f, we also write as 
.
 The important result for a behavior of the Fisher information on convolution (sum of independent random variables) is the Stam inequality, which was first stated by Stam in [
2] and subsequently proved by Blachman [
3],
      
      where we have the equality if and only if 
f and 
g are Gaussian.
The importance of the Stam inequality can be found in its applications, for instance, the entropy power inequality [
2]; the logarithmic Sobolev inequality [
4]; Cercignani conjecture [
5]; the Shannon conjecture on entropy and the central limit theorem [
6,
7].
For 
, we denote by 
 the convolution of 
f with the 
n-dimensional Gaussian density with mean vector 
 and covariance matrix 
, where 
 is the identity matrix. Namely, 
 is the heat semigroup acting on 
f and satisfies the partial differential equation
      
      which is called 
the heat equation. In this paper, we simply denote 
 by 
 and call it 
the Gaussian perturbation of 
f. Namely, letting 
 be the random variable on 
 with the density 
f and 
 be an 
n-dimensional Gaussian random variable independent of 
 with mean vector 
 and covariance matrix 
, the Gaussian perturbation 
 stands the density function 
 of the independent sum 
.
The remarkable relation between the entropy and the Fisher information can be established by a Gaussian perturbation (see, for instance, [
1], [
2] or [
8]);
      
      which is known as 
the de Bruijn identity.
Let f and g be probability measures on  such that  (f is absolutely continuous with respect to g). Setting the probability measure g as a reference, the relative entropy and the relative Fisher information can be introduced as follows:
Definition 1.2. The relative entropy of f with respect to g, 
 is defined by
      
      which takes always a non-negative value.
 We also define 
the relative Fisher information of f with respect to g by
      
      which is also non-negative. When random variables 
 and 
 have the densities 
f and 
g, respectively, the relative entropy and the relative Fisher information of 
 with respect to 
 are defined by 
 and 
, respectively.
In view of the de Bruijn identity, one might expect that there is a similar connection between the relative entropy and the relative Fisher information. Indeed, the gradient formulas for the relative entropy functionals were obtained in [
9,
10,
11], where the reference measures would not be changed in their cases.
Recently Verdú in [
12], however, investigated the derivative in 
t of 
 for two Gaussian perturbations 
 and 
. Here we should note that the reference measure does move by the same time parameter in this case. The following identity of de Bruijn type
      
      has been derived via MMSE in estimation theory (see also [
13], for general perturbations).
The main aim in this paper is that we shall give an alternative proof of this identity by direct calculation with integrations by part, the method of which is similar to ones in [
11,
14]. Moreover, it will be easily found that the above identity yields an integral representation of the relative entropy. We shall also see the simple proof of the logarithmic Sobolev inequality for centered Gaussian in univariate (
) case as an application of the integral representation.
  2. An Integral Representation of the Relative Entropy
We shall make the Gaussian perturbations  and , respectively, and consider the relative entropy , where the absolute continuity  remains true for .
Here, we regard 
 as a function of 
t and calculate the derivative,
      
      by integrations by part with help of the heat equation.
Proposition 2.1. Let  be probability measures on  with finite Fisher informations  and , and finite relative entropy . Then we obtainProof. First we should notice that the Fisher informations 
 and 
 are finite at any 
. Because, for instance, if an 
n-dimensional random variable 
 has the density 
f and 
 is an 
n-dimensional Gaussian random variable independent of 
 with mean vector 
 and covariance matrix 
, then by applying the Stam inequality (1) to independent random variables 
 and 
, we have that
      
      where 
 is by simple calculation. We shall also notice that the function 
 is non-increasing in 
t, that is, for 
,
      
      which can be found in [
15] (p. 101). Therefore, 
 is finite for 
. But by a nonlinear approximation argument in [
11], we can impose a stronger assumption without loss of generality that
      
 Concerning the first term in the most right hand side of (4), it follows immediately that
      
      by the de Bruijn identity (3), hence, we shall concentrate our attention upon the second term.
Since the densities 
 and 
 satisfy the heat equation (2), the second term can be reformulated as follows: 
      In this reformulation, we have changed integration and differentiation at the first equality, which is justified by a routine argument with the bounded convergence theorem (see, for instance, [
16]).
Applying integration by part to the first term in the last expression of (8), it becomes
      
      which can be asserted by the observation below. As 
 has finite Fisher information 
, 
 has finite 2-norm in 
 and must be bounded at infinity. Furthermore, from our technical assumption (6), 
 is also bounded. Hence if we factorize as
      
      then it can be found that 
 will vanish at infinity.
Applying integration by part to the second term in the last expression of (8), it becomes
      
      Here it should be noted that 
 will vanish at infinity by the following observation. Similarly, we factorize it as
      
      Then the boundedness of 
 comes from that 
, and one of 
 is by the assumption (6) same as before. Furthermore, the limit formula 
 ensures that 
 will vanish at infinity.
Substitute the Equation (
9) and Equation (
10) into (
8), it follows that
      
Combining the Equation (
7) and Equation (
11), we have that
      
      which means
      
Let  and  be n-dimensional random variables with the densities f and g, respectively, and  be an n-dimensional Gaussian random variable independent of  and  with mean vector  and covariance matrix .
Since the relative entropy is scale invariant, it follows that
      
      We know that both of 
 and 
, as 
 converge to 
Z in distribution. Thus, we have
      
      and the following integral representation for the relative entropy can be obtained:
Theorem 2.2. Let  be probability measures with finite Fisher informations and finite relative entropy . Then we have the integral representation,   3. An Application to the Logarithmic Sobolev Inequality
In this section, we shall give a proof of the logarithmic Sobolev inequality for a centered Gaussian measure in case of 
. Although several proofs of the logarithmic Sobolev inequality have already been given in many literatures (see, for instance, [
10,
17]), we shall give it here again as an application of the integral representation in Theorem 2.2.
Theorem 3.1. 
Let g be the centered Gaussian measure of variance . Then for any probability measure f on  of finite moment of order 2 with finite Fisher information , the following inequality holds:Proof. It is clear that the perturbed measure 
 is the centered Gaussian of variance 
 and the score of which is given by
      
      Then using the Stein relation (see, for instance, [
15]), the relative Fisher information 
 can be expanded as follows:
      As it was seen in (5), by Stam inequality, we have that
      
      where we put 
.
 Since 
f has finite moment of order 2, if we put the second moment of 
f as 
, then it is easy to see that the second moment of 
 is given by
      
      Substitute (13) and (14) into (12) and we obtain that
      
      Integrating for 
, we have
      
      Since 
 is dominated as 
 for 
, it follows that
      
On the other hand, the relative Fisher information 
 can be given as
      
      Combining (
15) and (
16), we have
      
      which means our desired inequality by Theorem 2.2.
Remark 3.2. Similar way to the proof of Theorem 3.1 can be found in the paper by Stam [
2], where it is not for relative case. Namely, based on convolution inequalities and the de Bruijn identity, the isoperimetric inequality on entropy for a standardized random variable 
X on 
,
      
      was shown. This inequality is essentially the same as the logarithmic Sobolev inequality for the standard Gaussian measure, where the left hand side in (
17) is the reciprocal of the entropy power.