Next Article in Journal
Effect of Scanning Strategy During Selective Laser Melting on Surface Topography, Porosity, and Microstructure of Additively Manufactured Ti-6Al-4V
Next Article in Special Issue
Training of Deep Convolutional Neural Networks to Identify Critical Liver Alterations in Histopathology Image Samples
Previous Article in Journal
Coordinated Dispatch of Multi-Energy Microgrids and Distribution Network with a Flexible Structure
Previous Article in Special Issue
Transparent Redirections in Content Delivery Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Asymptotic Performances of a Signal-To-Noise Ratio Moment-Based Estimator for Real Sinusoids in Additive Noise

Department of Engineering, University of Campania “L. Vanvitelli”, 81031 Aversa (CE), Italy
Appl. Sci. 2019, 9(24), 5556; https://doi.org/10.3390/app9245556
Submission received: 11 October 2019 / Revised: 4 December 2019 / Accepted: 11 December 2019 / Published: 17 December 2019

Abstract

:
We considered the problem of the estimation of signal-to-noise ratio (SNR) with a real deterministic sinusoid with unknown frequency, phase and amplitude in additive Gaussian noise of unknown variance. A blind SNR estimator that does not require the knowledge of the instantaneous frequency of the sinusoid, through separate estimation of signal and noise power, was derived using the method of moments, a general method to derive estimators based on high-order moments. Statistical performances of the proposed estimators were studied theoretically through derivation of Cramer–Rao lower bounds (CRLBs) and asymptotic variances. Furthermore, results from Monte-Carlo simulations that confirm the validity of the theoretical analysis are presented along with some comments on the use of proposed estimators in practical applications.

1. Introduction

We consider the problem of estimation of the signal-to-noise ratio (SNR) when a deterministic real sinusoid with unknown parameters is corrupted by additive white Gaussian noise. This problem is encountered in different fields of engineering; for example, in telecommunications [1], in radar systems [2] and in smart-grid applications. Indeed, this study started from the need for accurate measurement of the SNR in a test-bed used for synchrophasor measurement before estimation of the parameters of the sinusoidal voltage [3,4].
The same problem has been studied in [5], wherein Papic et al. proposed an algorithm for SNR estimation based on autocorrelation and modified covariance methods for autoregressive spectral estimation and required the reliable estimate of the frequency of the sinusoid. The case of complex rather than real deterministic sinusoids in additive noise has been addressed in [6], wherein an SNR estimator derived with the method of moments [7] that makes use of second and fourth order moments was proposed ( M 2 M 4 -estimator). It does not require the complex sinusoid frequency estimate, and for this reason, it belongs to the class of blind SNR estimators and may find applications in all those cases where the frequency estimate may not be available or frequency may be varying over the observation interval.
In recent years, moment-based SNR estimators have attracted the interest of researchers in the telecommunications field where second and fourth order moments based estimators have been employed for estimation of the SNR of modulated signals in additive white Gaussian noise and flat-fading channels. In [1,8], for example, the SNR M 2 M 4 -estimator was proposed for phase-shift keying (PSK) modulated signals as a blind estimator that does not require the knowledge of transmitted data. In [9] more general moments-based SNR estimators were derived for multi-level constellations, such as generic quadrature amplitude modulation (QAM) constellations, and the SNR M 2 M 4 -estimator was derived as a special case of moment-based estimators that make use of higher-order moments. In [10] an estimator that uses also the sixth-order moment of the received data for non-constant modulus constellations was proposed in order to improve performances. Further extensions of the method for multilevel constellations were proposed in [11]. In [9,10,11] performances of the proposed estimators were theoretically studied and results based on asymptotic analyses were reported.
In this paper we propose a moment-based SNR estimator for a deterministic real sinusoid in additive noise that is formed by the ratio of the estimators of the signal power and noise power. The proposed estimators belong to the class of the second-and-fourth-order moments estimators that can be obtained by application of the method of moments [7] and do not require the reliable frequency estimate, as in [5]. We derived a new general formula for even-order moments of the observed signal that is useful not only to provide the second and fourth moments required by the method of moments, but also for the study of statistical performances of the proposed estimators. Since asymptotic analyses carried out in [9,10,11] assume random signals in random noise and cannot immediately be extended to deterministic signals, we also derived a new general expression for covariance matrices useful for asymptotic analysis of performances of moment-based estimators when the signal corrupted by the noise is deterministic. Results from asymptotic analyses and Monte-Carlo simulations are presented along with the corresponding Cramer–Rao lower bounds (CRLB), and we reveal that the proposed estimators have good performances in a wide range of operating conditions. In same cases performances are very close to the optimum. This paper extends some of the results presented in [12] by providing new general formulas for any even-order moments and covariance matrices of the vectors of sample moments, and by studying statistical performances of the proposed estimators with theoretical asymptotic analyses that confirm the numerical results obtained with Monte-Carlo simulations.
The paper is organized as follows: in Section 2 we derive the estimators starting from a general result on the even-order moment for the given signal model; in Section 3 we derive the Cramer–Rao lower bounds for all proposed estimators; in Section 4 the asymptotic performances of the proposed estimators are derived; in Section 5 results from Monte-Carlo simulations are presented along with comparisons with the CRLBs and the asymptotic performances; finally, in Section 6 we draw the conclusions.

2. Moment-Based Estimators

We assume that samples from a real sinusoidal signal in additive noise are available over an observation window of size K, i.e.,
y k = A x k + w k , k = 0 , , K 1 ,
where x k = cos 2 π ν k + ϕ ; the amplitude A, the frequency ν and phase ϕ of the sinusoid are unknown; and the additive Gaussian noise w k is a random variable with zero-mean and unknown variance σ 2 ; i.e., w k N ( 0 ; σ 2 ) . The signal-to-noise ratio (SNR), defined as ρ A 2 / 2 σ 2 , is then unknown and represents the parameter we wish to estimate when the other parameters of the sinusoid, frequency and phase, remain unknown. To that purpose we derive an estimator based on the method of moments.
The method of moments is a well known statistical method used to derive estimators as functions of sample high-order moments [7]. The key idea is to express the parameter to be estimated as a function of the moments of y k of a different order—i.e., μ m = E y k m , where m is an integer—and use the natural estimators of the moments μ ^ m ( 1 / K ) k = 0 K 1 y k m in place of the true moments μ m to get an estimate of the parameter from the observed samples. In many cases, such estimators are computationally not intensive and present good statistical performances. Though, in general, the optimality of an estimator derived with this method cannot be guaranteed, in many cases the estimator turns out to be consistent. The method relies on the possibility of expressing moments from the observed random signal as functions of the parameters to be estimated. However, when in the observed signal model, one term is deterministic, as is our case, so the derivation of the moments might be cumbersome, as shown, for example, in [6]. For these cases, Kay suggests to assume a random nature for the "deterministic" sinusoid by assuming a random phase, and in this paper we follow such an approach [7] (p. 299).
To derive moment-based estimators of the signal power S = A 2 / 2 , noise power σ 2 and SNR ρ we use the second and fourth order moment, and therefore, we need to compute μ 2 and μ 4 . Instead of deriving each moment separately, as done in [12], we show here that a general formula expressing any even-order moment of the signal (1) exists that is valid asymptotically as the number of observed samples grows to infinity. Equivalently, in the signal model (1), where the sinusoidal signal is deterministic, we assume that the sinusoidal signal’s phase is a uniform random variable (r.v.), i.e., ϕ U 0 , 2 π , independent of the noise w k . Consequently, we can treat the sequence y k as the realization of a wide sense stationary (WSS) random sequence for which it is possible to calculate statistical moments as, under the given assumption,
1 K k = 0 K 1 x k 2 m E x k 2 m ,
for K .
The general formula for even-order moments for the signal model (1) is given by the following expression.
μ 2 m = n = 0 m 2 m 2 n 2 m 2 n 1 ! ! 2 n 1 ! ! 2 n ! ! A 2 n σ 2 m n ,
where ! ! denotes the double factorial operator and its derivation is given in Appendix A.
From (3) we can immediately get the second and fourth moments
μ 2 = A 2 2 + σ 2
μ 4 = 3 8 A 4 + 3 A 2 σ 2 + 3 σ 4 .
Let S = A 2 / 2 and N = σ 2 ; then, from (4) and (5) we obtain the estimators for S and N as
S ^ = 2 μ ^ 2 2 2 3 μ ^ 4
N ^ = μ ^ 2 S ^ ,
where, as required by the method of moments, instead of the true moments we use their corresponding natural estimators
μ ^ 2 = 1 K k = 0 K 1 y k 2
μ ^ 4 = 1 K k = 0 K 1 y k 4 .
The SNR estimator is immediately built as the ratio of the two estimators derived above; i.e.,
ρ ^ = S ^ N ^ .
In the following sections performances of the proposed estimators are studied both theoretically and numerically in terms of squared bias and variance. For a generic estimator θ ^ the squared bias is defined as [7,13]
b 2 ( θ ^ ) [ E [ θ ^ ] θ ] 2 ,
while the variance is defined as
var ( θ ^ ) = E [ | θ ^ E [ θ ^ ] | 2 ] .

3. Cramer–Rao Lower Bounds

Before deriving the performances of the proposed estimators, we derive the corresponding Cramer–Rao lower bounds (CRLB), as they represent the theoretical limit on the best achievable accuracy that can be obtained. First, we derive the CRLB for S ^ and N ^ and then the CRLB for the estimator of SNR ρ ^ .
We start from the joint probability density function
f y 0 , , y K 1 | S , N = 1 2 π σ 2 K / 2 exp 1 2 σ 2 k = 0 K 1 y k A x k 2 ,
where S = A 2 / 2 and N = σ 2 . To get the bound we need to first derive the Fisher information matrix that is defined as
I S , N E 2 ln f S 2 E 2 ln f S N E 2 ln f S N E 2 ln f N 2 .
In Appendix B we show that in this case, the Fisher matrix is given by
I S , N = K 2 A 2 σ 2 0 0 K σ 4 ,
and the Cramer–Rao lower bounds for the estimators S ^ and N ^ , obtained by inversion of the Fisher matrix, are
var ( S ^ ) 2 A 2 σ 2 K
var ( N ^ ) σ 4 K .
Based on the above results it is now possible to derive the CRLB for the proposed SNR estimator ρ ^ . Since the estimator is formed by the ratio of S ^ and N ^ , the bound for ρ ^ can be derived from the Fisher information matrix obtained previously and the derivatives of the transformation
g ( S , N ) = S N
according the following relationship [7]
var ( ρ ^ ) g S g N I 1 S , N g S g N T .
Since
g S g N T = 1 N S N 2 ,
we get
var ( ρ ^ ) A 2 σ 2 K 2 + A 2 4 σ 2 .
Similarly, we can derive the bound when the SNR is expressed in dB. In this case the estimator is given by the following transformation
g S , N = 10 log 10 S N .
The derivatives are
g S g N = 10 ln 10 1 S 1 N
= 10 ln 10 2 A 2 1 σ 2 ,
and we have that the lower bound is
var ( ρ ^ ) 10 ln 10 2 8 K σ 2 A 2 + 1 K dB 2 .

4. Asymptotic Performances

It is only possible to study the analytical performances of the proposed estimators asymptotically, i.e., for K that tends to infinity. In this section we derive the asymptotic variances of all proposed estimators and we will show in the next section that, though the analysis is only valid for K , in some cases we have found that the analytical results may predict performances for sufficiently high K as well. Unfortunately, it is not possible to predict the number of samples required to get the asymptotic results and we need to run Monte-Carlo simulations to have an idea of how large the number of samples needs to be. The mathematical method that we apply to derive the asymptotic analytical performances is outlined, for example, in [7], and makes it possible to overcome the mathematical intractability of the method that uses the transformation of random variables to get the pdf of the estimator.
The variance of the estimator derived with the method of moments can be obtained from the first-order Taylor expansion of g ( T ) about the point T = E T = μ , where T is the vector of the sample moments and g T is the function that maps the sample moments to the estimate. Under the first-order approximation the mean of the estimator is equal to g μ and by simple substitution using (4) and (5) it can be shown that the proposed estimators are asymptotically unbiased. The first-order approximation of the variance is given by
var ρ ^ g T T = μ T C T g T T = μ ,
where C T is the covariance matrix of T ; i.e.,
C T = E T μ T μ T .
Hence, we need to derive the covariance matrix under the assumption that in the signal model (1) x k is a deterministic signal. In Appendix C the covariance matrix is derived and it is shown that we cannot make the assumption suggested by Kay on the deterministic signal x k , i.e., assume that such signal is random by assuming a random initial phase, and that the covariance matrix used to get asymptotic results in similar studies [9,10,11] cannot be used here.
Given the signal model (1) and the vector of even-order sample moments T , each element of the covariance matrix C T is
C T i j = 1 K M i + j D i , j ,
where
M m = p = 0 m 2 m 2 p 2 p 1 ! ! 2 p ! ! 2 m 2 p 1 ! ! A 2 p σ 2 m p
and
D i , j = p = 0 i q = 0 j 2 i 2 p 2 j 2 q 2 i + j p q 1 ! ! 2 i + j p q ! ! 2 p 1 ! ! 2 q 1 ! ! A 2 i + j p q σ 2 p + q .
Note that the same covariance matrix can be used to study the asymptotic performances of all proposed estimators as the only difference is in the way the function g ( T ) is defined in each case. Furthermore, such a covariance matrix is useful to evaluate the performances of any moment based estimator given the signal model (1).
To derive the asymptotic variance of all proposed estimators we also need the partial derivatives defined in (24). As we use the second and fourth moment of y k we only need to consider the two sample moments T 2 and T 4 . The derivatives of the function g T 2 , T 4 are easily computed. First, consider the estimator S ^ , that is defined by the following function g T
g T = 2 T 2 2 2 3 T 4 = g T 2 , T 4 .
Hence, the partial derivatives are
g T 2 = 2 T 2 g T 2 , T 4
and
g T 4 = 1 3 1 g T 2 , T 4 .
The above results are useful for the estimator of the noise power N ^ defined by the function h T 2 , T 4 , as follows.
h T 2 , T 4 = T 2 g T 2 , T 4 .
The corresponding derivatives are
h T 2 = 1 g T 2
h T 4 = g T 4 .
Finally, the derivatives of the function f T 2 , T 4 = g T 2 , T 4 / h T 2 , T 4 , that define the estimator of the SNR ρ ^ , are
f T 2 = g T 2 1 h g h T 2 1 h 2
f T 4 = g T 4 1 h g h T 4 1 h 2 .
The general expression of the covariance matrix shows that the proposed estimators are consistent. However, it is difficult to know in advance when the asymptotic behavior describes the actual performances of the estimators well.

5. Numerical Examples

We present in this section some examples of the performances of the proposed moment-based estimators for S and N and signal-to-noise ratio ρ in terms of squared bias and variance. We report estimated squared biases and variances obtained through Monte-Carlo simulations and we will find that such numerical results confirm the asymptotic performances derived in Section 4. We also compare them to the corresponding CRLBs derived in Section 3 and we make some comments on the property of efficiency of the proposed estimators, i.e., whether they achieve (asymptotically) the corresponding CRLB, and on the property of consistency, i.e., whether the variance can be made to approach zero as the number of observations K [7,13].
For all proposed estimators we have estimated the squared bias (11) and variance (12) by averaging over L = 2 13 = 8192 estimates on a varying number of observations K. The mean of each estimator is estimated by means of a standard sample mean estimator; e.g., for S ^
E [ S ^ ] μ ^ S = 1 L l = 1 L S ^ l ,
where S ^ l is a single estimate of S with K observations. The squared bias is then obtained from (11), where the estimated mean is used in place of the true mean; e.g., b 2 ( S ^ ) = [ E [ S ^ ] S ] 2 . The variance of estimators is estimated by means of the standard sample variance estimator; e.g., for S ^
var ( S ^ ) σ ^ S 2 = 1 L 1 l = 1 L ( S ^ l μ ^ S ) 2 .
Estimation of squared bias and variance of estimators has been performed at different SNRs that, without loss of generality, are obtained by varying the power of the real sinusoid S while maintaining fixed noise power N = 1 . The frequency and phase of the deterministic sinusoid were chosen randomly at the beginning of each simulation.

5.1. Performances of S ^

First, we verified that S ^ is asymptotically unbiased, as expected from first-order approximation analysis. Results of simulations run at different SNRs show that the estimated squared bias is very small even with a small number of observations. As an example, we report in Figure 1 the estimated squared bias for S ^ at different SNRs obtained by increasing S and with N = 1 . Though the squared bias grows with increasing S, and then with the SNR in this case, it can be made quickly very small by increasing the number of observed samples. Some bias is noticeable at high SNR; e.g., for 76.9897 dB with a low number of samples.
In Figure 2 we show the results of estimated variance of the estimator S ^ at different SNRs, denoted with (sim), and we compare the results with the corresponding CRLBs that we have derived in Section 3 and asymptotic variances, denoted with (theor), derived in Section 4. Variances are reported on a logarithmic scale as 10 log 10 var versus the log 2 K , where K is the number of observed samples.
We note that estimated variances reach the corresponding asymptotic performances in the whole range of SNRs under consideration, confirming the validity of the asymptotic performances derivation. However, the number of samples required to get the asymptotic performances depends on the current value of S: for low values of S, the estimator reaches quickly the asymptotic variance, while for high values of S the number of required sample results to be quite large. Unfortunately, it is not possible to predict when the asymptotic result is valid as the assumption made in performance derivation is to have a large number of samples.
Plots in Figure 2 show also how the CRLBs for S ^ depend on the current value of S and become very high as S increases. In general, the CRLB for S ^ (16) depends on the product S N and then for fixed N it becomes larger as S increases; i.e., the best accuracy that can be obtained worsens with increasing signal power. On the other hand, the asymptotic variance depends on both S and N through terms of the type A 2 k σ 2 m , with k and m non negative integers, that are present in both the derivatives of the function that maps the moments to the estimate, i.e., g μ 2 , μ 4 , as defined in (29), and its derivatives (30) and (31), and the covariance matrix (26).
In Figure 2a,b we see that for SNRs equal to 3 . 0103 dB and 3 . 0103 dB the estimator S ^ is not asymptotically efficient though it is consistent, as the asymptotic variance and the corresponding CRLB are not superimposed. Note that the difference between the asymptotic variance and the CRLB, that cannot be filled by simply increasing the number of observations, can be calculated by means of the results derived in Section 3 and Section 4.
On the other hand, for higher values of S with N = 1 , the asymptotic variance is superimposed to the corresponding CRLB, showing that for a large range of SNRs under consideration the estimator is efficient and that such efficiency is reached in many cases with a relatively small number of samples. However, as S increases the bound is reached for an increasing number of observed samples. For example, while with an SNR of 50.9691 dB we need at least 2 16 samples to get the asymptotic behavior; at 76.9897 dB the estimated variance reaches the corresponding asymptotic variance with at least 2 22 samples. For even higher S we expect that the number of samples required to get asymptotic performances to increase exponentially.

5.2. Performances of N ^

We verified that the estimator N ^ is asymptotically unbiased. In fact, Figure 3 shows that, for most of the SNRs taken into consideration obtained by changing S with N = 1 , the estimator is practically unbiased even for small number of samples. For high SNRs the number of samples must be increased beyond 2 22 to get acceptable small squared bias; e.g., less than 10 2 .
In Figure 4 the estimated variances, denoted with (sim), along with the corresponding asymptotic variances, denoted with (theor), and CRLBs, at different SNRs are shown. Note that the CRLB of N ^ (17) depends on N 2 only and in the figures remains constant as, also in this case, the increase of the SNR is obtained by increasing S while the noise power remains constant with N = 1 . Variances are reported on a logarithmic scale as 10 log 10 var versus the log 2 K , where K is the number of observed samples.
Results show that the estimated variance achieves the asymptotic performances for the whole range of SNRs into consideration. As for the case of the signal power estimation, it is not possible to predict the number of samples from which the asymptotic variance expresses the exact variance because the asymptotic analysis presented in Section 4 is based on the fundamental assumption that K .
While the noise power estimator is asymptotically consistent in all cases, it is not efficient, since it cannot achieve the CRLB, in all cases under investigation. The difference between the asymptotic variance and the CRLB remains constant as the number of samples increases and depends on the current value of the signal power S. In Figure 4b we note that the difference decreases as S increases; e.g., for a given K the asymptotic variance of N ^ with an SNR of 3.0103 dB is closer to the CRLB than the asymptotic variance of N ^ at 0 dB. On the other hand, in Figure 4c–f we see that the difference becomes lager and larger and asymptotic variances of the noise power estimator are quite faraway from the CRLB. In other words, though the estimator N ^ is consistent, it may require too many samples to get acceptable performances. Moreover, it is not asymptotically efficient since the asymptotic variance is not superimposed to the corresponding CRLB.

5.3. Performances of ρ ^

In Figure 5 we show the estimated squared bias of the SNR estimator ρ ^ as a function of the number of observations K at different SNRs, obtained by changing S with N = 1 . Results reveal that, in certain cases, the SNR estimator exhibits some bias that depends on both S and N to be estimated and the number of samples observed over which the estimate of the SNR is computed. In fact, for a sufficiently large number of observations, the squared bias can be made very small, even if that means that for high values of S the number of observations required for an unbiased estimator is very large; e.g., with an SNR of 76.9897 dB the squared bias is still greater than one with 2 22 observations.
Estimated variances, denoted with (sim), at different SNRs as functions of the number of observations are shown in Figure 6 where asymptotic variances, denoted with (theor) and corresponding CRLBs are also shown.
In Figure 6a,b we show results for SNRs in the range from −3.0103 dB to 3.0103 dB obtained by varying S with N = 1 . Within this range the SNR estimator becomes quickly unbiased and the asymptotic performances are reached for small sizes of observation windows. The SNR estimator is consistent but not efficient, as the estimator’s variance and the corresponding CRLB are not superimposed. We note that such a gap becomes smaller as S increases.
In Figure 6c,d we show results for SNRs within the range from 10.9691 dB to 36.9897 dB. The estimated variances reach the corresponding asymptotic variances, though a greater number of samples is needed. In this case we see that the difference between the estimator’s variance and the corresponding CRLBs increases as S increases. Note that as the SNR increases the dominant term in (23) becomes
10 ln 10 2 1 K
that does not depend on the SNR to be estimated, while estimator’s asymptotic variances given, in general, by (24), and plotted in Figure 6, still depend on S and N.
Performances degrades as S increases with N = 1 . In Figure 6e,f we consider increasing SNRs from 44.9485 dB to 76.9897 dB, a range for which the estimator is biased even for relatively large number of samples. The estimated variances reach the asymptotic performances only when the number of samples is high enough and the SNR estimator becomes substantially unbiased. For example, when the SNR is equal to 76.9897 dB and N = 1 we need more than 2 22 samples to get the asymptotic variance. Most importantly, for higher values of S the asymptotic performances show that to get acceptable accuracies the number of samples must be increased even further.

6. Conclusions

It is possible to obtain with the method of moments, blind estimators for signal power, noise power and signal-to-noise ratio when a deterministic real sinusoid is corrupted by additive white Gaussian noise. We have derived new estimators by providing a general formula for the moments of the observed signal of any even order that is valid for a sufficiently large number of samples observed.
The performances of the proposed estimators have been studied through analytic derivation of asymptotic variances and Cramer–Rao lower bounds and confirmed through Monte-Carlo simulations. A general expression for the covariance matrix of the even-order sample moments has also been derived to get the asymptotic performances of the proposed estimators. Such an expression is valid asymptotically and can be useful to evaluate performances of any moment-based estimator for the same signal model, when a deterministic real sinusoid is corrupted by additive noise.
Analytical and numerical results show the estimator of signal power S ^ results to be asymptotically efficient and consistent for a large range of SNR values obtained by varying the signal power S and with N = 1 . On the other hand, at low SNRs and with N = 1 , below approximately 10 dB, the estimator S ^ is not efficient, though still consistent. Similar results have been obtained with the noise power estimator N ^ . The SNR estimator results were consistent in all the cases we studied. However, it is not efficient and we observed a performance degradation with respect to the corresponding CRLB that becomes larger as the signal power S increases and N = 1 .
As a general consideration, we found that the performances can be acceptable in a relatively wide range of SNRs. However, since we have found that there exists a strong dependence between S and N to be estimated and the number of observations, special care in the choice of K is required when this estimator is used in practice to get acceptable performances. Numerical examples show that in some cases the SNR estimator has very good performances and then it can be used to obtain an accuracy very close to the CRLB. The asymptotic results presented in this paper represent a useful mathematical tool that may provide useful guidance to expected performances before any practical implementation, letting any practitioner assess the performances, and then, usefulness, of the proposed estimators based on his application needs and operating conditions.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

First, we derive a general result on the moments of y k when in the signal model (1) x k is assumed as a generic random signal. Then, we provide the general formula for the moments of y k when x k is a deterministic sinusoidal signal.
We start from the definition of even-order moments
μ 2 m = E y k 2 m = E A x k + w k 2 m .
The application of the binomial theorem leads to
μ 2 m = E n = 0 2 m 2 m n A n x k n w k 2 m n
= n = 0 2 m 2 m n A n E x k n E w k 2 m n .
Since the odd-order moments of w k are zero, we have that only terms corresponding to even-order moments of w k remain in the summation
μ 2 m = n = 0 m 2 m 2 n A 2 n E x k 2 n E w k 2 m n .
The even-order moments of a real Gaussian random variable are
E w k 2 n = 2 n 1 ! ! σ 2 n , n = 1 , 2 ,
and then, we obtain
μ 2 m = n = 0 m 2 m 2 n 2 m 2 n 1 ! ! A 2 n E x k 2 n σ 2 m n .
In the case of real sinusoids, under the assumption of uniform random initial phase in ( 0 , 2 π ) , by definition the even-order moment of x k is
E x k 2 n = 1 2 π 0 2 π cos 2 π ν k + ϕ 2 n d ϕ .
By symmetry the integral above can also be written as
0 2 π cos 2 n θ d θ = 4 0 π / 2 cos 2 n θ d θ ,
where the integral on the interval ( 0 , π / 2 ) is given by
0 π / 2 cos 2 n θ d θ = π 2 2 n 1 ! ! 2 n ! ! .
Therefore,
1 2 π 0 2 π cos 2 n θ d θ = 2 n 1 ! ! 2 n ! ! .
Putting (A10) in (A6) we finally obtain the general formula (3) for even-order moments of the signal (1).

Appendix B

In this appendix we derive the Fisher information matrix needed to obtain the CRLBs of proposed estimators. We need to calculate the expectations of the first and second partial derivatives of the natural logarithm of the following conditional pdf
f y 0 , , y K 1 | S , N = 1 2 π N K / 2 exp 1 2 N k = 0 K 1 y k 2 S x k 2 ,
where S = A 2 / 2 and N = σ 2 , to get each element of thee matrix.
We start from the derivative of ln f with respect to S. We have
ln f S = 1 2 N k = 0 K 1 S y k 2 2 2 S x k y k + 2 S x k 2 .
= 1 2 N k = 0 K 1 2 S x k y k + 2 x k 2 .
Differentiating again with respect to S, we get
2 ln f S 2 = 1 2 N k = 0 K 1 2 1 2 S 3 x k y k
= 1 2 σ 2 k = 0 K 1 2 2 2 2 A 3 x k y k
= 1 A 3 σ 2 k = 0 K 1 x k y k .
Since E [ y k ] = A x k , the corresponding element in the Fisher matrix is given by
E 2 ln f S 2 = 1 A 3 σ 2 k = 0 K 1 A x k 2 .
Now, consider the term
1 K k = 0 K 1 x k 2 = 1 K k = 0 K 1 cos 2 2 π ν k + ϕ .
Under the assumption that the frequency ν is not close to 0 or 1/2 and for K large enough, we can write [7] (p. 33)
1 K k = 0 K 1 cos 4 π ν k + ϕ 0 .
Hence,
1 K k = 0 K 1 x k 2 1 2
and the first element of Fisher matrix is then given by
E 2 ln f S 2 = K 2 A 2 σ 2 .
Similarly, we obtain the first and second partial derivatives with respect to N. The first derivative is given by
ln f N = N K 2 ln 2 π N N 1 2 N k = 0 K 1 y k A x k 2 = K 2 N + 1 2 N 2 k = 0 K 1 y k A x k 2 .
Differentiating again with respect to N we get
2 ln f N 2 = K 2 N 2 1 N 3 k = 0 K 1 y k A x k 2 .
Hence, taking the expectation
E 2 ln f N 2 = K 2 N 2 + 1 N 3 E k = 0 K 1 y k A x k 2 .
Consider, now, the term
E k = 0 K 1 y k A x k 2 = k = 0 K 1 E y k 2 2 A x k E y k + A 2 x k 2 .
Since E y k 2 = A 2 x k 2 + N and E y k = A x k , the above term can be rewritten as
E k = 0 K 1 y k A x k 2 = k = 0 K 1 A 2 x k 2 + N 2 A 2 x k 2 + A 2 x k 2 = K N .
The second derivative is then
E 2 ln f N 2 = K 2 N 2 + K N 2 = K σ 4 .
The off-diagonal terms of the Fisher matrix are both equal to zero. In detail, we have
2 ln f S N = 2 ln f N S = 1 σ 4 k = 0 K 1 x k y k A x k 2 .
By taking the expectation, we immediately obtain that
E 2 ln f S N = 0 .

Appendix C

In this appendix we derive (26). Let us start from the definition of the covariance matrix. Each element of the covariance matrix is then given by
C T i j = cov T 2 i , T 2 j
= E 1 K k = 0 K 1 y k 2 i 1 K k = 0 K 1 y k 2 j E 1 K k = 0 K 1 y k 2 i E 1 K k = 0 K 1 y k 2 j .
Now, consider the first term of the right-hand side. After some manipulation we get
E 1 K k = 0 K 1 y k 2 i 1 K k = 0 K 1 y k 2 j = 1 K 2 k = 0 K 1 E y k 2 i + 2 j ,
where we have used the independence assumption of the Gaussian noise. Similarly, we obtain that the second term on the right-hand side of (A30) is
E 1 K k = 0 K 1 y k 2 i E 1 K k = 0 K 1 y k 2 j = 1 K 2 k = 0 K 1 E y k 2 i E y k 2 j .
Putting all together, the covariance (A29) can written as
cov T 2 i , T 2 j = 1 K k = 0 K 1 1 K E y k 2 i + j 1 K E y k 2 i E y k 2 j .
By using (3) we can explicitly write the first term as
1 K k = 0 K 1 E y k 2 i + j = p = 0 i + j 2 i + j 2 p A 2 p 1 K k = 0 K 1 x k 2 p 2 ( i + j ) 2 p 1 ! ! σ 2 j + j p .
Since the term
1 K k = 0 K 1 x k 2 p ,
as K grows to infinity, it is asymptotically equal to the moment of order 2 p of x k when the initial phase ϕ is random. Then, using (A10) we have that asymptotically
1 K k = 0 K 1 E y k 2 i + j M i + j = p = 0 i + j 2 i + j 2 p A 2 p 2 p 1 ! ! 2 p ! ! 2 ( i + j ) 2 p 1 ! ! σ 2 j + j p .
The remaining term is written as
1 K k = 0 K 1 E y k 2 i E y k 2 j = 1 K k = 0 K 1 p = 0 i 2 i 2 p A 2 i p x k 2 i p E w k 2 p q = 0 j 2 j 2 q A 2 j q x k 2 j q E w k 2 q = p = 0 i q = 0 j 2 i 2 p 2 j 2 q A 2 i + j p q 1 K k = 0 K 1 x k 2 i + j p q E w k 2 p E w k 2 q .
Now, asymptotically we have
1 K k = 0 K 1 x k 2 i + j p q E x k 2 i + j p q = 2 i + j p q 1 ! ! 2 i + j p q ! ! .
Hence, the term D i , j is
p = 0 i q = 0 j 2 i 2 p 2 j 2 q A 2 i + j p q 2 i + j p q 1 ! ! 2 i + j p q ! ! 2 p 1 ! ! 2 q 1 ! ! σ 2 p + q .

References

  1. Pauluzzi, D.R.; Beaulieu, N.C. A comparison of SNR estimation techniques for the AWGN channel. IEEE Trans. Commun. 2000, 48, 1681–1691. [Google Scholar] [CrossRef]
  2. Papic, V.D.; Djurovic, Z.M.; Kovacevic, B.D. Adaptive Doppler-Kalman filter for radar systems. Image Signal Process. IEE Proc.-Vis. 2006, 153, 379–387. [Google Scholar] [CrossRef]
  3. Cuccaro, P.; Gallo, D.; Landi, C.; Luiso, M.; Romano, G. Recursive phasor estimation algorithm for synchrophasor measurement. In Proceedings of the IEEE International Workshop Applied Measurements for Power Systems (AMPS), Aachen, Germany, 23–25 September 2015; pp. 90–95. [Google Scholar] [CrossRef]
  4. Cuccaro, P.; Gallo, D.; Landi, C.; Luiso, M.; Romano, G. Phase-based estimation of synchrophasors. In Proceedings of the IEEE International Workshop Applied Measurements for Power Systems (AMPS), Aachen, Germany, 28–30 September 2016; pp. 1–6. [Google Scholar] [CrossRef]
  5. Papic, V.; Djurovic, Z.; Kvascev, G.; Tadic, P. On signal-to-noise ratio estimation. In Proceedings of the Melecon 2010—2010 15th IEEE Mediterranean Electrotechnical Conference IEEE, Valletta, Malta, 25–28 April 2010; Volume 5, pp. 160–165. [Google Scholar] [CrossRef]
  6. Sekhar, S.C.; Sreenivas, T.V. Signal-to-noise ratio estimation using higher-order moments. Signal Process. 2006, 86, 716–732. [Google Scholar] [CrossRef]
  7. Kay, S. Fundamentals of Statistical Signal Processing; Prentice Hall PTR: Upper Saddle River, NJ, USA, 1993; Volume 1. [Google Scholar]
  8. Matzner, R. An SNR estimation algorithm for complex baseband signals using higher order statistics. Facta Univ. 1993, 6, 41–52. [Google Scholar]
  9. Ping, G.; Tepedelenlioglu, C. SNR estimation for nonconstant modulus constellations. IEEE Trans. Signal Process. 2005, 53, 865–870. [Google Scholar] [CrossRef]
  10. López-Valcarce, R.; Mosquera, C. Sixth-order statistics-based non-data-aided SNR estimation. IEEE Commun. Lett. 2007, 11, 351–353. [Google Scholar] [CrossRef]
  11. Álvarez-Díaz, M.; López-Valcarce, R.; Mosquera, C. SNR estimation for multilevel constellations using higher-order moments. IEEE Trans. Signal Process. 2010, 58, 1515–1526. [Google Scholar] [CrossRef]
  12. Romano, G. Blind Signal-to-Noise Ratio Estimation of Real Sinusoid in Additive Noise. In Proceedings of the 42nd International Conference Telecommunications and Signal Processing (TSP), Budapest, Hungary, 1–3 July 2019; pp. 224–227. [Google Scholar] [CrossRef]
  13. Manolakis, D.G.; Ingle, V.K.; Kogon, S.M. Statistical and Adaptive Signal Processing: Spectral Estimation, Signal Modeling, Adaptive Filtering, and Array Processing; Artech House: Norwood, MA, USA, 2005. [Google Scholar]
Figure 1. Estimated squared bias b 2 ( S ) ^ of signal power estimator S ^ versus the number of observations K for different values of signal-to-noise ratio (SNR) obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows estimated squared biases for different SNRs as reported in the legend.
Figure 1. Estimated squared bias b 2 ( S ) ^ of signal power estimator S ^ versus the number of observations K for different values of signal-to-noise ratio (SNR) obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows estimated squared biases for different SNRs as reported in the legend.
Applsci 09 05556 g001
Figure 2. Asymptotic and estimated variances of the signal power estimator S ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows asymptotic (thvar) and estimated (sim) variances for different SNRs along with corresponding CRLBs.
Figure 2. Asymptotic and estimated variances of the signal power estimator S ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows asymptotic (thvar) and estimated (sim) variances for different SNRs along with corresponding CRLBs.
Applsci 09 05556 g002
Figure 3. Estimated squared bias b 2 ( N ) ^ of the noise power estimator N ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows estimated squared biases for different SNRs as reported in the legend.
Figure 3. Estimated squared bias b 2 ( N ) ^ of the noise power estimator N ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows estimated squared biases for different SNRs as reported in the legend.
Applsci 09 05556 g003
Figure 4. Asymptotic and estimated variances of the noise power estimator N ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows asymptotic (thvar) and estimated (sim) variances for different SNRs along with CRLBs.
Figure 4. Asymptotic and estimated variances of the noise power estimator N ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows asymptotic (thvar) and estimated (sim) variances for different SNRs along with CRLBs.
Applsci 09 05556 g004
Figure 5. Estimated squared bias b 2 ( ρ ^ ) of the SNR estimator ρ ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows estimated squared biases for different SNRs as reported in the legend.
Figure 5. Estimated squared bias b 2 ( ρ ^ ) of the SNR estimator ρ ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows estimated squared biases for different SNRs as reported in the legend.
Applsci 09 05556 g005
Figure 6. Asymptotic and estimated variances of the SNR estimator ρ ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows asymptotic (thvar) and estimated (sim) variances for different SNRs along with corresponding CRLBs.
Figure 6. Asymptotic and estimated variances of the SNR estimator ρ ^ versus the number of observations K for different values of SNR obtained by changing the signal power S and with noise power N = 1 . Each subfigure shows asymptotic (thvar) and estimated (sim) variances for different SNRs along with corresponding CRLBs.
Applsci 09 05556 g006

Share and Cite

MDPI and ACS Style

Romano, G. Asymptotic Performances of a Signal-To-Noise Ratio Moment-Based Estimator for Real Sinusoids in Additive Noise. Appl. Sci. 2019, 9, 5556. https://doi.org/10.3390/app9245556

AMA Style

Romano G. Asymptotic Performances of a Signal-To-Noise Ratio Moment-Based Estimator for Real Sinusoids in Additive Noise. Applied Sciences. 2019; 9(24):5556. https://doi.org/10.3390/app9245556

Chicago/Turabian Style

Romano, Gianmarco. 2019. "Asymptotic Performances of a Signal-To-Noise Ratio Moment-Based Estimator for Real Sinusoids in Additive Noise" Applied Sciences 9, no. 24: 5556. https://doi.org/10.3390/app9245556

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop