Open Access
This article is

- freely available
- re-usable

*Entropy*
**2017**,
*19*(6),
276;
https://doi.org/10.3390/e19060276

Article

Noise Enhancement for Weighted Sum of Type I and II Error Probabilities with Constraints

College of Communication Engineering, Chongqing University, Chongqing 400044, China

^{*}

Author to whom correspondence should be addressed.

Received: 12 April 2017 / Accepted: 12 June 2017 / Published: 14 June 2017

## Abstract

**:**

In this paper, the noise-enhanced detection problem is investigated for the binary hypothesis-testing. The optimal additive noise is determined according to a criterion proposed by DeGroot and Schervish (2011), which aims to minimize the weighted sum of type I and II error probabilities under constraints on type I and II error probabilities. Based on a generic composite hypothesis-testing formulation, the optimal additive noise is obtained. The sufficient conditions are also deduced to verify whether the usage of the additive noise can or cannot improve the detectability of a given detector. In addition, some additional results are obtained according to the specificity of the binary hypothesis-testing, and an algorithm is developed for finding the corresponding optimal noise. Finally, numerical examples are given to verify the theoretical results and proofs of the main theorems are presented in the Appendix.

Keywords:

noise enhancement; hypothesis testing; weighted sum; error probability## 1. Introduction

In the binary hypothesis testing problem, there are usually a null hypothesis ${H}_{0}$ and an alternative hypothesis ${H}_{1}$, and the objective of testing is to be determine truthfulness of them based on the observation data and a decision rule. Due to the presence of noise, the decision result obviously cannot be absolutely correct. Generally, two erroneous decisions may occur in the signal detection: type I error that rejects a true null hypothesis and type II error that accepts a false null hypothesis [1].

In the classical statistical theory, the Neyman–Pearson criterion is usually applied to obtain a decision rule that minimizes the type II error probability $\beta $ with a constraint on the type I error probability $\alpha $. However, the minimum $\beta $ may not always correspond to the optimal decision result. For instance, in the example 1 of [2], a binary hypothesis testing is designed to determine the mean of a normal data: the mean equals to −1 under ${H}_{0}$ and equals to 1 under ${H}_{1}$. Under the constraint that the type I error probability is fixed to 0.05, the type II error probability is decreased from 0.0091 to 0.00000026 when the data size increases from 20 to 100, whereas the rejection region of the null hypothesis is changed from (0.1, +∞) to (−0.51, +∞). In such case, more information brings a worse decision result even a smaller type II error probability is achieved with the fixed type I error probability. Similarly, the decision rule that minimizes the type I error probability for a fixed type II error probability may not perform well. Therefore, it could not be appropriate to simply minimize one of the two error probabilities in practice. The ideal case is to develop a decision criterion that minimizes the two types of error probabilities simultaneously, but it is almost impossible in practical applications.

In order to obtain a better decision result to balance the type I and II error probabilities, DeGroot and Schervish [1] proposed a criterion to minimize a weighted sum of type I and II error probabilities, i.e., $\underset{\varphi}{\mathrm{min}}[{c}_{1}\cdot \alpha (\varphi )+{c}_{2}\cdot \beta (\varphi )]$, where $\varphi $ represents the decision rule, ${c}_{1}$ and ${c}_{2}$ are the weight coefficients corresponding to $\alpha $ and $\beta $, respectively, and ${c}_{1},{c}_{2}>0$. Furthermore, DeGroot also provided the optimal decision procedure to minimize the weighted sum. The decision rule is given as follows. If ${c}_{1}f(x|{H}_{0})<{c}_{2}f(x|{H}_{1})$, the null hypothesis ${H}_{0}$ is rejected, where $f(x|{H}_{0})$ and $f(x|{H}_{1})$ are the respective probability density functions (pdfs) of the observation $\mathbf{x}$ under ${H}_{0}$ and ${H}_{1}$. If ${c}_{1}f(x|{H}_{0})>{c}_{2}f(x|{H}_{1})$, the alternative hypothesis ${H}_{1}$ is rejected. In addition, if ${c}_{1}f(x|{H}_{0})={c}_{2}f(x|{H}_{1})$, the hypothesis ${H}_{0}$ can be either rejected or not. The optimal detector in this case is closely related to the distribution of the observation. This implies that once the distribution changes, the detector should be adjusted accordingly. But in the cases where the detector is fixed, this weighted sum rule cannot be directly applied. In such a case, finding an alternative method to minimize the weighted sum of type I and II error probabilities instead of changing the detector is important. Fortunately, the stochastic resonance (SR) theory provides a means to solve this problem.

The SR, first discovered by Benzi et al. [3] in 1981, is a phenomenon where noise plays a positive role in enhancing signal and system through a nonlinear system under certain conditions. The phenomenon of SR in the signal detection is also called noise-enhanced detection. Recent studies indicate that the system output performance can be improved significantly by adding noise to the system input or increasing the background noise level [4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22]. The improvements achieved via noise can be measured in the forms of increased signal-to-noise ratio (SNR) [7,8,9,10], mutual information (MI) [11,12] or detection probability [13,14,15,16], or in the form of decreased Bayes risk [17,18]. For example, the SNR gain of a parallel uncoupled array of bistable oscillators, operating in a mixture of sinusoidal signal and Gaussian white noise, is maximized via extra array noise [8]. In addition, due to the added array noise, the performance of a finite array closely approaches to an infinite array. In [11], the throughput MI of threshold neurons is increased by increasing the intensity of faint input noise. The optimal additive noise to maximize the detection probability with a constraint on false-alarm probability is studied in [13], and the sufficient conditions for improvability and non-improvability are deduced. In [17], the effects of additive independent noise on the performance of suboptimal detectors are investigated according to the restricted Bayes criterion, where the minimum noise modified Bayes risk is explored with certain constraints on the conditional risks. Inspired by this concept, it is reasonable to conjecture that a proper noise can decrease the weighted sum of type I and II error probabilities for a fixed detector.

In the absence of constraints, it is obvious that the additive noise that minimizes the weighted sum is a constant vector, whereas the corresponding type I or II error probability may exceed a certain value to cause a bad decision result. To avoid this problem, two constraints are enforced on type I and II error probabilities, respectively, to keep a balance. The aim of this work is to find the optimal additive noise that minimizes the weighted sum of type I and II error probabilities with the constraints on type I and II error probabilities for a fixed detector. Furthermore, the work can also be extended to some applications, such as the energy detection in sensor networks [23,24] and the independent Bernoulli trials [25]. The main contributions of this paper are summarized as follows:

- Formulation of the optimization problem for minimizing the noise modified weighted sum of type I and II error probabilities under the constraints on the two error probabilities is presented.
- Derivations of the optimal noise that minimizes the weighted sum and sufficient conditions for improvability and nonimprovability for a general composite hypothesis testing problem are provided.
- Analysis of the characteristics of the optimal additive noise that minimizes the weighted sum for a simple hypothesis testing problem is studied and the corresponding algorithm to solve the optimization problem is developed.
- Numerical results are presented to verify the theoretical results and to demonstrate the superior performance of the proposed detector.

The remainder of this paper is organized as follows: in Section 2, a noise modified composite hypothesis testing problem is formulated first for minimizing the weighted sum of type I and II error probabilities under different constraints. Then the sufficient conditions for improvability and nonimprovability are given and the optimal additive noise is derived. In Section 3, additional theoretical results are analyzed for a simple hypothesis testing problem. Finally, simulation results are shown in Section 4 and conclusions are made in Section 5.

Notation: Lower-case bold letters denote vectors, with ${\upsilon}_{i}$ denoting the i-th element of $\mathit{\upsilon}$; $\theta $ denotes the value of parameter $\mathrm{\Theta}$; ${f}_{\mathit{\upsilon}}(\mathit{\upsilon}|\theta )$ denotes the pdf of $\mathit{\upsilon}$ for a given parameter value $\mathrm{\Theta}=\theta $; ${\mathrm{\Omega}}_{i}$ denotes the set of all possible parameter values of $\mathrm{\Theta}$ under ${H}_{i}$; $\delta (\cdot )$ denotes the Dirac function; $\cap $, $\cup $ and $\varnothing $ denote intersection, union and null set, respectively; $\ast $, ${(\cdot )}^{T}$, $\int$, $E\{\cdot \}$, $\mathrm{min}$, $\mathrm{max}$ and $\mathrm{arg}$ denote convolution, transpose, integral, expectation, minimum, maximum and argument operators, respectively; $\mathrm{inf}\{\cdot \}$ and $\mathrm{sup}\{\cdot \}$ denote the infimum and supremum operators, respectively; $\sum$ means summation; $\nabla $ and $H$ denote the respective gradient and Hessian operators.

## 2. Noise Enhanced Composite Hypothesis Testing

#### 2.1. Problem Formulation

Consider the following binary composite hypothesis testing problem:
where $\mathbf{x}\in {\mathbb{R}}^{N}$ is the observation vector, ${H}_{0}$ and ${H}_{1}$ are the null and the alternative hypothesizes, respectively, $\theta $ denotes the value of parameter $\mathrm{\Theta}$, ${f}_{\mathbf{x}}(\mathbf{x}|\theta )$ represents the pdf of $\mathbf{x}$ for a given parameter value $\mathrm{\Theta}=\theta $. The parameter $\mathrm{\Theta}$ has multiple possible values under each hypothesis and denote the pdf of any parameter value $\mathrm{\Theta}=\theta $ under ${H}_{0}$ and ${H}_{1}$ by ${\varpi}_{0}(\theta )$ and ${\varpi}_{1}(\theta )$. In addition, ${\mathrm{\Omega}}_{0}$ and ${\mathrm{\Omega}}_{1}$ denote the respective sets of all possible values of $\mathrm{\Theta}$ under ${H}_{0}$ and ${H}_{1}$. It is true that ${\mathrm{\Omega}}_{0}\cap {\mathrm{\Omega}}_{1}=\varnothing $ and the union of them forms the parameter space $\mathrm{\Omega}$, i.e., $\mathrm{\Omega}={\mathrm{\Omega}}_{0}\cup {\mathrm{\Omega}}_{1}$.

$$\{\begin{array}{l}{H}_{0}:{f}_{\mathbf{x}}(\mathbf{x}|\theta ),\theta \in {\mathrm{\Omega}}_{0}\\ {H}_{1}:{f}_{\mathbf{x}}(\mathbf{x}|\theta ),\theta \in {\mathrm{\Omega}}_{1}\end{array}$$

Without loss of generality, a decision rule (detector) is considered as:
where ${\mathsf{\Gamma}}_{0}$ and ${\mathsf{\Gamma}}_{1}$ form the observation space $\mathsf{\Gamma}$. Actually, the detector chooses ${H}_{1}$ if $\mathbf{x}\in {\mathsf{\Gamma}}_{1}$, otherwise chooses ${H}_{0}$ if $\mathbf{x}\in {\mathsf{\Gamma}}_{0}$.

$$\varphi (\mathbf{x})=\{\begin{array}{l}1,\mathbf{x}\in {\mathsf{\Gamma}}_{1}\\ 0,\mathbf{x}\in {\mathsf{\Gamma}}_{0}\end{array}$$

In order to investigate the performance of the detector achieved via an additive noise, a noise modified observation $\mathbf{y}$ is obtained by adding an independent additive noise $\mathbf{n}$ to the original observation $\mathbf{x}$, i.e., $\mathbf{y}=\mathbf{x}+\mathbf{n}$. For a given parameter value $\mathrm{\Theta}=\theta $, the pdf of $\mathbf{y}$ is calculated by the convolution of the pdfs of $\mathbf{x}$ and $\mathbf{n}$, given by:
where ${p}_{\mathbf{n}}(\mathbf{n})$ denotes the pdf of $\mathbf{n}$. For a fixed detector, the noise modified type I and II error probabilities of the detector for given parameter values now is expressed as:

$${f}_{\mathbf{y}}(\mathbf{y}|\theta )={f}_{\mathbf{x}}(\mathbf{x}|\theta )\ast {p}_{\mathbf{n}}(\mathbf{n})={\displaystyle {\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n}){f}_{\mathbf{x}}(\mathbf{y}-\mathbf{n}|\theta )d\mathbf{n}}$$

$${\alpha}^{\mathbf{y}}(\varphi ;\theta )={\displaystyle {\int}_{\mathsf{\Gamma}}\varphi (\mathbf{y}){f}_{\mathbf{y}}(\mathbf{y}|\theta )d}\mathbf{y}={\displaystyle {\int}_{{\mathsf{\Gamma}}_{1}}{f}_{\mathbf{y}}(\mathbf{y}|\theta )d}\mathbf{y},\theta \in {\mathrm{\Omega}}_{0}$$

$${\beta}^{\mathbf{y}}(\varphi ;\theta )={\displaystyle {\int}_{\mathsf{\Gamma}}(1-\varphi (\mathbf{y})){f}_{\mathbf{y}}(\mathbf{y}|\theta )d}\mathbf{y}={\displaystyle {\int}_{{\mathsf{\Gamma}}_{0}}{f}_{\mathbf{y}}(\mathbf{y}|\theta )d}\mathbf{y},\theta \in {\mathrm{\Omega}}_{1}$$

Correspondingly, the average noise modified type I and II error probabilities are calculated by:

$${\alpha}^{\mathbf{y}}(\varphi )={\displaystyle {\int}_{{\mathrm{\Omega}}_{0}}{\alpha}^{\mathbf{y}}(\varphi ;\theta ){\varpi}_{0}(\theta )d}\theta $$

$${\beta}^{\mathbf{y}}(\varphi )={\displaystyle {\int}_{{\mathrm{\Omega}}_{1}}{\beta}^{\mathbf{y}}(\varphi ;\theta ){\varpi}_{1}(\theta )d}\theta $$

From (6) and (7), the weighted sum of the two types of average error probabilities is obtained as:
where ${c}_{1}$ and ${c}_{2}$ are the weights assigned for the type I and II error probabilities, which can be predefined according to the actual situations. For example, if the prior probabilities are known, the value of ${c}_{1}$ and ${c}_{2}$ equal the prior probabilities corresponding to ${H}_{0}$ and ${H}_{1}$, respectively. Besides, the values of ${c}_{1}$ and ${c}_{2}$ can also be determined based on the expected decision results.

$$\begin{array}{ll}E{r}^{\mathbf{y}}& ={c}_{1}\cdot {\alpha}^{\mathbf{y}}(\varphi )+{c}_{2}\cdot {\beta}^{\mathbf{y}}(\varphi )\\ \text{}& ={c}_{1}\cdot {\displaystyle {\int}_{{\mathrm{\Omega}}_{0}}{\alpha}^{\mathbf{y}}(\varphi ;\theta ){\varpi}_{0}(\theta )d}\theta +{c}_{2}\cdot {\displaystyle {\int}_{{\mathrm{\Omega}}_{1}}{\beta}^{\mathbf{y}}(\varphi ;\theta ){\varpi}_{1}(\theta )d}\theta \end{array}$$

In this work, the aim is to find the optimal independent additive noise, which minimizes the weighted sum of the average error probabilities under the constraints on the maximum type I and II error probabilities for different parameter values. The optimization problem can be formulated as below:
subject to
where ${\alpha}_{o}$ and ${\beta}_{o}$ are the upper limits for the type I and II error probabilities, respectively.

$${p}_{\mathbf{n}}^{opt}(\mathbf{n})=\mathrm{arg}\underset{{p}_{\mathbf{n}}(\mathbf{n})}{\mathrm{min}}E{r}^{\mathbf{y}}$$

$$\{\begin{array}{l}\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{\alpha}^{\mathbf{y}}(\varphi ;\theta )\le {\alpha}_{o}\\ \underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{\beta}^{\mathbf{y}}(\varphi ;\theta )\le {\beta}_{o}\end{array}$$

In order to explicitly express the optimization problem described in (9) and (10), substituting (3) into (4) produces:
where

$$\begin{array}{ll}{\alpha}^{\mathbf{y}}(\varphi ;\theta )& ={\displaystyle {\int}_{{\mathsf{\Gamma}}_{1}}{\displaystyle {\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n}){f}_{\mathbf{x}}(\mathbf{y}-\mathbf{n}|\theta )d\mathbf{n}}d}\mathbf{y}\\ \text{}& ={\displaystyle {\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n}){\displaystyle {\int}_{{\mathsf{\Gamma}}_{1}}{f}_{\mathbf{x}}(\mathbf{y}-\mathbf{n}|\theta )d}\mathbf{y}d\mathbf{n}}\\ \text{}& ={\displaystyle {\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n}){A}_{\theta}(\mathbf{n})d\mathbf{n}}\\ \text{}& =E\{{A}_{\theta}(\mathbf{n})\}\end{array},\theta \in {\mathrm{\Omega}}_{0}$$

$${A}_{\theta}(\mathbf{n})={\displaystyle {\int}_{{\mathsf{\Gamma}}_{1}}{f}_{\mathbf{x}}(\mathbf{y}-\mathbf{n}|\theta )d}\mathbf{y}={\displaystyle {\int}_{\mathsf{\Gamma}}\varphi (\mathbf{y}){f}_{\mathbf{x}}(\mathbf{y}-\mathbf{n}|\theta )d}\mathbf{y},\theta \in {\mathrm{\Omega}}_{0}$$

It should be noted that ${A}_{\theta}(\mathbf{n})$ can be viewed as the type I error probability obtained by adding a constant vector $\mathbf{n}$ to $\mathbf{x}$ for $\theta \in {\mathrm{\Omega}}_{0}$. Therefore, ${\alpha}^{\mathbf{x}}(\varphi ;\theta )={A}_{\theta}(0)={\displaystyle {\int}_{\mathsf{\Gamma}}\varphi (\mathbf{x}){f}_{\mathbf{x}}(\mathbf{x}|\theta )d}\mathbf{x}$ denotes the type I error probability for the original observation $\mathbf{x}$.

Similarly, ${\beta}^{\mathbf{y}}(\varphi ;\theta )$ in (5) can be expressed as:
where

$${\beta}^{\mathbf{y}}(\varphi ;\theta )=E\{{B}_{\theta}(\mathbf{n})\},\text{}\theta \in {\mathrm{\Omega}}_{1}$$

$${B}_{\theta}(\mathbf{n})={\displaystyle {\int}_{{\mathsf{\Gamma}}_{0}}{f}_{\mathbf{x}}(\mathbf{y}-\mathbf{n}|\theta )d}\mathbf{y}={\displaystyle {\int}_{\mathsf{\Gamma}}(1-\varphi (\mathbf{y})){f}_{\mathbf{x}}(\mathbf{y}-\mathbf{n}|\theta )d}\mathbf{y},\text{}\theta \in {\mathrm{\Omega}}_{1}$$

The ${B}_{\theta}(\mathbf{n})$ can be treated as the type II error probability obtained by adding a constant vector $\mathbf{n}$ to $\mathbf{x}$ for $\theta \in {\mathrm{\Omega}}_{1}$ and ${\beta}^{\mathbf{x}}(\varphi ;\theta )={B}_{\theta}(0)={\displaystyle {\int}_{\mathsf{\Gamma}}(1-\varphi (\mathbf{x})){f}_{\mathbf{x}}(\mathbf{x}|\theta )d}\mathbf{x}$ is the original type II error probability without adding noise for $\theta \in {\mathrm{\Omega}}_{1}$.

With (11) and (13), (8) becomes:
where

$$E{r}^{\mathbf{y}}={\displaystyle {\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n})\left[{c}_{1}\cdot {\displaystyle {\int}_{{\mathrm{\Omega}}_{0}}{A}_{\theta}(\mathbf{n}){\varpi}_{0}(\theta )d}\theta +{c}_{2}\cdot {\displaystyle {\int}_{{\mathrm{\Omega}}_{1}}{B}_{\theta}(\mathbf{n}){\varpi}_{1}(\theta )d}\theta \right]d\mathbf{n}}=E\{Er(\mathbf{n})\}$$

$$Er(\mathbf{n})={c}_{1}\cdot {\displaystyle {\int}_{{\mathrm{\Omega}}_{0}}{A}_{\theta}(\mathbf{n}){\varpi}_{0}(\theta )d}\theta +{c}_{2}\cdot {\displaystyle {\int}_{{\mathrm{\Omega}}_{1}}{B}_{\theta}(\mathbf{n}){\varpi}_{1}(\theta )d}\theta $$

Accordingly, $Er(\mathbf{n})$ is the weighted sum of two types of average error probabilities achieved by adding a constant vector $\mathbf{n}$ to the original observation $\mathbf{x}$. Naturally, $E{r}^{\mathbf{x}}=Er(0)$ denotes the weighted sum of type I and II average error probabilities for the original observation $\mathbf{x}$.

Combined (11), (13) and (15), the optimization problem in (9) and (10) now is:
subject to

$${p}_{\mathbf{n}}^{opt}(\mathbf{n})=\mathrm{arg}\underset{{p}_{\mathbf{n}}(\mathbf{n})}{\mathrm{min}}E\{Er(\mathbf{n})\}$$

$$\{\begin{array}{l}\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}E\{{A}_{\theta}(\mathbf{n})\}\le {\alpha}_{o}\\ \underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}E\{{B}_{\theta}(\mathbf{n})\}\le {\beta}_{o}\end{array}$$

#### 2.2. Sufficient Conditions for Improvability and Non-improvability

In practice, the solution of the optimization problem in (17) and (18) requires a research over all possible noises and this procedure is complicated. Therefore, it is worthwhile to determine whether the detector can or cannot be improved by adding additive noise in advance. From (17) and (18), a detector is considered to be improvable if there exists one noise $\mathbf{n}$ that satisfies $E\{Er(\mathbf{n})\}<E{r}^{\mathbf{x}}=Er(0)$, $\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}E\{{A}_{\theta}(\mathbf{n})\}\le {\alpha}_{o}$ and $\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}E\{{B}_{\theta}(\mathbf{n})\}\le {\beta}_{o}$ simultaneously; otherwise, the detector is considered to be non-improvable.

The sufficient conditions for non-improvability can be obtained according to the characteristics of ${A}_{\theta}(\mathbf{n})$, ${B}_{\theta}(\mathbf{n})$ and $Er(\mathbf{n})$, which are provided in Theorem 1.

**Theorem**

**1.**

If there exists ${\theta}^{*}\in {\mathrm{\Omega}}_{0}$ (${\theta}^{*}\in {\mathrm{\Omega}}_{1}$) such that ${A}_{{\theta}^{*}}(\mathbf{n})\le {\alpha}_{o}$ (${B}_{{\theta}^{*}}(\mathbf{n})\le {\beta}_{o}$) implies $Er(\mathbf{n})\ge Er(0)$ for any $\mathbf{n}\in {P}_{\mathbf{n}}$, where ${P}_{\mathbf{n}}$ represents the convex set of all possible additive noises, and if ${A}_{{\theta}^{*}}(\mathbf{n})$ (${B}_{{\theta}^{*}}(\mathbf{n})$ ) and $Er(\mathbf{n})$ are convex functions over ${P}_{\mathbf{n}}$, then the detector is non-improvable.

The proof is provided in Appendix A.

Under the conditions in Theorem 1, the detector cannot be improved and it is unnecessary to solve the optimization problem in (17) and (18). In other words, if the conditions in Theorem 1 are satisfied, the three inequities $E{r}^{\mathbf{y}}\ge E{r}^{\mathbf{x}}$, $\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}E\{{A}_{\theta}(\mathbf{n})\}\le {\alpha}_{o}$ and $\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}E\{{B}_{\theta}(\mathbf{n})\}\le {\beta}_{o}$ cannot be achieved simultaneously by adding any additive noise. In addition, even if the conditions in Theorem 1 are not satisfied, the detector can also be non-improvable. This implies the sufficient conditions for improvability need to be addressed.

The sufficient conditions for improvability are discussed now. Suppose that ${A}_{\theta}(\mathbf{x})$ ($\forall \theta \in {\mathrm{\Omega}}_{0}$), ${B}_{\theta}(\mathbf{x})$ ($\forall \theta \in {\mathrm{\Omega}}_{1}$) and $Er(\mathbf{x})$ are second-order continuously differentiable around $\mathbf{x}=0$. In order to facilitate the subsequent analysis, six auxiliary functions are predefined as follows based on the first and second partial derivatives of ${A}_{\theta}(\mathbf{x})$, ${B}_{\theta}(\mathbf{x})$ and $Er(\mathbf{x})$ with respect to the elements of $\mathbf{x}$. The first three auxiliary functions ${a}_{\theta}^{(1)}(\mathbf{x},\mathbf{g})$, ${b}_{\theta}^{(1)}(\mathbf{x},\mathbf{g})$ and $e{r}^{(1)}(\mathbf{x},\mathbf{g})$ are defined as the weight sums of the first partial derivatives of ${A}_{\theta}(\mathbf{x})$, ${B}_{\theta}(\mathbf{x})$ and $Er(\mathbf{x})$, respectively, based on the coefficient vector $\mathbf{g}$. Specifically:
where $\mathbf{g}$ is a $N$-dimensional column vector, ${\mathbf{g}}^{T}$is the transposition of $\mathbf{g}$, ${\mathbf{x}}_{i}$ and ${\mathbf{g}}_{i}$ are the i-th elements of $\mathbf{x}$ and $\mathbf{g}$, respectively. In addition, $\nabla $ denotes the gradient operator, thereby $\nabla {A}_{\theta}(\mathbf{x})$ ($\nabla {B}_{\theta}(\mathbf{x})$,$\nabla Er(\mathbf{x})$) is a $N$-dimensional column vector with i-th element $\partial {A}_{\theta}(\mathbf{x})/\partial {\mathbf{x}}_{i}$ ($\partial {B}_{\theta}(\mathbf{x})/\partial {\mathbf{x}}_{i}$, $\partial {E}_{r}(\mathbf{x})/\partial {\mathbf{x}}_{i}$), $i=1,\dots ,N$. The last three auxiliary functions ${a}_{\theta}^{(2)}(\mathbf{x},\mathbf{g})$, ${b}_{\theta}^{(2)}(\mathbf{x},\mathbf{g})$ and $e{r}^{(2)}(\mathbf{x},\mathbf{g})$ are defined as the weight sums of the second partial derivatives of ${A}_{\theta}(\mathbf{x})$, ${B}_{\theta}(\mathbf{x})$ and $Er(\mathbf{x})$ based on the $N\times N$ coefficient matrix $\mathbf{g}{\mathbf{g}}^{T}$, i.e.,
where $H$ denote the Hessian operator, $H({A}_{\theta}(\mathbf{x}))$ ($H({B}_{\theta}(\mathbf{x}))$,$H(Er(\mathbf{x}))$) is a $N\times N$ matrix with its $(j,i)$-th element denoted by ${\partial}^{2}{A}_{\theta}(\mathbf{x})/(\partial {\mathbf{x}}_{j}\partial {\mathbf{x}}_{i})$ (${\partial}^{2}{B}_{\theta}(\mathbf{x})/(\partial {\mathbf{x}}_{j}\partial {\mathbf{x}}_{i})$,${\partial}^{2}Er(\mathbf{x})/(\partial {\mathbf{x}}_{j}\partial {\mathbf{x}}_{i})$), where $i,j=1,\dots ,N$.

$${a}_{\theta}^{(1)}(\mathbf{x},\mathbf{g})\triangleq {\displaystyle \sum _{i=1}^{N}{g}_{i}\frac{\partial {A}_{\theta}(\mathbf{x})}{\partial {\mathbf{x}}_{i}}={\mathbf{g}}^{T}\nabla {A}_{\theta}(\mathbf{x})},\theta \in {\mathrm{\Omega}}_{0}$$

$${b}_{\theta}^{(1)}(\mathbf{x},\mathbf{g})\triangleq {\displaystyle \sum _{i=1}^{N}{g}_{i}\frac{\partial {B}_{\theta}(\mathbf{x})}{\partial {\mathbf{x}}_{i}}}={\mathbf{g}}^{T}\nabla {B}_{\theta}(\mathbf{x}),\theta \in {\mathrm{\Omega}}_{1}$$

$$e{r}^{(1)}(\mathbf{x},\mathbf{g})\triangleq {\displaystyle \sum _{i=1}^{N}{g}_{i}\frac{\partial Er(\mathbf{x})}{\partial {\mathbf{x}}_{i}}}={\mathbf{g}}^{T}\nabla Er(\mathbf{x})$$

$${a}_{\theta}^{(2)}(\mathbf{x},\mathbf{g})\triangleq {\displaystyle \sum _{j=1}^{N}{\displaystyle \sum _{i=1}^{N}{g}_{j}{g}_{i}\frac{{\partial}^{2}{A}_{\theta}(\mathbf{x})}{(\partial {\mathbf{x}}_{j}\partial {\mathbf{x}}_{i})}={\mathbf{g}}^{T}H({A}_{\theta}(\mathbf{x}))\mathbf{g}}},\theta \in {\mathrm{\Omega}}_{0}$$

$${b}_{\theta}^{(2)}(\mathbf{x},\mathbf{g})\triangleq {\displaystyle \sum _{j=1}^{N}{\displaystyle \sum _{i=1}^{N}{g}_{j}{g}_{i}\frac{{\partial}^{2}{B}_{\theta}(\mathbf{x})}{(\partial {\mathbf{x}}_{j}\partial {\mathbf{x}}_{i})}}}={\mathbf{g}}^{T}H({B}_{\theta}(\mathbf{x}))\mathbf{g},\theta \in {\mathrm{\Omega}}_{1}$$

$$e{r}^{(2)}(\mathbf{x},\mathbf{g})\triangleq {\displaystyle \sum _{j=1}^{N}{\displaystyle \sum _{i=1}^{N}{g}_{j}{g}_{i}\frac{{\partial}^{2}Er(\mathbf{x})}{(\partial {\mathbf{x}}_{j}\partial {\mathbf{x}}_{i})}}}={\mathbf{g}}^{T}H(Er(\mathbf{x}))\mathbf{g}$$

Based on the definitions in (19)–(24), Theorem 2 presents the sufficient conditions for improvability.

**Theorem**

**2.**

Suppose that ${\mathrm{\Lambda}}_{0}$ and ${\mathrm{\Lambda}}_{1}$ are the sets of all possible values of $\theta $ that maximize ${A}_{\theta}(0)$ and ${B}_{\theta}(0)$, respectively, ${\alpha}_{o}=\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}(0)$ and ${\beta}_{o}=\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{B}_{\theta}(0)$. The detector is improvable, if there exists a $N$-dimensional column vector $\mathbf{g}$ that satisfies one of the following conditions for all ${\theta}_{0}\in {\mathrm{\Lambda}}_{0}$ and ${\theta}_{1}\in {\mathrm{\Lambda}}_{1}$ :

- (1)
- $e{r}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$, ${a}_{{\theta}_{0}}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$, ${b}_{{\theta}_{1}}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$;
- (2)
- $e{r}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}>0$, ${a}_{{\theta}_{0}}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}>0$, ${b}_{{\theta}_{1}}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}>0$;
- (3)
- $e{r}^{(2)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$, ${a}_{{\theta}_{0}}^{(2)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$, ${b}_{{\theta}_{1}}^{(2)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$.

The proof is presented in Appendix B.

Theorem 2 indicates that under the condition (1), (2) or (3), there always exist noises that decrease the weighted sum of average error probabilities under the constraints on the type I and II error probabilities. In addition, alternative sufficient conditions for improvability can be obtained by defining the following two functions, and they are:
where $I(t)$ and $S(t)$ are the minimum weighted sum of two types of average error probabilities and the maximum type II error probability for a given maximum type I error probability obtained via adding a constant vector, respectively. If there is a ${t}_{0}\le {\alpha}_{o}$ such that $I({t}_{0})\le Er(0)$ and $S(0)\le {\beta}_{o}$, the detector is improvable. More specifically, there exists a constant vector ${\mathbf{n}}_{0}$ that satisfies $\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}({\mathbf{n}}_{0})={t}_{0}\le {\alpha}_{o}$, $Er({\mathbf{n}}_{0})\le Er(0)$ and $\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{B}_{\theta}({\mathbf{n}}_{0})\le {\beta}_{o}$ simultaneously. However, in most cases, the solution of the optimization problem in (17) and (18) is not a constant vector. A more practical sufficient condition for improvability is shown in Theorem 3.

$$I(t)=\mathrm{inf}\{Er(\mathbf{n})|\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}(\mathbf{n})=t,\mathbf{n}\in {\mathbb{R}}^{N}\}$$

$$S(t)=\mathrm{sup}\{\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{B}_{\theta}(\mathbf{n})|\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}(\mathbf{n})=t,\mathbf{n}\in {\mathbb{R}}^{N}\}$$

**Theorem**

**3.**

Let $\tilde{\alpha}=\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{\alpha}^{\mathbf{x}}(\varphi ;\theta )$ and $\tilde{\beta}=\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{\beta}^{\mathbf{x}}(\varphi ;\theta )$ be the respective maximum type I and II error probabilities without adding any noise, and suppose that $\tilde{\alpha}\le {\alpha}_{o}$, $\tilde{\beta}\le {\beta}_{o}$ and $S(\tilde{\alpha})=\tilde{\beta}$. If $I(t)$ and $S(t)$ are second-order continuously differentiable around $t=\tilde{\alpha}$, and ${I}^{\u2033}(\tilde{\alpha})<0$ and ${S}^{\u2033}(\tilde{\alpha})<0$ hold at the same time, then the detector is improvable.

The proof is given in Appendix C.

Additionally, the following functions $J(\epsilon )$ and $G(\epsilon )$ are defined:

$$J(\epsilon )=\mathrm{inf}\{Er(\mathbf{n})|\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{B}_{\theta}(\mathbf{n})=\epsilon ,\mathbf{n}\in {\mathbb{R}}^{N}\}$$

$$G(\epsilon )=\mathrm{sup}\{\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}(\mathbf{n})|\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{B}_{\theta}(\mathbf{n})=\epsilon ,\mathbf{n}\in {\mathbb{R}}^{N}\}$$

A similar conclusion to the Theorem 3 can be made as well, provided in Corollary 1.

**Corollary**

**1.**

The detector is improvable, if ${J}^{\u2033}(\tilde{\beta})\le 0$ and ${G}^{\u2033}(\tilde{\beta})\le 0$ hold, where $J(\epsilon )$ and $G(\epsilon )$ are second-order continuously differentiable around $\epsilon =\tilde{\beta}$, and $G(\tilde{\beta})=\tilde{\alpha}$.

The proof is similar to that of Theorem 3 and it is omitted here.

#### 2.3. Optimal Additive Noise

In general, it is difficult to solve the optimization problem in (17) and (18) directly, because the solution is obtained based on the search over all possible additive noises. Hence, in order to reduce the computational complexity, one can utilize Parzen window density estimation to obtain an approximate solution. Actually, the pdf of the optimal additive noise can be approximated by:
where ${\eta}_{l}\ge 0$ and ${\sum}_{l=1}^{L}{\eta}_{l}}=1$, while ${\vartheta}_{l}(\cdot )$ represents the window function that satisfies ${\vartheta}_{l}(\mathbf{x})\ge 0$ for any $\mathbf{x}$ and $\int {\vartheta}_{l}(\mathbf{x})d\mathbf{x}=0$ for $l=1,\dots ,L$. The window function can be a cosine window, rectangular window, or Gauss window function. With (29), the optimization problem is simplified to obtain the parameter values corresponding to each window function. In such cases, global optimization algorithms can be applied such as Particle swarm optimization (PSO), Ant colony algorithm (ACA), and Genetic algorithm (GA) [26,27,28].

$${p}_{\mathbf{n}}(\mathbf{n})={\displaystyle \sum _{l=1}^{L}{\eta}_{l}{\vartheta}_{l}(\mathbf{n})}$$

If the numbers of parameter values in ${\mathrm{\Omega}}_{0}$ and ${\mathrm{\Omega}}_{1}$ are finite, the optimal additive noise for (17) and (18) is a randomization of no more than $M+K$ constant vectors. In this case, ${\mathrm{\Omega}}_{0}$ and ${\mathrm{\Omega}}_{1}$ can be expressed by ${\mathrm{\Omega}}_{0}=\{{\theta}_{01},{\theta}_{02},\dots ,{\theta}_{0M}\}$ and ${\mathrm{\Omega}}_{1}=\{{\theta}_{11},{\theta}_{12},\dots ,{\theta}_{1K}\}$, where $M$ and $K$ are finite positive integers. The Theorem 4 states this claim.

**Theorem**

**4.**

Suppose that each component in the optimal additive noise is finite, namely ${n}_{i}\in [{a}_{i},{b}_{i}]$ for $i=1,\dots ,N$, where ${a}_{i}$ and ${b}_{i}$ are two finite values. If ${A}_{{\theta}_{0i}}(\cdot )$ and ${B}_{{\theta}_{1i}}(\cdot )$ are continuous functions, the pdf of the optimal additive noise for the optimization problem in (17) and (18) can be expressed as:
where ${\eta}_{l}\ge 0$ and ${\sum}_{l=1}^{M+K}{\eta}_{l}}=1$.

$${p}_{\mathbf{n}}(\mathbf{n})={\displaystyle \sum _{l=1}^{M+K}{\eta}_{l}\delta (\mathbf{n}-{\mathbf{n}}_{l})}$$

The proof is similar to that of Theorem 4 in [17] and Theorem 3 in [13], and omitted here. In some special cases, the optimal additive can be solved directly based on the characteristics of $I(t)$ ($H(\epsilon )$). For example, let $E{r}_{\mathrm{min}}=\underset{t}{\mathrm{min}}I(t)=I({t}_{m})$ ($E{r}_{\mathrm{min}}=\underset{\epsilon}{\mathrm{min}}H(\epsilon )=I({\epsilon}_{m})$) and $\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}({\mathbf{n}}_{m})={t}_{m}$ ($\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{B}_{\theta}({\mathbf{n}}_{m})={\epsilon}_{m}$). If ${t}_{m}\le {\alpha}_{o}$ (${\epsilon}_{m}\le {\beta}_{o}$) and $\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{B}_{\theta}({\mathbf{n}}_{m})\le {\beta}_{o}$ ($\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}({\mathbf{n}}_{m})\le $${\alpha}_{o}$), the optimal additive noise is a constant vector with pdf of ${p}_{\mathbf{n}}(\mathbf{n})=\delta (\mathbf{n}-{\mathbf{n}}_{m})$. In addition, equality of $\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}E\{{A}_{\theta}(\mathbf{n})\}={\alpha}_{o}$ ($\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}E\{{B}_{\theta}(\mathbf{n})\}={\beta}_{o}$) holds if ${t}_{m}>{\alpha}_{o}$ (${\epsilon}_{m}>{\beta}_{o}$).

## 3. Noise Enhanced Simple Hypothesis Testing

In this section, the noise enhanced binary simple hypothesis testing problem is considered, which is a special case of the optimization problem in (9) and (10). Therefore, the conclusions obtained in Section 2 are also applicable in this section. Furthermore, due to the specificity of simple binary hypothesis testing problem, some additional results are also obtained.

#### 3.1. Problem Formulation

When ${\mathrm{\Omega}}_{i}=\{{\theta}_{i}\}$, $i=0,1$, the composite binary hypothesis testing problem described in (1) is simplified to a simple binary hypothesis testing problem. In this case, the probability of ${\theta}_{i}$ under ${H}_{i}$ equals to 1, i.e., ${\varpi}_{i}(\theta )=1$ for $i=0,1$. Therefore, the corresponding noise modified type I and II error probabilities is rewritten as:
where ${f}_{0}(\cdot )$ and ${f}_{1}(\cdot )$ represent the pdfs of $\mathbf{x}$ under ${H}_{0}$ and ${H}_{1}$, respectively, and ${A}_{0}(\mathbf{n})$ and ${B}_{1}(\mathbf{n})$ are:

$${\alpha}^{\mathbf{y}}(\varphi )={\alpha}^{\mathbf{y}}(\varphi ;{\theta}_{0})={\displaystyle {\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n}){\displaystyle {\int}_{{\mathsf{\Gamma}}_{1}}{f}_{0}(\mathbf{y}-\mathbf{n})d}\mathbf{y}d\mathbf{n}}=E\{{A}_{0}(\mathbf{n})\}$$

$${\beta}^{\mathbf{y}}(\varphi )={\beta}^{\mathbf{y}}(\varphi ;{\theta}_{1})={\displaystyle {\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n}){\displaystyle {\int}_{{\mathsf{\Gamma}}_{0}}{f}_{1}(\mathbf{y}-\mathbf{n})d}\mathbf{y}d\mathbf{n}}=E\{{B}_{1}(\mathbf{n})\}$$

$${A}_{0}(\mathbf{n})={\displaystyle {\int}_{{\mathsf{\Gamma}}_{1}}{f}_{0}(\mathbf{y}-\mathbf{n})d}\mathbf{y}$$

$${B}_{1}(\mathbf{n})={\displaystyle {\int}_{{\mathsf{\Gamma}}_{0}}{f}_{1}(\mathbf{y}-\mathbf{n})d}\mathbf{y}$$

Correspondingly, the weighted sum of noise modified type I and II error probabilities is calculated by:
where

$$\begin{array}{ll}E{r}^{\mathbf{y}}& ={c}_{1}{\alpha}^{\mathbf{y}}(\varphi )+{c}_{2}{\beta}^{\mathbf{y}}(\varphi )={c}_{1}E\{{A}_{0}(\mathbf{n})\}+{c}_{2}E\{{B}_{1}(\mathbf{n})\}\\ \text{}& ={\displaystyle {\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n})({c}_{1}{A}_{0}(\mathbf{n})+{c}_{2}{B}_{1}(\mathbf{n}))d\mathbf{n}}\\ \text{}& =E\{Er(\mathbf{n})\}\end{array}$$

$$Er(\mathbf{n})={c}_{1}{A}_{0}(\mathbf{n})+{c}_{2}{B}_{1}(\mathbf{n})$$

As a result, the optimization problem in (9) and (10) becomes:
$${p}_{\mathbf{n}}^{opt}(\mathbf{n})=\mathrm{arg}\underset{{p}_{\mathbf{n}}(\mathbf{n})}{\mathrm{min}}E\{Er(\mathbf{n})\}$$
subject to

$$\{\begin{array}{l}E\{{A}_{0}(\mathbf{n})\}\le {\alpha}_{o}\\ E\{{B}_{1}(\mathbf{n})\}\le {\beta}_{o}\end{array}$$

Based on the definitions in (33) and (34), ${A}_{0}(\mathbf{n})$ and ${B}_{1}(\mathbf{n})$ are viewed as the noise modified type I and II error probabilities obtained by adding a constant vector noise. Furthermore, ${A}_{0}(0)$ and ${B}_{1}(0)$ are the original type I and II error probabilities, respectively.

#### 3.2. Algorithm for the Optimal Additive Noise

According to the Theorem 4 in Section 2.3, the optimal additive noise for the optimization problem in (37) and (38) is a randomization of most two constant vectors with the pdf ${p}_{\mathbf{n}}^{opt}(\mathbf{n})=\eta \delta (\mathbf{n}-{\mathbf{n}}_{1})+(1-\eta )\delta (\mathbf{n}-{\mathbf{n}}_{2})$. In order to find the values of $\eta $, ${\mathbf{n}}_{1}$ and ${\mathbf{n}}_{2}$, we first divide each constant vector $\mathbf{n}$ into four disjoint sets according to the relationships of ${A}_{0}(\mathbf{n})$ and ${\alpha}_{o}$, ${B}_{1}(\mathbf{n})$ and ${\beta}_{o}$. To be specific, the four disjoint sets are ${Q}_{1}=\{\mathbf{n}|{A}_{0}(\mathbf{n})\le {\alpha}_{o},{B}_{1}(\mathbf{n})\le {\beta}_{o}\}$, ${Q}_{2}=\{\mathbf{n}|{A}_{0}(\mathbf{n})\le {\alpha}_{o},{B}_{1}(\mathbf{n})>{\beta}_{o}\}$, ${Q}_{3}=\{\mathbf{n}|{A}_{0}(\mathbf{n})>{\alpha}_{o},{B}_{1}(\mathbf{n})\le {\beta}_{o}\}$, and ${Q}_{4}=\{\mathbf{n}|{A}_{0}(\mathbf{n})>{\alpha}_{o},{B}_{1}(\mathbf{n})>{\beta}_{o}\}$. Then, we calculate the minimum $Er(\mathbf{n})$, the corresponding set of all possible values of $\mathbf{n}$ is denoted by ${Q}_{e}=\{\mathbf{n}|\mathbf{n}=\mathrm{arg}\underset{\mathbf{n}}{\mathrm{min}}Er(\mathbf{n})\}$. It should be noted that $\mathbf{n}\in {Q}_{e}$ is the optimal additive noise that minimizes the weighted sum without constraints.

It is obvious that ${Q}_{1}$, ${Q}_{2}$ and ${Q}_{3}$ do not exist if all the elements of ${Q}_{e}$ belong to ${Q}_{4}$. In other words, if ${Q}_{e}\subset {Q}_{4}$, there is no additive noise that satisfies $E\{Er(\mathbf{n})\}<Er(0)$ under the constraints of $E\{{A}_{0}(\mathbf{n})\}\le {\alpha}_{o}$ and $E\{{B}_{1}(\mathbf{n})\}\le {\beta}_{o}$. Therefore, if the detector is improvable, the elements of ${Q}_{e}$ must come from ${Q}_{1}$, ${Q}_{2}$ and/or ${Q}_{3}$. Theorem 5 is now provided to find the values of $\eta $, ${\mathbf{n}}_{1}$ and ${\mathbf{n}}_{2}$.

**Theorem**

**5.**

Let ${\eta}_{1}={\scriptscriptstyle \frac{{\alpha}_{o}-{A}_{0}({\mathbf{n}}_{2})}{{A}_{0}({\mathbf{n}}_{1})-{A}_{0}({\mathbf{n}}_{2})}}$ and ${\eta}_{2}={\scriptscriptstyle \frac{{\beta}_{o}-{B}_{1}({\mathbf{n}}_{2})}{{B}_{1}({\mathbf{n}}_{1})-{B}_{1}({\mathbf{n}}_{2})}}$.

- (1)
- If ${Q}_{e}\cap {Q}_{1}\ne \varnothing $, then $\eta =1$ and ${\mathbf{n}}_{1}\in {Q}_{e}\cap {Q}_{1}$ such that $E{r}_{opt}^{\mathbf{y}}=Er({\mathbf{n}}_{1})=\underset{\mathbf{n}}{\mathrm{min}}Er(\mathbf{n})$.
- (2)
- If ${Q}_{e}\cap {Q}_{2}\ne \varnothing $ and ${Q}_{e}\cap {Q}_{3}\ne \varnothing $ are true, then we have ${\mathbf{n}}_{1}\in {Q}_{e}\cap {Q}_{2}$, ${\mathbf{n}}_{2}\in {Q}_{e}\cap {Q}_{3}$, ${\eta}_{1}\le \eta \le {\eta}_{2}$, and $E{r}_{opt}^{\mathbf{y}}=\underset{\mathbf{n}}{\mathrm{min}}Er(\mathbf{n})$.
- (3)
- If ${Q}_{e}\subset {Q}_{2}$, then $E{r}_{opt}^{\mathbf{y}}$ is obtained when $\eta ={\eta}_{2}$, and the corresponding $E\{{A}_{0}(\mathbf{n})\}$ achieves the minimum and $E\{{B}_{1}(\mathbf{n})\}={\beta}_{o}$.
- (4)
- If ${Q}_{e}\subset {Q}_{3}$, then $E{r}_{opt}^{\mathbf{y}}$ is achieved when $\eta ={\eta}_{1}$, and the corresponding $E\{{A}_{0}(\mathbf{n})\}={\alpha}_{o}$ and $E\{{B}_{1}(\mathbf{n})\}$ reaches the minimum.

The corresponding proofs are provided in Appendix D.

From (3) and (4) in Theorem 5, under the constraints on $E\{{A}_{0}(\mathbf{n})\}\le {\alpha}_{o}$ and $E\{{B}_{1}(\mathbf{n})\}\le {\beta}_{o}$, the solution of the optimization problem in (37) and (38) is identical with the additive noise that minimizes $E\{{A}_{0}(\mathbf{n})\}$ ($E\{{B}_{1}(\mathbf{n})\}$) when ${Q}_{e}\subset {Q}_{2}$ (${Q}_{e}\subset {Q}_{3}$). In such cases, the optimal solution can be obtained easily by referring the algorithm provided in [14].

## 4. Numerical Results

In this section, a binary hypothesis testing problem is studied to verify the theoretical analysis, and it is:
where $x\in \mathbb{R}$ is an observation, $\mathrm{\Theta}$ is a constant or random variable, and $v$ is the background noise with pdf ${p}_{v}(\cdot )$. From (39), the pdf of $x$ under ${H}_{0}$ is ${f}_{0}(x)={p}_{v}(x)$, and the pdf of $x$ under ${H}_{1}$ for a given parameter value $\mathrm{\Theta}=\theta $ is denoted by ${f}_{\theta}(x)={p}_{v}(\cdot )\ast {p}_{\theta}(\cdot )$, where ${p}_{\theta}(\cdot )$ represents the pdf of $\mathrm{\Theta}=\theta $. A noise modified observation $y$ is obtained via adding an additive independent noise $n$ to the observation $x$, i.e., $y=x+n$. If the additive noise $n$ is a constant vector, the pdf of $y$ under ${H}_{0}$ is calculated as ${f}_{0}(y)={f}_{0}(x-n)$, and the pdf of $y$ under ${H}_{1}$ for $\mathrm{\Theta}=\theta $ is ${f}_{\theta}(y)={f}_{\theta}(x-n)$. In addition, a linear- quadratic detector is utilized here, given by:
where ${d}_{0}$, ${d}_{1}$ and ${d}_{2}$ are detector parameters, and $\gamma $ denotes the detection threshold. In the numerical examples, ${\alpha}_{o}={\alpha}^{x}$ and ${\beta}_{o}={\beta}^{x}$, where ${\alpha}^{x}$ and ${\beta}^{x}$ are the original type I and II error probabilities, respectively.

$$\{\begin{array}{l}{H}_{0}:x=v\\ {H}_{1}:x=\mathrm{\Theta}+v\end{array}$$

$$T(y)={d}_{0}{y}^{2}+{d}_{1}y+{d}_{2}\stackrel{{H}_{1}}{\underset{{H}_{0}}{\underset{<}{>}}}\gamma $$

#### 4.1. Rayleigh Distribution Background Noise

Suppose that $\mathrm{\Theta}=s$ is a constant, the problem shown in (39) represents a simple binary hypothesis testing problem. Here, we set ${d}_{0}={d}_{2}=0$ and ${d}_{1}=1$, then the detector becomes

$$T(y)=y\stackrel{{H}_{1}}{\underset{{H}_{0}}{\underset{<}{>}}}\gamma $$

It is assumed that the background noise $v$ obeys the mixture of Rayleigh distributions with zero-means such that ${p}_{v}(v)={\displaystyle {\sum}_{i=1}^{M}{m}_{i}{\phi}_{i}(v-{\mu}_{i})}$, where ${m}_{i}\ge 0$ for $i=1,\dots ,M$, ${\sum}_{i=1}^{M}{m}_{i}}=1$, and

$${\phi}_{i}(x)=\{\begin{array}{ll}{\scriptscriptstyle \frac{x}{{\sigma}_{i}^{2}}}\mathrm{exp}(-{\scriptscriptstyle \frac{{x}^{2}}{2{\sigma}_{i}^{2}}}),& x\ge 0\\ 0,& x<0\end{array}$$

In the simulations, the variances of all the Rayleigh components are assumed to be the same, i.e., ${\sigma}_{i}=\sigma $ for $i=1,\dots ,M$. In addition, the parameters are specified as $M=4$, ${u}_{1}=0.2$, ${u}_{2}=0.4$, ${u}_{3}=-2\sigma \sqrt{{\scriptscriptstyle \frac{\pi}{2}}}-0.2$, ${u}_{4}=-2\sigma \sqrt{{\scriptscriptstyle \frac{\pi}{2}}}-0.4$ and ${m}_{i}=0.25$ for $i=1,\dots ,4$. From (33) and (34), the noise modified type I error probability ${A}_{0}(n)$ and type II error probability ${B}_{1}(n)$ obtained by adding a constant vector $n$ is calculated as:
where $\mathrm{\Phi}(x)={\displaystyle {\int}_{x}^{\infty}{\scriptscriptstyle \frac{x}{{\sigma}^{2}}}\mathrm{exp}(-{\scriptscriptstyle \frac{{x}^{2}}{2{\sigma}^{2}}})}dt$, when $x>0$; $\mathrm{\Phi}(x)=1$, when $x\le 0$. Accordingly, ${\alpha}^{x}={A}_{0}(0)$ $={\displaystyle {\sum}_{i=1}^{4}{m}_{i}\mathrm{\Phi}(\gamma -{\mu}_{i})}$ and ${\beta}^{x}={B}_{1}(0)=1-{\displaystyle {\sum}_{i=1}^{4}{m}_{i}\mathrm{\Phi}(\gamma -{\mu}_{i}-s)}$. Let ${c}_{1}={\beta}^{x}/({\alpha}^{x}+{\beta}^{x})$ and ${c}_{2}=$ ${\alpha}^{x}/({\alpha}^{\mathbf{x}}+{\beta}^{x})$, the noise modified weighted sum of the two types of error probabilities obtained via adding a constant vector is $Er(n)={c}_{1}{\displaystyle {\sum}_{i=1}^{4}{m}_{i}\mathrm{\Phi}(\gamma}-{\mu}_{i}-n)+{c}_{2}{\displaystyle {\sum}_{i=1}^{4}{m}_{i}\mathrm{\Phi}(\gamma -{\mu}_{i}}-n-s)$.

$${A}_{0}(n)={\displaystyle {\int}_{\gamma}^{+\infty}{f}_{0}(y)dy=}{\displaystyle \sum _{i=1}^{4}{m}_{i}}\mathrm{\Phi}(\gamma -{\mu}_{i}-n)$$

$${B}_{1}(n)={\displaystyle {\int}_{-\infty}^{\gamma}{f}_{1}(y)dy=}1-{\displaystyle \sum _{i=1}^{4}{m}_{i}}\mathrm{\Phi}(\gamma -{\mu}_{i}-n-s)$$

From Section 3.2, the pdf of the optimal additive noise that minimizes weighted sum of type I and II error probabilities is denoted by ${p}_{\mathbf{n}}^{opt}(\mathbf{n})=\eta \delta (\mathbf{n}-{\mathbf{n}}_{1})+(1-\eta )\delta (\mathbf{n}-{\mathbf{n}}_{2})$, under the two constraints that ${\alpha}^{y}\le {\alpha}^{x}$ and ${\beta}^{y}\le {\beta}^{x}$. Moreover, the optimal additive noise for the case without any constraints is a constant vector.

Figure 1 plots the minimum noise modified weighted sums of type I and II error probabilities obtained under no constraint and two constraints that ${\alpha}^{y}\le {\alpha}^{x}$ and ${\beta}^{y}\le {\beta}^{x}$, and the original weighted sum without adding any noise for different values of $\sigma $ when $s=3$ and $\gamma =s/2$. When $\sigma \le {\sigma}_{1}$, there is no noise that decreases the weighted sum. With the increase of $\sigma $, noise exhibits a positive effect on the detection performance. To be specific, when ${\sigma}_{1}<\sigma <{\sigma}_{2}$, the weighted sum can be decreased by adding a constant vector for the no constraint case. When $\sigma >{\sigma}_{2}$, the weighted sum can be decreased adding the noise under two constraints. The noise modified weighted sum obtained without any constraints is less than or equal to that obtained under the two constraints, and the difference between them first decreases to zero for ${\sigma}_{3}<\sigma <{\sigma}_{4}$ and then gradually increases when $\sigma >{\sigma}_{4}$. In addition, once $\sigma $ exceeds a certain value, no noise exists that can decrease the weighted sum for any cases.

Figure 2 shows the type I and II error probabilities corresponding to the weighted sum in Figure 1. From both Figure 1 and Figure 2, it is observed that one of the noise modified Type I and II error probabilities performs worse than the original one for the no constraint case. Therefore, though the noise modified weighted sum obtained with no constraint is less than that obtained under the two constraints, the corresponding noise is actually not suitable to add to the observation. Furthermore, when the minimum value of the noise modified weighted sum is obtained under the two constraints, the corresponding type II probability equals to the original one and the type I probability achieves the minimum for ${\sigma}_{2}<\sigma <{\sigma}_{3}$. Conversely, when $\sigma >{\sigma}_{4}$, the corresponding type I probability equals to the original one and the type II probability achieves the minimum. The results are consistent with part (3) and part (4) in Theorem 5. Especially, for ${\sigma}_{3}<\sigma <{\sigma}_{4}$, the minimum values of the noise modified weighted sum obtained under no constraint is equal to that obtained under two constraints, and the corresponding type I and II error probabilities are the same, which also agrees with part (2) in Theorem 5. In order to further illustrate the results in Figure 1 and Figure 2, Table 1 provides the optimal additive noises added for the two different cases.

Figure 3 depicts the minimum noise modified weighted sums of the type I and II error probabilities versus $s$ for the cases of no constraint and two constraints, and the original weighted sum, when $\sigma =1$ and $\gamma =s/2$. The corresponding type I and II error probabilities are depicted in Figure 4a,b, respectively. It is seen in Figure 3, the improvement of the weighted sum obtained by adding noise first increases and then decreases with the increase of $s$, and finally they all converge to the same value. The differences for the cases with and without constraints are very small in most cases. In the small interval of $s$, i.e., $s\in ({s}_{1},{s}_{2})$, the difference even decreases to zero. On the other hand, the noise modified type I error probability obtained under no constraint is significantly greater than the original one for $s<{s}_{1}$, while the corresponding type II error probability is less than that obtained under the two constraints. The situation, however, is reversed for ${s}_{2}<s<{s}_{3}$. When $s>{s}_{3}$, there is no noise that decreases the weighted sum under the two constraints, while the weighted sum is still decreased by adding a constant vector for no constraint case. When $s>{s}_{4}$, the weighted sum cannot be decreased by adding any noise for all the cases. Furthermore, Table 2 shows the optimal additive noises that minimize the weighted sum under the cases of no and two constraints.

Figure 5 shows the minimum noise modified weighted sums of type I and II error probabilities versus $\gamma $ for the cases of no constraint and two constraints, and the original weighted sum, when $\sigma =1$ and $s=3$. The corresponding type I and II error probabilities are depicted in Figure 6a,b, respectively. As illustrated in Figure 5, when $\gamma $ is close to zero, the original weighted sum $E{r}^{x}$ approaches to zero. In such case, no additive noise exists to decrease the weighted sum. For the case of two constraints, the improvement of the weighted sum first increases for ${\gamma}_{1}<\gamma <{\gamma}_{2}$ and then decreases for ${\gamma}_{2}<\gamma <{\gamma}_{3}$, and no improvement can be obtained when $\gamma >{\gamma}_{2}$. On the other hand, the minimum noise modified weighted sum obtained under no constraint is smaller than that obtained under the two constraints for ${\gamma}_{1}<\gamma <{\gamma}_{3}$, and the difference between them first increases and then decreases for both ${\gamma}_{1}<\gamma <{\gamma}_{2}$ and ${\gamma}_{2}<\gamma <{\gamma}_{3}$. When $\gamma >{\gamma}_{3}$, there still exists a constant vector that decreases the weighted sum, but it may be not a suitable noise in the practical application according to the type II probability depicted in Figure 6b. Furthermore, in order to study the results illustrated in Figure 5 and Figure 6, Table 3 shows the optimal additive noises that minimize the weighted sum for the cases of no and two constraints.

#### 4.2. Gaussian Mixture Background Noise

Suppose that $\mathrm{\Theta}$ is a random variable with following pdf:

$${\varpi}_{1}(\theta )=\rho \delta (\theta -s)+(1-\rho )\delta (\theta +s)$$

Therefore, we have ${\mathrm{\Omega}}_{0}=\{0\}$ and ${\mathrm{\Omega}}_{1}=\{s,-s\}$. In the simulations, we set ${d}_{0}=1$, ${d}_{1}=0$, ${d}_{2}=-{s}^{2}/4$ and $\gamma =0$, the detector is expressed as:

$$T(y)={y}^{2}-\frac{{s}^{2}}{4}\stackrel{{H}_{1}}{\underset{{H}_{0}}{\underset{<}{>}}}0$$

Moreover, we assume that $v$ is a zero-mean symmetric Gaussian mixture noise with pdf of ${p}_{v}(v)={\displaystyle {\sum}_{i=1}^{M}{m}_{i}{\psi}_{i}(v-{\mu}_{i})}$, where ${m}_{i}\ge 0$, ${\sum}_{i=1}^{M}{m}_{i}}=1$ and:

$${\psi}_{i}(v)={\scriptscriptstyle \frac{1}{\sqrt{2\pi {\sigma}_{i}^{2}}}}\mathrm{exp}(-{\scriptscriptstyle \frac{{v}^{2}}{2{\sigma}_{i}^{2}}})$$

Let $M=4$ and the mean values of the symmetric Gaussian components are set as [0.05 0.52 −0.52 −0.05] with corresponding weights [0.35 0.15 0.15 0.35]. In addition, the variances of Gaussian components are the same, i.e., ${\sigma}_{i}=\sigma $ for $i=0,\dots ,4$. According to (12) and (14), the noise modified type I error probability obtained by adding a constant vector $n$ to $x$ is calculated by:
and the corresponding type II error probabilities for $\mathrm{\Theta}=s$ and $-s$ are respectively calculated as:
where $\Psi (x)={\displaystyle {\int}_{x}^{\infty}{\scriptscriptstyle \frac{1}{\sqrt{2\pi}}}\mathrm{exp}(-{\scriptscriptstyle \frac{{t}^{2}}{2{\sigma}^{2}}})}dt$. Accordingly:

$${A}_{0}(n)={\displaystyle \sum _{i=1}^{4}{m}_{i}}(\Psi (\frac{s/2+{\mu}_{i}+n}{\sigma})+\Psi (\frac{s/2-{\mu}_{i}-n}{\sigma}))$$

$${B}_{s}(n)=1-{\displaystyle \sum _{i=1}^{4}{m}_{i}}(\Psi (\frac{3s/2+{\mu}_{i}+n}{\sigma})+\Psi (\frac{-s/2-{\mu}_{i}-n}{\sigma}))$$

$${B}_{-s}(n)=1-{\displaystyle \sum _{i=1}^{4}{m}_{i}}(\Psi (\frac{-s/2+{\mu}_{i}+n}{\sigma})+\Psi (\frac{3s/2-{\mu}_{i}-n}{\sigma}))$$

$${B}_{1}(n)=\rho {B}_{s}(n)+(1-\rho ){B}_{-s}(n)$$

Therefore, the original type I and type II error probabilities for $\mathrm{\Theta}=s$ and $-s$ are ${\alpha}^{x}(\varphi ;0)={A}_{0}(0)$, ${\beta}^{x}(\varphi ;s)={B}_{s}(0)$ and ${\beta}^{x}(\varphi ;-s)={B}_{-s}(0)$, respectively.

Due to the symmetry property of $v$, one obtains ${B}_{s}(0)={B}_{-s}(0)$. In this case, the original average type II error probability is ${\beta}^{x}={B}_{1}(0)=\rho {\beta}^{x}(\varphi ;s)+(1-\rho ){\beta}^{x}(\varphi ;-s)={B}_{s}(0)={B}_{-s}(0)$. The noise modified weighted sum of type I and average type II error probabilities corresponding to the constant vector is expressed by $Er(n)={c}_{1}{A}_{0}(n)+{c}_{2}{B}_{1}(n)$. The values of ${c}_{1}$ and ${c}_{2}$ are still specified as ${\beta}^{x}/({\alpha}^{x}+{\beta}^{x})$ and ${\alpha}^{x}/({\alpha}^{x}+{\beta}^{x})$, respectively. From Theorem 4 in Section 2.3, the optimal additive noise that minimizes the weighted sum is a randomization with a pdf of ${p}_{n}^{opt}(n)={\eta}_{1}\delta (n-{n}_{1})+{\eta}_{2}\delta (n-{n}_{2})+{\eta}_{3}\delta (n-{n}_{3})$, where ${\eta}_{i}\ge 0$ for $i=1,\dots ,3$, and ${\sum}_{i=1}^{3}{\eta}_{i}}=1$.

Figure 7 shows the detection performance of the original detector and the noise enhanced detector that minimizes the weighted sum of type I and average type II error probabilities under the constraints that ${\alpha}^{y}(\varphi ;0)\le {\alpha}_{o}$ and $\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{\beta}^{y}(\varphi ;\theta )\le {\beta}_{o}$, for different values of $\sigma $ where $s=1$ and $\rho =0.6$. The minimum achievable noise modified weighted sum is plotted in Figure 7a, and the corresponding type I error probability and type II error probabilities for $\mathrm{\Theta}=s$ and $-s$ are depicted in Figure 7b–d, respectively.

From Figure 7, the original weighted sums, type I error probabilities, and type II error probabilities for $\mathrm{\Theta}=s$ and $-s$ increase as $\sigma $ decreases towards zero. In Figure 7a, when $\sigma $ is close to zero, the weighted sum can be decreased significantly. With the increase of $\sigma $, the improvement obtained by adding noise is reduced gradually to zero. In other words, the phenomenon of noise-enhanced detection performance cannot occur when $\sigma $ exceeds a certain value. In Figure 7b, the noise modified type I error probability stays at 0.1500 for $\sigma <0.07$ and then increases gradually to equal to the original type I error probability. Moreover, the noise modified type II error probabilities for $\mathrm{\Theta}=s$ corresponding to the minimum weighted sum increases from zero to that of original detector, shown in Figure 7c, while the type II error probabilities for $\mathrm{\Theta}=-s$ of the noise enhanced detector is equal to that of the original detector all the time. In fact, the type II error probability for $\mathrm{\Theta}=s$ also reaches the minimums under the constraints that ${\alpha}^{y}(\varphi ;0)\le {\alpha}_{o}$ and $\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{\beta}^{y}(\varphi ;\theta )\le {\beta}_{o}$ in this example. In addition, Table 4 offers the optimal additive noises that minimize the weighted sum for different values of $\sigma $ to explain the results in Figure 7. It should be noted that the optimal noise is not unique.y.

Figure 8a demonstrates the weighted sums of type I and average type II error probabilities of the original detector and the noise enhanced detector versus $s$, where $\sigma =0.08$ and $\rho =0.6$. The corresponding type I error probability and type II error probabilities for $\mathrm{\Theta}=s$ and $-s$ are depicted in Figure 8b–d, respectively. From Figure 8a, the weighted sum cannot be decreased under the constraints on different error probabilities for $s<{s}_{1}$ and $s<{s}_{2}$. Conversely, there exists additive noise under the constraints that reduces the weighted sum for ${s}_{1}<s<{s}_{2}$, and the corresponding improvement first increases and then decreases with the increase of $s$. Comparing Figure 8b with Figure 8a, it is noted that the change of the noise modified type I error probability is similar to that of the noise modified weighted sum. In Figure 8c, the noise modified type II error probability for $\mathrm{\Theta}=s$ first decreases to the minimum and then increases as $s$ increases, while the type II error probability for $\mathrm{\Theta}=-s$ of the noise modified detector is always equal to that of the original detector, shown in Figure 8d. In addition, in order to further illustrate the results in Figure 8, Table 5 shows the optimal noises that minimize the weighted sum under the case of two constraints.

## 5. Conclusions

In this paper, a noise-enhanced detection problem has been investigated for a general composite hypothesis testing. Under the constraints of type I and II error probabilities, the minimization of the weighted sum of average type I and II error probabilities has been explored by adding an additive independent noise. The sufficient conditions for improvability of the weighted sum are provided, and a simple algorithm to search the optimal noise is developed. Then some additional theoretical results are made based on the specificity of the binary simple hypothesis testing problem. The studies on different noise distributions confirm the theoretical analysis that the optimal additive noise indeed minimizes the weighted sum under certain conditions. To be noted that, theoretical results can also be extended to a broad class of noise enhanced optimization problems under two inequality constraints such as the minimization of Bayes risk under the different constraints of condition risks for a binary hypothesis testing problem.

## Acknowledgments

This research is partly supported by the Basic and Advanced Research Project in Chongqing (Grant No. cstc2016jcyjA0134, No. cstc2016jcyjA0043), the National Natural Science Foundation of China (Grant No. 61501072, No. 41404027, No. 61571069, No. 61675036, No. 61471073) and the Project No. 106112017CDJQJ168817 supported by the Fundamental Research Funds for the Central Universities.

## Author Contributions

Shujun Liu raised the idea of the framework to solve noise enhanced detection problems. Ting Yang and Shujun Liu contributed to the drafting of the manuscript, interpretation of the results, some experimental design and checked the manuscript. Kui Zhang contributed to develop the algorithm of finding the corresponding optimal noise and Ting Yang contributed to the proofs of the theories developed in this paper. All authors have read and approved the final manuscript.

## Conflicts of Interest

The authors declare no conflict of interest.

## Appendix A. Proof of Theorem 1

**Proof.**

Due to the convexity of ${A}_{{\theta}^{*}}(\mathbf{n})$ and according to the Jensen’s inequality, the type I error probability in (4) is calculated as:

$${\alpha}^{\mathbf{y}}(\varphi ;{\theta}^{*})=E\{{A}_{{\theta}^{*}}(\mathbf{n})\}\ge {A}_{{\theta}^{*}}(E\{\mathbf{n}\})$$

The contradiction method is utilized to prove this theorem. Suppose that the detector can be improved by adding noise. The improvability means that ${\alpha}^{\mathbf{y}}(\varphi ;{\theta}^{*})\le {\alpha}_{o}$ for any ${\theta}^{*}\in {\mathrm{\Omega}}_{0}$, and then ${A}_{{\theta}^{*}}(E\{\mathbf{n}\})\le {\alpha}_{o}$ from (A1). Since $E\{\mathbf{n}\}\in {P}_{\mathbf{n}}$, ${A}_{{\theta}^{*}}(E\{\mathbf{n}\})\le {\alpha}_{o}$ implies $Er(E\{\mathbf{n}\})\ge Er(0)$ based on the assumption in Theorem 1, (15) can be recalculated as:
where the first inequality holds according to the convexity of $Er(\mathbf{n})$. From (A1) and (A2), the inequality $E{r}^{\mathbf{y}}<E{r}^{\mathbf{x}}$ cannot be achieved by adding any noise under the conditions presented in Theorem 1. Therefore, the detector is nonimprovable, which contradicts the assumption. Similarly, the alternative conditions for nonimprovability stated in the parentheses can also be proved. □

$$E{r}^{\mathbf{y}}=E\{Er(\mathbf{n})\}\ge Er(E\{\mathbf{n}\})\ge Er(0)=E{r}^{\mathbf{x}}$$

## Appendix B. Proof of Theorem 2

**Proof.**

According to the definitions in (9) and (10), improvability for a detector means that there exists at least one pdf ${p}_{\mathbf{n}}(\mathbf{n})$ to satisfy three conditions, i.e., $E{r}^{\mathbf{y}}={\displaystyle {\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n})Er(\mathbf{n})d\mathbf{n}}<Er(0)$, ${\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n}){A}_{\theta}(\mathbf{n})d\mathbf{n}}\le {\alpha}_{o$ for any $\theta \in {\mathrm{\Omega}}_{0}$ and ${\int}_{{\mathbb{R}}^{N}}{p}_{\mathbf{n}}(\mathbf{n}){B}_{\theta}(\mathbf{n})d\mathbf{n}}\le {\beta}_{o$ for any $\theta \in {\mathrm{\Omega}}_{1}$. Suppose that the noise pdf ${p}_{\mathbf{n}}(\mathbf{n})$ consists of $L$ infinitesimal noise components, i.e., ${p}_{\mathbf{n}}(\mathbf{n})={\displaystyle {\sum}_{l=1}^{L}{\lambda}_{l}\delta (\mathbf{n}-{\mathsf{\epsilon}}_{l})}$. The three conditions can be rewritten as follows:

$$E{r}^{\mathbf{y}}={\displaystyle \sum _{l=1}^{L}{\lambda}_{l}}Er({\mathsf{\epsilon}}_{l})<Er(0)$$

$$\sum _{l=1}^{L}{\lambda}_{l}}{A}_{\theta}({\mathsf{\epsilon}}_{l})\le {\alpha}_{o},\forall \theta \in {\mathrm{\Omega}}_{0$$

$$\sum _{l=1}^{L}{\lambda}_{l}}{B}_{\theta}({\mathsf{\epsilon}}_{l})\le {\beta}_{o},\forall \theta \in {\mathrm{\Omega}}_{1$$

Since ${\mathsf{\epsilon}}_{l}$, $l=1,\dots ,L$, is infinitesimal, $Er({\mathsf{\epsilon}}_{l})$, ${A}_{\theta}({\mathsf{\epsilon}}_{l})$ and ${B}_{\theta}({\mathsf{\epsilon}}_{l})$ can be expressed approximately with Taylor series expansion as $Er(0)+{\mathsf{\epsilon}}_{l}^{T}\mathbf{E}\mathbf{r}+0.5{\mathsf{\epsilon}}_{l}^{T}\mathbf{H}{\mathsf{\epsilon}}_{l}$, ${A}_{\theta}(0)+{\mathsf{\epsilon}}_{l}^{T}{\mathbf{A}}_{\theta}+0.5{\mathsf{\epsilon}}_{l}^{T}{\mathbf{H}}_{\theta}^{A}{\mathsf{\epsilon}}_{l}$ and ${B}_{\theta}(0)+{\mathsf{\epsilon}}_{l}^{T}{\mathbf{B}}_{\theta}+0.5{\mathsf{\epsilon}}_{l}^{T}{\mathbf{H}}_{\mathbf{\theta}}^{B}{\mathsf{\epsilon}}_{l}$, where $\mathbf{E}\mathbf{r}$(${\mathbf{A}}_{\theta}$,${\mathbf{B}}_{\theta}$) and $\mathbf{H}$(${\mathbf{H}}_{\theta}^{A}$,${\mathbf{H}}_{\theta}^{B}$) are the gradient and the Hessian matrix of $Er(\mathbf{x})$(${A}_{\theta}(\mathbf{x})$,${B}_{\theta}(\mathbf{x})$) around $\mathbf{x}=0$, respectively. Therefore, (A3)–(A5) are rewritten as:

$$\sum _{l=1}^{L}{\lambda}_{l}{\mathsf{\epsilon}}_{l}^{T}\mathbf{E}\mathbf{r}}+0.5{\displaystyle \sum _{l=1}^{L}{\lambda}_{l}{\mathsf{\epsilon}}_{l}^{T}\mathbf{H}{\mathsf{\epsilon}}_{l}}<0$$

$$\sum _{l=1}^{L}{\lambda}_{l}{\mathsf{\epsilon}}_{l}^{T}{\mathbf{A}}_{\theta}}+0.5{\displaystyle \sum _{l=1}^{L}{\lambda}_{l}{\mathsf{\epsilon}}_{l}^{T}{\mathbf{H}}_{\theta}^{A}{\mathsf{\epsilon}}_{l}}\le {\alpha}_{o}-{A}_{\theta}(0),\forall \theta \in {\mathrm{\Omega}}_{0$$

$$\sum _{l=1}^{L}{\lambda}_{l}{\mathsf{\epsilon}}_{l}^{T}{\mathbf{B}}_{\theta}}+0.5{\displaystyle \sum _{l=1}^{L}{\lambda}_{l}{\mathsf{\epsilon}}_{l}^{T}{\mathbf{H}}_{\theta}^{B}{\mathsf{\epsilon}}_{l}}\le {\beta}_{o}-{B}_{\theta}(0),\forall \theta \in {\mathrm{\Omega}}_{1$$

Let ${\mathsf{\epsilon}}_{l}$ be expressed by ${\mathsf{\epsilon}}_{l}={\tau}_{l}\mathbf{g}$, where $\mathbf{g}$ is a $N$-dimensional real vector and ${\tau}_{l}$ is an infinitesimal real value, $l=1,\dots ,L$. Accordingly, one obtains:

$$\sum _{l=1}^{L}{\lambda}_{l}{\tau}_{l}{\mathbf{g}}^{T}\mathbf{E}\mathbf{r}}+0.5{\displaystyle \sum _{l=1}^{L}{\lambda}_{l}{\tau}_{l}^{2}{\mathbf{g}}^{T}\mathbf{H}\mathbf{g}}<0$$

$$\sum _{l=1}^{L}{\lambda}_{l}{\tau}_{l}{\mathbf{g}}^{T}{\mathbf{A}}_{\theta}}+0.5{\displaystyle \sum _{l=1}^{L}{\lambda}_{l}{\tau}_{l}^{2}{\mathbf{g}}^{T}{\mathbf{H}}_{\theta}^{A}\mathbf{g}}\le {\alpha}_{o}-{A}_{\theta}(0),\forall \theta \in {\mathrm{\Omega}}_{0$$

$$\sum _{l=1}^{L}{\lambda}_{l}{\tau}_{l}{\mathbf{g}}^{T}{\mathbf{B}}_{\theta}}+0.5{\displaystyle \sum _{l=1}^{L}{\lambda}_{l}{\tau}_{l}^{2}{\mathbf{g}}^{T}{\mathbf{H}}_{\theta}^{B}\mathbf{g}}\le {\beta}_{o}-{B}_{\theta}(0),\forall \theta \in {\mathrm{\Omega}}_{1$$

Based on the definitions given in (19)–(24), (A9)–(A11) are simplified as:
where $k=2{\displaystyle {\sum}_{l=1}^{L}{\lambda}_{l}{\tau}_{l}}/{\displaystyle {\sum}_{l=1}^{L}{\lambda}_{l}{\tau}_{l}^{2}}$. As ${\alpha}_{o}={A}_{\theta}(0)$ for $\theta \in {\mathrm{\Lambda}}_{0}$ and ${\alpha}_{o}>{A}_{\theta}(0)$ for $\theta \in {\mathrm{\Omega}}_{0}{\mathrm{\Lambda}}_{0}$, the right-hand side of (A13) approaches to plus infinity for $\theta \in {\mathrm{\Omega}}_{0}{\mathrm{\Lambda}}_{0}$. Similarly, when ${\beta}_{o}={B}_{\theta}(0)$ for $\theta \in {\mathrm{\Lambda}}_{1}$ and ${\beta}_{o}>{B}_{\theta}(0)$ for $\theta \in {\mathrm{\Omega}}_{1}{\mathrm{\Lambda}}_{1}$, the right-hand side of (A14) also goes to plus infinity for $\theta \in {\mathrm{\Omega}}_{1}{\mathrm{\Lambda}}_{1}$. Therefore, we only need to consider the cases of $\theta \in {\mathrm{\Lambda}}_{0}$ and $\theta \in {\mathrm{\Lambda}}_{1}$. In doing so, (A12)–(A14) are now:

$$(k\cdot e{r}^{(1)}(\mathbf{x},\mathbf{g})+e{r}^{(2)}(\mathbf{x},\mathbf{g})){|}_{\mathbf{x}=0}<0$$

$$(k\cdot {a}_{\theta}^{(1)}(\mathbf{x},\mathbf{g})+{a}_{\theta}^{(2)}(\mathbf{x},\mathbf{g})){|}_{\mathbf{x}=0}<\frac{2({\alpha}_{o}-{A}_{\theta}(0))}{{\displaystyle {\sum}_{l=1}^{L}{\lambda}_{l}{\tau}_{l}^{2}}},\forall \theta \in {\mathrm{\Omega}}_{0}$$

$$(k\cdot {b}_{\theta}^{(1)}(\mathbf{x},\mathbf{g})+{b}_{\theta}^{(2)}(\mathbf{x},\mathbf{g})){|}_{\mathbf{x}=0}<\frac{2({\beta}_{o}-{B}_{\theta}(0))}{{\displaystyle {\sum}_{l=1}^{L}{\lambda}_{l}{\tau}_{l}^{2}}},\forall \theta \in {\mathrm{\Omega}}_{1}$$

$$(k\cdot e{r}^{(1)}(\mathbf{x},\mathbf{g})+e{r}^{(2)}(\mathbf{x},\mathbf{g})){|}_{\mathbf{x}=0}<0$$

$$(k\cdot {a}_{\theta}^{(1)}(\mathbf{x},\mathbf{g})+{a}_{\theta}^{(2)}(\mathbf{x},\mathbf{g})){|}_{\mathbf{x}=0}<0,\forall \theta \in {\mathrm{\Lambda}}_{0}$$

$$(k\cdot {b}_{\theta}^{(1)}(\mathbf{x},\mathbf{g})+{b}_{\theta}^{(2)}(\mathbf{x},\mathbf{g})){|}_{\mathbf{x}=0}<0,\forall \theta \in {\mathrm{\Lambda}}_{1}$$

It is obvious that $k$ can be set as any real value by choosing appropriate ${\lambda}_{l}$ and ${\tau}_{l}$. As a result, (A15)–(A17) can be satisfied by selecting a suitable value of $k$ under each condition in Theorem 2. That is:

- (1)
- Inequalities (A15)–(A17) can be satisfied by setting $k$ as a sufficiently large positive number, if $e{r}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$, ${a}_{{\theta}_{0}}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$, ${b}_{{\theta}_{1}}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$ hold.
- (2)
- Inequalities (A15)–(A17) can be satisfied by setting $k$ as a sufficiently large negative number, if $e{r}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}>0$, $e{a}_{{\theta}_{0}}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}>0$, ${b}_{{\theta}_{1}}^{(1)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}>0$ hold.
- (3)
- Inequalities (A15)–(A17) can be satisfied by setting $k$ as zero, if $e{r}^{(2)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$, ${a}_{{\theta}_{0}}^{(2)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$, ${b}_{{\theta}_{1}}^{(2)}(\mathbf{x},\mathbf{g}){|}_{\mathbf{x}=0}<0$ hold. □

## Appendix C. Proof of Theorem 3

**Proof.**

Since $I(t)$ and $S(t)$ are second-order continuously differentiable around $t=\tilde{\alpha}$, there exists a $\xi >0$ such that ${I}^{\u2033}(\cdot )<0$ and ${S}^{\u2033}(\cdot )<0$ for $\mathsf{\Delta}=(\tilde{\alpha}-\xi ,\tilde{\alpha}+\xi )$. If one adds a noise with pdf ${p}_{\widehat{\mathbf{n}}}(\mathbf{n})=0.5\delta (\mathbf{n}-{\mathbf{n}}_{1})+0.5\delta (\mathbf{n}-{\mathbf{n}}_{2})$, where $\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}({\mathbf{n}}_{1})=\tilde{\alpha}+\xi $ and $\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}({\mathbf{n}}_{2})=\tilde{\alpha}-\xi $, to the original observation $\mathbf{x}$, the maximum values of corresponding noise modified type I and II error probabilities are:

$$\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}E\{{A}_{\theta}(\widehat{\mathbf{n}})\}\le E\{\underset{\theta \in {\mathrm{\Omega}}_{0}}{\mathrm{max}}{A}_{\theta}(\widehat{\mathbf{n}})\}\le 0.5(\tilde{\alpha}+\xi )+0.5(\tilde{\alpha}-\xi )=\tilde{\alpha}\le {\alpha}_{o}$$

$$\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}E\{{B}_{\theta}(\widehat{\mathbf{n}})\}\le E\{\underset{\theta \in {\mathrm{\Omega}}_{1}}{\mathrm{max}}{B}_{\theta}(\widehat{\mathbf{n}})\}\le 0.5S(\tilde{\alpha}+\xi )+0.5S(\tilde{\alpha}-\xi )\le S(\tilde{\alpha})=\tilde{\beta}\le {\beta}_{o}$$

In addition:

$$E\{Er(\widehat{\mathbf{n}})\}=E\{I(t)\}=0.5I(\tilde{\alpha}+\xi )+0.5I(\tilde{\alpha}-\xi )<I(\tilde{\alpha})$$

One obtains $E\{Er(\widehat{\mathbf{n}})\}<Er(0)$ because $I(\tilde{\alpha})\le Er(0)$ according to the definition of $I(t)$. As a result, the detector is improvable. □

## Appendix D. Proof of Theorem 5

**Proof.**

Part (1): If ${Q}_{e}\cap {Q}_{1}\ne \varnothing $, any ${\mathbf{n}}_{1}\in {Q}_{e}\cap {Q}_{1}$ satisfies the constraints of ${A}_{0}({\mathbf{n}}_{1})\le {\alpha}_{o}$ and ${B}_{1}({\mathbf{n}}_{1})\le {\beta}_{o}$ based on the definition of ${Q}_{1}$ and $E{r}_{opt}^{\mathbf{y}}=Er({\mathbf{n}}_{1})=\underset{\mathbf{n}}{\mathrm{min}}Er(\mathbf{n})<Er(0)$ according to the definition of ${Q}_{e}$.

Part (2): If ${Q}_{e}\cap {Q}_{2}\ne \varnothing $ and ${Q}_{e}\cap {Q}_{3}\ne \varnothing $ simultaneously, there exists ${\mathbf{n}}_{1}\in {Q}_{e}\cap {Q}_{2}$ and ${\mathbf{n}}_{2}\in {Q}_{e}\cap {Q}_{3}$ such that $Er({\mathbf{n}}_{1})=Er({\mathbf{n}}_{2})=\underset{\mathbf{n}}{\mathrm{min}}Er(\mathbf{n})$ based on the definition of ${Q}_{e}$. In order to meet the constraints that $E\{{A}_{0}(\mathbf{n})\}\le {\alpha}_{o}$ and $E\{{B}_{1}(\mathbf{n})\}\le {\beta}_{o}$, the noise components $\eta $, ${\mathbf{n}}_{1}$ and ${\mathbf{n}}_{2}$ should satisfy the following two inequalities:

$$\eta {A}_{0}({\mathbf{n}}_{1})+(1-\eta ){A}_{0}({\mathbf{n}}_{2})\le {\alpha}_{o}$$

$$\eta {B}_{1}({\mathbf{n}}_{1})+(1-\eta ){B}_{1}({\mathbf{n}}_{2})\le {\beta}_{o}$$

Consequently, $\eta \ge {\eta}_{1}={\scriptscriptstyle \frac{{\alpha}_{o}-{A}_{0}({\mathbf{n}}_{2})}{{A}_{0}({\mathbf{n}}_{1})-{A}_{0}({\mathbf{n}}_{2})}}$ and $\eta \le {\eta}_{2}={\scriptscriptstyle \frac{{\beta}_{o}-{B}_{1}({\mathbf{n}}_{2})}{{B}_{1}({\mathbf{n}}_{1})-{B}_{1}({\mathbf{n}}_{2})}}$ according to the definitions of ${Q}_{2}$ and ${Q}_{3}$. If ${\eta}_{1}\le \eta \le {\eta}_{2}$, the noise with pdf ${p}_{\mathbf{n}}^{opt}(\mathbf{n})=\eta \delta (\mathbf{n}-{\mathbf{n}}_{1})+(1-\eta )\delta (\mathbf{n}-{\mathbf{n}}_{2})$ can minimize $E\{Er(\mathbf{n})\}$ and satisfy the two inequalities, and $E{r}_{opt}^{\mathbf{y}}=\eta Er({\mathbf{n}}_{1})+(1-\eta )Er({\mathbf{n}}_{2})=\underset{\mathbf{n}}{\mathrm{min}}Er(\mathbf{n})$.

Part (3): If ${Q}_{e}\subset {Q}_{2}$, the optimal additive noise is not a constant vector, i.e., $\eta \ne 1$. Therefore, one of ${\mathbf{n}}_{1}$ and ${\mathbf{n}}_{2}$ belongs to ${Q}_{2}$ and the second one comes from ${Q}_{1}$ or ${Q}_{3}$. In addition, $\eta $, ${\mathbf{n}}_{1}$ and ${\mathbf{n}}_{2}$ should also satisfy the two constraints in (A21) and (A22).

First, suppose that ${\mathbf{n}}_{1}\in {Q}_{2}$ and ${\mathbf{n}}_{2}\in {Q}_{1}$, then (A21) holds based on the definitions of ${Q}_{1}$ and ${Q}_{2}$. We should only consider the constraint in (A22), which implies $\eta \le {\eta}_{2}$. It is true that ${A}_{0}({\mathbf{n}}_{2})\le {\alpha}_{o}$ and ${B}_{1}({\mathbf{n}}_{2})\le {\beta}_{o}$ according to the definition of ${Q}_{1}$. If $Er({\mathbf{n}}_{1})>Er({\mathbf{n}}_{2})$, we have $Er({\mathbf{n}}_{2})<E{r}_{opt}^{\mathbf{y}}=\eta Er({\mathbf{n}}_{1})+(1-\eta )Er({\mathbf{n}}_{2})$, which contradicts with the definition of ${p}_{\mathbf{n}}^{opt}(\mathbf{n})$. Hence, $Er({n}_{1})<Er({n}_{2})$ and the minimum of $E\{Er(\mathbf{n})\}$ is obtained when $\eta ={\eta}_{2}$.

Next, suppose that ${\mathbf{n}}_{1}\in {Q}_{2}$ and ${\mathbf{n}}_{2}\in {Q}_{3}$. The two inequalities in (A21) and (A22) require that ${\eta}_{1}\le \eta \le {\eta}_{2}$. If $Er({\mathbf{n}}_{1})>Er({\mathbf{n}}_{2})$, the minimum of $E\{Er(\mathbf{n})\}$ is obtained when $\eta ={\eta}_{1}$. In such case, there exists a noise with pdf ${p}_{\widehat{\mathbf{n}}}(\mathbf{n})=\varsigma {p}_{\mathbf{n}}^{opt}(\mathbf{n})+(1-\varsigma )\delta (\mathbf{n}-{\mathbf{n}}_{e})$ that satisfies $E\{{A}_{0}(\widehat{\mathbf{n}})\}\le {\alpha}_{o}$ and $E\{{B}_{1}(\widehat{\mathbf{n}})\}\le {\beta}_{o}$ simultaneously, where ${\mathbf{n}}_{e}\in {Q}_{e}$ and $0\le \varsigma \le 1$. Therefore, $E\{Er(\widehat{\mathbf{n}})\}=\varsigma E{r}_{opt}^{\mathbf{y}}+$ $(1-\varsigma )Er({\mathbf{n}}_{e})<E{r}_{opt}^{\mathbf{y}}$ since $Er({\mathbf{n}}_{e})=\underset{\mathbf{n}}{\mathrm{min}}Er(\mathbf{n})<E{r}_{opt}^{\mathbf{y}}$, which contradicts with the definition of ${p}_{\mathbf{n}}^{opt}(\mathbf{n})$. As a result, $Er({\mathbf{n}}_{1})<Er({\mathbf{n}}_{2})$ and the minimum of $E\{Er(\mathbf{n})\}$ is obtained when $\eta ={\eta}_{2}$.

When $\eta ={\eta}_{2}$, one obtains $E\{{B}_{1}(\mathbf{n})\}=\eta {B}_{1}({\mathbf{n}}_{1})+(1-\eta ){B}_{1}({\mathbf{n}}_{2})={\beta}_{o}$. In other words, the minimum of $E\{Er(\mathbf{n})\}$ is obtained when $E\{{A}_{0}(\mathbf{n})\}$ achieves the minimum and $E\{{A}_{0}(\mathbf{n})\}$. Accordingly, one obtains $E{r}_{opt}^{\mathbf{y}}={c}_{1}{\alpha}_{opt}^{\mathbf{y}}+{c}_{2}{\beta}_{o}$.

Part (4): The proof of Part (4) is similar to that of Part (3) and it is omitted here. □

## References

- DeGroot, M.H.; Sxhervish, M.J. Probability and Statistics, 4nd ed.; Addison-Wesley: Boston, MA, USA, 2011. [Google Scholar]
- Pericchi, L.; Pereira, C. Adaptative significance levels using optimal decision rules: Balancing by weighting the error probabilities. Braz. J. Probab. Stat.
**2016**, 30, 70–90. [Google Scholar] [CrossRef] - Benzi, R.; Sutera, A.; Vulpiani, A. The mechanism of stochastic resonance. J. Phys. A Math.
**1981**, 14, 453–457. [Google Scholar] [CrossRef] - Patel, A.; Kosko, B. Noise benefits in quantizer-array correlation detection and watermark decoding. IEEE Trans. Signal Process.
**2011**, 59, 488–505. [Google Scholar] [CrossRef] - Han, D.; Li, P.; An, S.; Shi, P. Multi-frequency weak signal detection based on wavelet transform and parameter compensation band-pass multi-stable stochastic resonance. Mech. Syst. Signal Process.
**2016**, 70–71, 995–1010. [Google Scholar] [CrossRef] - Addesso, P.; Pierro, V.; Filatrella, G. Interplay between detection strategies and stochastic resonance properties. Commun. Nonlinear Sci. Numer. Simul.
**2016**, 30, 15–31. [Google Scholar] [CrossRef] - Gingl, Z.; Makra, P.; Vajtai, R. High signal-to-noise ratio gain by stochastic resonance in a double well. Fluct. Noise Lett.
**2001**, 1, L181–L188. [Google Scholar] [CrossRef] - Makra, P.; Gingl, Z. Signal-to-noise ratio gain in non-dynamical and dynamical bistable stochastic resonators. Fluct. Noise Lett.
**2002**, 2, L147–L155. [Google Scholar] [CrossRef] - Makra, P.; Gingl, Z.; Fulei, T. Signal-to-noise ratio gain in stochastic resonators driven by coloured noises. Phys. Lett. A
**2003**, 317, 228–232. [Google Scholar] [CrossRef] - Duan, F.; Chapeau-Blondeau, F.; Abbott, D. Noise-enhanced SNR gain in parallel array of bistable oscillators. Electron. Lett.
**2006**, 42, 1008–1009. [Google Scholar] [CrossRef] - Mitaim, S.; Kosko, B. Adaptive stochastic resonance in noisy neurons based on mutual information. IEEE Trans. Neural Netw.
**2004**, 15, 1526–1540. [Google Scholar] [CrossRef] [PubMed] - Patel, A.; Kosko, B. Mutual-Information Noise Benefits in Brownian Models of Continuous and Spiking Neurons. In Proceedings of the 2006 International Joint Conference on Neural Network, Vancouver, BC, Canada, 16–21 July 2006; pp. 1368–1375. [Google Scholar]
- Chen, H.; Varshney, P.K.; Kay, S.M.; Michels, J.H. Theory of the stochastic resonance effect in signal detection: Part I – fixed detectors. IEEE Trans. Signal Process.
**2007**, 55, 3172–3184. [Google Scholar] [CrossRef] - Patel, A.; Kosko, B. Optimal noise benefits in Neyman–Pearson and inequality constrained signal detection. IEEE Trans. Signal Process.
**2009**, 57, 1655–1669. [Google Scholar] - Bayram, S.; Gezici, S. Stochastic resonance in binary composite hypothesis-testing problems in the Neyman–Pearson framework. Digit. Signal Process.
**2012**, 22, 391–406. [Google Scholar] [CrossRef] - Bayrama, S.; Gultekinb, S.; Gezici, S. Noise enhanced hypothesis-testing according to restricted Neyman–Pearson criterion. Digit. Signal Process.
**2014**, 25, 17–27. [Google Scholar] [CrossRef] - Bayram, S.; Gezici, S.; Poor, H.V. Noise enhanced hypothesis-testing in the restricted Bayesian framework. IEEE Trans. Signal Process.
**2010**, 58, 3972–3989. [Google Scholar] [CrossRef] - Bayram, S.; Gezici, S. Noise enhanced M-ary composite hypothesis-testing in the presence of partial prior information. IEEE Trans. Signal Process.
**2011**, 59, 1292–1297. [Google Scholar] [CrossRef] - Chen, H.; Varshney, L.R.; Varshney, P.K. Noise-enhanced information systems. Proc. IEEE
**2014**, 102, 1607–1621. [Google Scholar] [CrossRef] - Weber, J.F.; Waldman, S.D. Stochastic Resonance is a Method to Improve the Biosynthetic Response of Chondrocytes to Mechanical Stimulation. J. Orthop. Res.
**2015**, 34, 231–239. [Google Scholar] [CrossRef] [PubMed] - Duan, F.; Chapeau-Blondeau, F.; Abbott, D. Non-Gaussian noise benefits for coherent detection of narrow band weak signal. Phys. Lett. A
**2014**, 378, 1820–1824. [Google Scholar] [CrossRef] - Lu, Z.; Chen, L.; Brennan, M.J.; Yang, T.; Ding, H.; Liu, Z. Stochastic resonance in a nonlinear mechanical vibration isolation system. J. Sound Vib.
**2016**, 370, 221–229. [Google Scholar] [CrossRef] - Rossi, P.S.; Ciuonzo, D.; Ekman, T.; Dong, H. Energy Detection for MIMO Decision Fusion in Underwater Sensor Networks. IEEE Sen. J.
**2015**, 15, 1630–1640. [Google Scholar] [CrossRef] - Rossi, P.S.; Ciuonzo, D.; Kansanen, K.; Ekman, T. Performance Analysis of Energy Detection for MIMO Decision Fusion in Wireless Sensor Networks Over Arbitrary Fading Channels. IEEE Trans. Wirel. Commun.
**2016**, 15, 7794–7806. [Google Scholar] [CrossRef] - Ciuonzo, D.; de Maio, A.; Rossi, P.S. A Systematic Framework for Composite Hypothesis Testing of Independent Bernoulli Trials. IEEE Signal Proc. Lett.
**2015**, 22, 1249–1253. [Google Scholar] [CrossRef] - Parsopoulos, K.E.; Vrahatis, M.N. Particle Swarm Optimization Method for Constrained Optimization Problems; IOS Press: Amsterdam, The Netherlands, 2002; pp. 214–220. [Google Scholar]
- Hu, X.; Eberhart, R. Solving constrained nonlinear optimization problems with particle swarm optimization. In Proceedings of the sixth world multiconference on systemics, cybernetics and informatics, Orlando, FL, USA, 14–18 July 2002. [Google Scholar]
- Price, K.V.; Storn, R.M.; Lampinen, J.A. Differential Evolution: A Practical Approach to Global Optimization; Springer: New York, NY, USA, 2005. [Google Scholar]

**Figure 1.**The minimum noise modified weighted sums of the type I and II error probabilities obtained under no constraint and two constraints, and the original weighted sum for different $\sigma $ when s = 3 and γ = s/2.

**Figure 2.**The type I (

**a**) and II (

**b**) error probabilities corresponding to the weighted sum in Figure 1.

**Figure 3.**The minimum noise modified weighted sums of the type I and II error probabilities obtained under no constraint and two constraints, and the original weighted sum for different $s$ when σ = 1 and γ = s/2.

**Figure 4.**The type I and II error probabilities corresponding to the weighted sum in Figure 3 are shown in (

**a**) and (

**b**), respectively.

**Figure 5.**The minimum noise modified weighted sums of the type I and II error probabilities obtained under no constraint and two constraints, and the original weighted sum for different $\gamma $ when σ = 1 and s = 3.

**Figure 6.**The type I and II error probabilities corresponding to the weighted sum in Figure 5 are shown in (

**a**) and (

**b**), respectively.

**Figure 7.**The weighted sums, type I error probabilities, and type II error probabilities for $\mathrm{\Theta}=s$ and −s of the original detector and the noise enhanced detector for different σ where s = 1 and ρ = 0.6 shown in (

**a**), (

**b**), (

**c**) and (

**d**), respectively

**Figure 8.**The weighted sums, type I error probabilities, and type II error probabilities for $\mathrm{\Theta}=s$ and −s of the original detector and the noise enhanced detector for different s where σ = 0.08 and ρ = 0.6 shown in (a), (b), (c) and (d), respectively.

**Table 1.**The optimal additive noises that minimize the weighted sum under two constraints and no constraint for various $\sigma $ where s = 3 and γ = s/2.

$\mathit{\sigma}$ | Two Constraints | No Constraints | ||
---|---|---|---|---|

${\mathit{n}}_{\mathbf{1}}$ | ${\mathit{n}}_{\mathbf{2}}$ | $\eta $ | ${\mathit{n}}_{\mathit{o}}$ | |

0.950 | - | - | - | −1.7089 |

1.250 | −1.9082 | 1.7963 | 0.6950 | −1.9218 |

2.125 | −2.5136 | 3.1896 | 0.7862 | −2.5136/3.1896 |

3.000 | −3.3771 | 4.6942 | 0.3770 | 4.7449 |

**Table 2.**The optimal additive noises that minimize the weighted sum under two constraints and no constraint for various $s$ where σ = 1 and γ = s/2.

$\mathit{s}$ | Two Constraints | No Constraints | ||
---|---|---|---|---|

${\mathit{n}}_{\mathbf{1}}$ | ${\mathit{n}}_{\mathbf{2}}$ | $\mathit{\eta}$ | ${\mathit{n}}_{\mathit{o}}$ | |

1.25 | −1.3682 | 1.7327 | 0.2918 | 1.7474 |

1.75 | −1.4408 | 1.6563 | 0.7265 | −1.4408/1.6563 |

2.5 | −1.6052 | 1.4690 | 0.6983 | −1.6201 |

3.25 | - | - | - | −0.5866 |

**Table 3.**The optimal additive noises that minimize the weighted sum under two constraints and no constraint for various $\gamma $ where σ = 1 and s = 3.

$\mathit{\gamma}$ | Two Constraints | No Constraints | ||
---|---|---|---|---|

${\mathit{n}}_{\mathbf{1}}$ | ${\mathit{n}}_{\mathbf{2}}$ | $\mathit{\eta}$ | ${\mathit{n}}_{\mathit{o}}$ | |

0.050 | - | - | - | - |

1.100 | −2.1213 | 0.9341 | 0.2878 | 0.9691 |

1.425 | −1.7947 | 1.2585 | 0.5355 | −1.7957 |

2.250 | −0.9693 | 2.0836 | 0.8867 | −1.1763 |

3.375 | - | - | - | −0.5775 |

**Table 4.**The optimal additive noises that minimize the weighted sum under two constraints for various $\sigma $ where s = 1 and ρ = 0.6.

$\mathit{\sigma}$ | ${\mathit{n}}_{\mathbf{1}}$ | ${\mathit{n}}_{\mathbf{2}}$ | ${\mathit{n}}_{\mathbf{3}}$ | ${\mathit{\eta}}_{\mathbf{1}}$ | ${\mathit{\eta}}_{\mathbf{2}}$ | ${\mathit{\eta}}_{\mathbf{3}}$ |
---|---|---|---|---|---|---|

0.0001 | 0.2286 | - | - | 1.0000 | - | - |

0.02 | 0.2286 | −0.2255 | - | 0.8413 | 0.1587 | - |

0.05 | 0.2287 | −0.2208 | 0.2421 | 0.5310 | 0.3446 | 0.1244 |

0.08 | 0.2180 | −0.2185 | −0.2168 | 0.5943 | 0.2449 | 0.1608 |

**Table 5.**The optimal additive noises that minimize the weighted sum under two constraints for various $s$ where σ = 0.08 and ρ = 0.6.

$\mathit{s}$ | ${\mathit{n}}_{1}$ | ${\mathit{n}}_{2}$ | ${\mathit{n}}_{3}$ | ${\mathit{\eta}}_{1}$ | ${\mathit{\eta}}_{2}$ | ${\mathit{\eta}}_{3}$ |
---|---|---|---|---|---|---|

0.65 | 0.1613 | −0.1613 | - | 0.6267 | 0.3733 | - |

0.75 | 0.2026 | −0.2026 | - | 0.7949 | 0.2051 | - |

0.85 | 0.2148 | −0.2149 | -0.2150 | 0.8262 | 0.1300 | 0.0438 |

0.95 | 0.2195 | −0.2196 | -0.2190 | 0.7006 | 0.1916 | 0.1078 |

© 2017 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).