Next Article in Journal
Light Propagation and Multi-Scale Enhanced DeepLabV3+ for Underwater Crack Detection
Next Article in Special Issue
MHD-Protonet: Margin-Aware Hard Example Mining for SAR Few-Shot Learning via Dual-Loss Optimization
Previous Article in Journal
Algorithm Providing Ordered Integer Sequences for Sampling with Replacement Confidence Intervals
Previous Article in Special Issue
High-Performance Computing and Parallel Algorithms for Urban Water Demand Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Frequency Domain Image Filtering Based on a Multilayer Neural Network with Multi-Valued Neurons

1
Department of Computer Science, Manhattan University, Riverdale, NY 10471, USA
2
Department of Systems Analysis and Optimization Theory, State University “Uzhhorod National University”, 88000 Uzhhorod, Ukraine
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Algorithms 2025, 18(8), 461; https://doi.org/10.3390/a18080461
Submission received: 3 July 2025 / Revised: 18 July 2025 / Accepted: 21 July 2025 / Published: 24 July 2025

Abstract

Neural networks have shown significant promise in the field of image processing, particularly for tasks such as denoising and restoration, due to their capacity to model complex nonlinear relationships between inputs and outputs. In this study, we explored the application of a complex-valued neural network—a multilayer neural network with multi-valued neurons (MLMVN)—for filtering two types of noise in digital images: additive Gaussian noise and multiplicative speckle noise. The proposed approach involves processing images as a set of overlapping patches in the frequency domain using MLMVN. Training was performed using a batch learning algorithm, which proved to be more efficient for big learning sets: it results in fewer learning epochs and a better generalization capability. Experimental results demonstrated that MLMVN achieves noise filtering quality comparable to well-established methods, such as the BM3D, Lee, and Frost filters. These findings suggest that MLMVN offers a viable framework for image denoising, particularly in scenarios where frequency domain processing is advantageous. Also, complex-valued logistic and hyperbolic tangent activation functions were used for multi-valued neurons for the first time and have shown their efficiency.

1. Introduction

The primary challenge in effective image and signal processing is the presence of noise. Noise refers to random deviations of a signal from its true value, caused by various external or internal factors within an image. As a result, a significant portion of image processing research and techniques focuses on noise reduction and removal.
In this article, we consider intelligent filtering of two common types of noise: additive Gaussian noise and speckle noise. Additive Gaussian noise with the Gaussian probability distribution function is usually caused by the following factors [1,2]:
  • Discrete nature of light (photons), resulting in the fluctuation in the number of photons detected by an image sensor in a given period of time;
  • Imperfections in electronic circuits, including amplifiers and analog-to-digital converters;
  • Distribution of microscopic grains of metallic silver or dye particles in film emulsions while working with photographic film.
Speckle noise is a multiplicative noise, which may also contain an additive component. Speckle noise can be observed in images that are acquired using a coherent source (for example, laser, radar, and ultrasound). The cause of speckles is the microscopic variations in the surface roughness within one pixel, so when coherent radiation strikes the surface and becomes reflected, a received signal is subjected to random variations in phase and amplitude [3].
Noise filtering in images is a long-standing problem. However, it remains highly relevant today. Since the advent of digital imaging, numerous noise filtering methods have been developed. These range from a variety of spatial and frequency domain linear and nonlinear filters to filtering methods based on the use of neural networks and fuzzy logic.
One of the most efficient filters to deal with additive and speckle noise in images is the BM3D filter [4]. It is an efficient noise reduction method that operates in the frequency domain, utilizing the 3D Fourier transform on blocks constructed from statistically similar image patches. The BM3D filter has demonstrated high efficiency in reducing both Gaussian and speckle noise.
A variety of image denoising methods, including anisotropic diffusion, total variation minimization, weighted nuclear norm minimization (WNNM), trainable nonlinear reaction diffusion (TNRD), and shrinkage fields, have been explored in [5,6]. Techniques based on anisotropic diffusion, detailed in [7,8,9,10,11,12], employ a partial differential equation to iteratively smooth pixel intensities in low-gradient regions while preserving edges in high-gradient areas, effectively reducing noise while maintaining critical image features. Similarly, TNRD, introduced in [13,14], was demonstrated to be effective for filtering additive Gaussian noise in [5]. Total variation minimization, described in [15], reduces noise by minimizing the integral of the signal gradient, producing a denoised image that retains essential features. Other interesting noise filtering techniques are based on the use of shrinkage fields and nuclear norm minimization. The former one utilizes a structure called cascade of shrinkage fields (CSF) [16], offering computationally efficient denoising for high-resolution images. In turn, WNNM [17] constructs a low-rank matrix from similar image patches to estimate denoised patches, demonstrating high effectiveness in filtering additive Gaussian noise.
It is also worth mentioning such classical filters as the Lee filter [18], the Frost filter [19], and a family of order statistics-based filters [20,21,22,23]. The Lee and Frost filters are usually used to filter multiplicative speckle noise. These filters use the local statistics of an image fragment to calculate the updated intensity value of a target pixel. For example, the Lee filter uses the local mean value and weights based on the local and global variance of the image to calculate the pixel intensity. The Frost filter, in turn, uses an exponentially weighted value, with weights calculated using the local variance and mean.
Order statistics filters are based on the analysis of a variational series and the application of linear or nonlinear averaging over the intensities selected according to some special ranking criteria [24]. This family of filters includes basic filters such as the median filter and minimum and maximum filters [20,21,22]. But the main representatives of this family are based on ranking intensities in a local window around a pixel to be processed based on their closeness to the intensity of a pixel of interest in terms of values or their ranks [23,24,25]. These latter filters provide a decent ability to suppress noise and better preserve image details.
The widespread popularity of neural networks has led to their extensive adoption in various denoising methods, as well as other image processing tasks. A notable example is provided in [26], where machine learning techniques, and specifically neural networks, were employed for digital image restoration, noise removal, and object detection.
A comprehensive overview of deep learning-based image denoising methods using feedforward and U-shaped neural network architectures was presented in [27]. The study analyzed the advantages and limitations of these architectures, as well as various learning strategies employed for efficient image denoising.
Convolutional neural networks (CNNs) have captured the attention of researchers worldwide, including those working in the field of image filtering. Multiple CNN architectures were analyzed for filtering different types of noise in digital images in [28,29,30]. Among these, the Hybrid CNN [31] and the Pre-trained RLN [32] were specifically designed to address speckle noise and were reported to achieve performance comparable to, and in some cases surpassing, classical filters such as the Lee and Frost filters.
The denoising convolutional neural network (DnCNN) was proposed in [33]. DnCNN employs a deep architecture with multiple convolutional layers of various types and does not include pooling layers. The network was trained to reduce Gaussian noise in grayscale and color images, achieving results that were comparable to, and in some cases better than, those of BM3D, WNNM, TNRD, and CSF.
CNNs have also been adapted for specialized imaging domains. In [34], a deep CNN inspired by DnCNN was effectively utilized to filter a combination of additive Gaussian and Poisson noise in X-ray images obtained during the Multi-Shock (MShock) experiments conducted at the National Ignition Facility (NIF). According to the research, this approach achieved superior filtering results in terms of PSNR compared to filters such as the mean and Butterworth.
In addition to CNNs, neural networks with different architectures have also been successfully applied to denoising tasks. In [35], a multilayer perceptron (MLP) with four hidden layers, each containing 2047 neurons, was proposed for filtering additive Gaussian noise. The network was trained on a dataset designed from image fragments in the spatial domain, comprising 362 million training samples. The MLP with this architecture slightly outperformed the BM3D algorithm or showed comparable results in Gaussian noise filtering for specific test images.
Various types of neural networks, including those based on despeckling autoencoders, were evaluated for reducing speckle noise in ultrasound images in [36]. The study demonstrated that denoising solutions utilizing neural networks, such as Di-Conv-AE-Net, DGAN-Net [37,38], and D-U-NET, based on U-Net [39], outperformed widely used filtering algorithms like BM3D.
In [40], the stacked sparse denoising autoencoder (SSDA) architecture was proposed for denoising and inpainting grayscale images. This approach is based on sparse coding methods, which reconstruct images from a sparse linear combination of an overcomplete dictionary. It was demonstrated that SSDA could achieve performance comparable to the Bayes least squares with Gaussian scale mixture (BLS-GSM) method [41] and K-SVD [42].
A neural network framework named QIS-SPFT (QIS Serial–Parallel Fusion Transformer), which integrates CNN and transformer components, was employed to suppress Poisson (photon shot) noise during the QIS (quanta image sensor) imaging process was proposed in [43]. QIS-SPFT demonstrated superior performance in terms of PSNR compared to other filtering techniques, such as MLE, TD-BM3D, and QIS-Net.
In this paper, we focus on reducing the effects of speckle noise and additive Gaussian noise on digital images in the frequency domain using a complex-valued neural network.
Our idea is to perform image filtering by processing overlapping image fragments (patches) similarly how it was performed in [44,45]. However, while in that work a neural network is used to process patches taken from the spatial domain, we will focus on processing patches represented in the frequency domain using a multilayer neural network with multi-valued neurons (MLMVN). The idea behind this processing can be considered somewhat similar to the idea behind the BM3D filter. This filter performs processing by grouping similar image fragments into 3D blocks and processing them as a single unit. The idea behind this is that if patches are similar, then the noise affecting them should also be similar and can be filtered in a similar way. In our case, we would like to employ MLMVN as an intelligent processor that performs frequency domain convolution of image patches with adaptive convolutional kernels generated from the learning process.
The idea of using MLMVN as a framework for this task is based on the nature of image representation in the frequency domain, which relies on complex numbers. In turn, MLMVN is a type of neural network that shares the same architecture as the classical MLP but at the same time is based on multi-valued neurons (MVNs). MVNs, which are comprehensively described in [46], are artificial neurons whose weights, inputs, and outputs are complex numbers. This fact makes them a perfect processing unit for a vast number of problems where complex numbers are essential. This includes image processing in the frequency domain. MLMVN, along with its derivative-free backpropagation learning algorithm, was introduced in [47]. A number of studies have been conducted that build upon the concepts of MVNs and MLMVN [48,49,50,51,52]. Since their introduction, MVNs and MLMVN have demonstrated successful practical applications in various areas [53,54,55,56,57,58,59].
The structure of this paper is as follows. The Section 2 describes the process of noise filtering based on the use of a set of overlapping patches in the frequency domain. The respective algorithm is described, and the process of finding convolutional filter kernels in the frequency domain using MLMVN is presented. Then the training set design is described.
In Section 3, experimental results on filtering additive Gaussian noise and speckle noise using the MLMVN are presented. A detailed description of the learning process is provided, followed by a comparison of the filtering results obtained using our approach with those of some well-known popular filtering algorithms. Additionally, an analysis of the convolutional kernels generated from the learning process is conducted.

2. Frequency Domain Filtering of Overlapping Patches

2.1. Frequency Domain Filtering of Overlapping Patches

For some simplicity, but without loss of generality, we focus in this work on filtering grayscale images. Filtering of color images can be reduced to channel-wise processing of grayscale images, and therefore this simplification really does not affect anything.
Let f ( · ) be an image. We may consider it as a superposition of a desired component g ( · ) and a noise component q ( · ) . As we focus here on additive and multiplicative noises, the respective representations of this superposition are additive f ( · ) = g ( · ) + q ( · ) and multiplicative f ( · ) = g ( · ) × q ( · ) . The process of image filtering consists of finding an approximation of an ideal image f ^ ( · ) such that the difference between a filtered image and the ideal (noise-free) image is minimized according to a given metric. Usually, the root mean square error (RMSE) is used as this metric.
RMSE = 1 M N i = 1 M j = 1 N A ( i , j ) B ( i , j ) 2
Minimization of RMSE means maximization of the peak signal-to-noise ratio (PSNR).
PSNR = 20 log 10 L RMSE ,
where L is the maximum possible pixel intensity value for a grayscale 8-bit image L = 255 . Specifically, PSNR is usually used to evaluate the quality of filtering.
Many classical image filtering methods [18,19,20,21,22,23] usually utilize linear and nonlinear filtering in the spatial domain. However, filtering in the frequency domain can also be highly effective. For example, one of the most effective filters for Gaussian and speckle noise reduction—BM3D—processes images in the frequency domain. It is a common understanding that a visible noise component in the case of additive and multiplicative noise primarily affects higher frequencies. This can be observed in Figure 1.
Hence, filtering of this kind of noise should typically be utilized via low-pass filtering preserving low frequencies and carefully and selectively suppressing high-frequency components affected by noise. This means that we may consider the process of filtering in the frequency domain as the detection of high-frequency components affected by noise followed by their suppression. A crucial part of this process is in fact the accurate identification of the components affected by noise and their correction. It is important to keep in mind that high-frequency components also contain information about small details and object boundaries. Therefore, while we are interested in noise filtering, we need to do our best to preserve useful information. By simply suppressing all high-frequency components, we may lose important information in a filtered image.
The design of filters is a sophisticated task. Some filters work effectively for a specific type of noise but prove to be ineffective for other types. This demands either modifying an existing filter or developing a new one. Such a situation drives the search for tools that enable the efficient design of new filters by implementing complex functional dependencies between the ideal and noisy images, thereby facilitating high-quality noise removal.
We employ here a neural network as such a tool. The idea behind the use of a neural network in image filtering is based on the ability of the network to learn from its environment. As a neural network may learn a lot of things from data, there are many reasons to believe that it may learn how to detect and suppress noise while at the same time preserving image details. Neural networks have successfully been applied to image filtering for more than a decade. Successful use of MLP for noise filtering was shown, for example, in [44,45]. So far, all applications of neural networks in image filtering have been in the spatial domain. That is, in all these applications a neural network developed a certain spatial domain filter. Here we suggest using a neural network for frequency domain filtering. We would like to use a neural network as a tool that is able to determine which frequencies are affected by noise and which are not or less affected. Simultaneously a neural network should be able to synthesize (design) convolutional kernels for noise filtering in the frequency domain resulting from the learning process. It is natural to use the MLMVN [47] for solving this problem, as a complex-valued network that is suitable to work with the complex-valued data in the frequency domain.
Nowadays images typically may have a large size. Thus, on the one hand, processing a large digital image as a whole using a neural network can be a highly resource-consuming task. On the other hand, to design a robust filter through the learning process, we would need to use many patches from many various images rather than some large image(s) as a whole. Thus, our idea is to focus on filtering relatively small overlapping patches. Each patch in this case should be filtered as a whole, while a resulting image should be created by averaging over all the overlapping pixels.
Hence, we focus on training MLMVN to design frequency domain convolutional kernels based on taking the Fourier transform of a noisy patch from an artificially corrupted image as an input and the Fourier transform of a respective clean patch from a noise-free image as a desired output. This approach also simplifies the adaptation of a neural network to specific data, as fragments contain significantly fewer details compared to the entire image. While processing local regions, the global context of an image is preserved through the overlapping areas of the fragments. Since fragments are much smaller than an entire image, their processing requires considerably fewer computational resources. Additionally, each fragment can be treated as an independent unit, enabling the parallel processing of multiple fragments.
Thus, our idea is basically to reconstruct the Fourier transform of a noise-free patch from the Fourier transform of its noisy version using a neural network whose input is the Fourier transform of a noisy patch and desired output is the Fourier transform of a corresponding noise-free patch. To create a respective representative learning set, many clean images should be corrupted by noise, noisy patches should be randomly selected from each image, their Fourier transforms should be used as inputs,  corresponding clean patches should be selected, starting with the same coordinates from the respective clean images, and their Fourier transforms should be used as desired outputs.
We suggest employing MLMVN with a single hidden layer and an output layer with the same number of output neurons as the number of Fourier coefficients in a respective patch’s Fourier transform. With this topology of MLMVN, its two layers of neurons perform the following tasks. Every neuron in a single hidden layer develops a frequency domain convolutional kernel through the learning process. Therefore, each neuron in a single hidden layer performs a frequency domain convolution, multiplying component-wise its weights by the respective Fourier transform coefficients of a patch to be processed.
F 1 ( F ( w ) F ( p ) ) = w p
Each neuron in the output layer estimates a respective Fourier coefficient of a noise-free patch based on the outputs of all hidden layer neurons.
The selection of patches for constructing the training set and the learning process is described below in Section 2.2. After a learning set is created, the learning process should start. After a neural network develops its weights through the learning process, it should be used to filter images. To extract patches from an image, we use a window of size m × n . By moving this window across the image using steps s x and s y (where s x corresponds to the step along the x-axis and s y to the step along the y-axis), we extract the intensities contained within the window into a patch.
As a result of this operation, a set of overlapping patches is generated if s x < m and s y < n . This procedure is illustrated in Figure 2.
For the reader’s convenience, we provide an algorithm (Algorithm 1) for extracting a set of overlapping patches from an image to be processed. It should also be mentioned that the provided algorithm describes the process of building a set of patches for the actual image filtering process.
Algorithm 1: Extracting a set of overlapping patches from the input image
  1:
Input:
  2:
I: An image of size M × N
  3:
m: The patch width
  4:
n: The patch height
  5:
s x : The step size along x-axis
  6:
s y : The step size along y-axis
  7:
Output:
  8:
P: A set of overlapping patches
  9:
Procedure:
10:
P
11:
for  i 1 to M m by s x  do
12:
    for  j 1 to N n by s y  do
13:
        Extract a patch p: p I [ i : i + m , j : j + n ]
14:
         P P { p }
15:
    end for
16:
end for
17:
return P
Since we perform processing in the frequency domain, after a set of overlapping patches P is built, we need to create P F —a set of Fourier transforms of these patches—in the following way:
P F = 1 n m F ( p ) | p P .
The vectorized Fourier transform p F P F should be used as an input sample for MLMVN. As was mentioned above, we employ a shallow MLMVN with a k q k topology for processing frequency domain data. This means that a neural network consists of an input layer with k inputs, a single hidden layer with q neurons, and an output layer with k neurons. The number of inputs k, which is also equal to the number of neurons in the output layer, depends on the size of the image fragments being processed ( k = m n ). The number of hidden neurons q should be determined based on experimental testing.
It is important to make a remark about activation functions, which we use in our network. Two activation functions have been considered so far for MVNs—discrete and continuous ones. But both these functions project a weighted sum of MVNs onto the unit circle [46]. In MLMVN, which is used here, all hidden neurons employ a standard continuous MVN activation function f ( z ) = z | z | . At the same time, neither discrete nor continuous activation functions producing an output located on the unit circle can be used for output layer neurons. As our goal is to use MLMVN to estimate the Fourier transform of a clean patch, output neurons should produce an output not necessarily located on the unit circle. This requires the use of a different type of activation function capable of producing an output with an arbitrary (not necessarily unitary) magnitude. We used three such activation functions in our experiments: a linear activation function (when a weighted sum becomes an output), a complex-valued sigmoid activation function, and complex-valued hyperbolic tangent activation function. The respective results and analysis of the network behavior with all three of these activation functions are presented in Section 3.
As a result of processing elements of P F with a neural network, we obtain a set of vectorized filtered overlapping patches P ˜ F in the frequency domain. Elements of P ˜ F can be used to obtain a set of filtered overlapping patches in the spatial domain in the following way:
P ˜ = crop n m F 1 ( p ˜ F ) | p ˜ F P ˜ F .
The goal of cropping the resulting image patch in the spatial domain is to remove possible unwanted artifacts and distortions that occur while processing the boundary regions of an image. The set P ˜ is used for the “synthesis” of a filtered image.
It is important to note that after processing a patch with MLMVN, we restore a zero-frequency coefficient of the Fourier transform by setting it equal to a zero-frequency coefficient of the respective input (unprocessed) patch. This step is valid because our modeled noise has a zero mean. Thus, we can preserve a respective mean over a patch that is being processed. This is important to avoid a shift in patch intensities, which would occur if the mean value were modified.
For reconstructing an image from P ˜ , we use a window of size t × r . By “moving” this window starting from the top-left corner with vertical step t in the range [ 1 , M t ] and with horizontal step r in the range [ 1 , N r ] , and collecting all patches that lie inside the window at each step, we are able to reconstruct an image fragment of size t × r . This should be carried out by combining the intensities of these overlapping patches. Intensities in areas where patches overlap are restored by applying an aggregation function. In this work, we employed median and mean aggregation functions. Below, an algorithm (Algorithm 2) for the “synthesis” of a filtered image from processed overlapping patches in the frequency domain is provided.
In general, the image noise filtering process proposed in this paper can be illustrated in Figure 3.
The algorithm for the complete process of filtering a noisy image is outlined here (Algorithm 3).
Algorithm 2: “Synthesis” of a filtered image from processed overlapping patches
  1:
Input:
  2:
P ˜ : A set of filtered overlapping patches
  3:
m: The width of a patch
  4:
n: The height of a patch
  5:
t: The width of a window
  6:
r: The height of a window
  7:
a g r : An aggregation function
  8:
Output:
  9:
I filtered : A filtered image
10:
Procedure:
11:
for  i 1 to M t by t do
12:
    for  j 1 to N r by r do
13:
         R
14:
        Collect the patches from P ˜ :
15:
            R p ( x , y ) p ( x , y ) P ˜ , ¬ ( x + m < i x > i + t y + n < j y > j + r )
16:
        Set the image intensities in the window:
17:
            I filtered [ i : i + t , j : j + r ] a g r p ( x , y ) p ( x , y ) R
18:
    end for
19:
end for
20:
return  I filtered

2.2. Organization of the Learning Process

The learning process we utilized in this work for training MLMVN is based on batch learning with validation. The main advantage of learning with validation is that it makes it possible to verify whether a neural network has developed a generalization capability. As learning with validation is based on the minimization of the validation error, it stops when a network is capable of generalizing with a desired accuracy. This also helps to avoid overfitting, which may often affect learning based on the minimization of the learning error (that is, the error on the learning set). Overfitting may occur from repeated attempts to “memorize” a learning set by achieving a low learning error. But the actual goal of learning is to develop a generalization capability (that is, to deal with the data that were not used to adjust the weights), not to “memorize” a learning set. This leads to improvement in robustness. In the learning process with validation, a dataset is divided into three subsets: a learning (training) subset, a validation subset, and a test subset. The training subset is used to adjust the weights, while the validation subset is used to verify whether a desired generalization capability has been achieved. Then the test subset is used to verify the results of the learning process.
Algorithm 3: The process of filtering noise in a digital image using MLMVN
  1:
Input:
  2:
I noisy : A noisy image
  3:
m: The width of a patch
  4:
n: The height of a patch
  5:
Output:
  6:
I filtered : A filtered image
  7:
Procedure:
  8:
Normalize the image: I ^ noisy 1 255 I noisy
  9:
Split the image I ^ noisy into patches forming the set P: P split ( I ^ noisy )
10:
P ˜
11:
for  p P  do
12:
    Convert from the spatial to the frequency domain: p F F ( p )
13:
    Normalize p F : p ^ F 1 n m p F
14:
    Process p ^ F with the MLMVN: p ^ F out MLMVN ( p ^ F )
15:
    Restore the zero-frequency coefficient: p ^ F out [ 0 , 0 ] p ^ F [ 0 , 0 ]
16:
    Denormalize p ^ F out : p F out ( n m ) p ^ F out
17:
    Convert p F out to the spatial domain: p out F 1 ( p F out )
18:
     P ˜ P ˜ { p out }
19:
end for
20:
Synthesize the filtered image I filtered using set P ˜ : I filtered synthesize ( P ˜ )
21:
return  I filtered
Model training was performed separately for each type of noise. To design training, validation, and test sets, a set of grayscale images of different sizes was used. It was split into three subsets, S t (training), S v (validation), and S f (test), such that S t S v S f = . The training dataset consists of pairs of vectorized Fourier transforms of noisy and corresponding clean image patches. The training dataset was built using set S t , which comprised 300 grayscale images. It is important to note that as a part of preprocessing, we normalize the images by dividing their intensities by 255, so they have a range [ 0 , 1 ] after normalization. This is important because a neural network learns faster and generalizes better when it deals with normalized data. For each image I S t , an image I noisy was created by applying noise (additive Gaussian noise or multiplicative speckle noise). Additive Gaussian noise was modeled with the following levels, 0.1 σ , 0.2 σ , and 0.3 σ , while multiplicative speckle noise was modeled with levels 0.2 σ , 0.3 σ , and 0.4 σ , where σ is the standard deviation of the ideal (clean) image. After applying the corresponding noise, h random patches of size m × n were picked from the clean image I. In our experiments, we selected 200 patches per image. The same number of patches, of the same size and at the same spatial coordinates, were picked from each corresponding noisy image I noisy . As a result, two sets of image fragments were created:
P ideal = C ( x j , y j ) ( i ) | i = 1 , l ¯ , j = 1 , h ¯
and
P noisy = D ( x j , y j ) ( i ) | i = 1 , l ¯ , j = 1 , h ¯
where ( x j , y j ) are spatial coordinates of top-left corner of patch and l is the number of images in S t . Each element of the P ideal and P noisy sets was converted from the spatial domain to the frequency domain by computing a two-dimensional Fourier transform and normalized using the factor 1 n m . After vectorizing each image fragment in the frequency domain, the following sets were built:
P F ideal = vec 1 n m F C T | C P ideal
and
P F noisy = vec 1 n m F D T | D P noisy
Thus, the training set for MLMVN was formed from elements of P F ideal and P F noisy in the following way:
L = T i in , T i out | i = 1 , l h ¯ , T i in P F noisy , T i out P F ideal
The validation dataset was built using g full-size images from S v . Additive Gaussian or multiplicative speckle noise was applied to each image from S v , thereby creating a set of noisy images S v noisy similarly to how noise was added to create a training set. Using pairs of ideal and noisy images, the validation dataset was constructed in the following way:
V = V i in , V i out | i = 1 , g ¯ , V i in S v noisy , V i out S v
In our work, MLMVN training was performed using the batch learning algorithm proposed in [60,61]. This algorithm is based on a derivative-free approach and enables the correction of neuron weights across multiple learning samples simultaneously (that is, across an entire batch).
As noted above, correction of the network weights was performed using samples from the training dataset L. During our experiments, we used a training dataset with 60,000 learning samples created from 300 grayscale images (as 200 patches were randomly selected from each of the 300 images). To employ the batch learning algorithm, learning samples were grouped into batches containing b samples each. We employed a batch size of 20,000 learning samples. The justification for this choice is provided in Section 3.2. Our learning process is based on the maximization of the mean value of the validation PSNR evaluated over the images from the validation set. During our experiments, we employed a validation set, which consists of five pairs of noisy and clean grayscale images.
After each batch learning step was completed, a validation step was performed. During the validation, image pairs from the validation dataset V were processed using the algorithm described in the previous section of this article (see Algorithm 3). The PSNR of the filtered image relative to the clean image was computed for each pair of images from V. The average PSNR value across all elements from V was compared with the threshold PSNR value, and the learning process either was stopped if the current average validation PSNR reached or exceeded a pre-determined threshold value or continued if it was below it. The algorithm for the learning process is outlined here (Algorithm 4)
The final performance evaluation of the proposed filtering approach was performed using the images from the S f set. It is important to note again that these images were not used in the design of either the training set or the validation set.
Algorithm 4: The MLMVN learning process
  1:
Input:
  2:
L: A training set
  3:
V: A validation set
  4:
PSNR v : The desired PSNR on the validation set
  5:
PSNR b : The besired PSNR on a batch
  6:
n: Tha size of a batch
  7:
iterations_limit: The maximum number of iterations per batch
  8:
Output:
  9:
A trained MLMVN
10:
Procedure:
11:
Build a set of batches B: B { b 1 , , b k b i L , b i b j = when i j , | b i | = n }
12:
PSNR v c u r 0
13:
repeat
14:
    for  b B  do
15:
         PSNR b c u r 0
16:
         iteration 0
17:
        while  PSNR b c u r < PSNR b  and  iteration < iteration_limit  do
18:
           Update the MLMVN weights
19:
           Process the samples from b with the MLMVN: b ˜ { MLMVN ( s i ) s i b }
20:
            PSNR b c u r mean ( { PSNR ( s ˜ i , s i ) s ˜ i b ˜ , s i b } )
21:
            iteration iteration + 1
22:
        end while
23:
        Process the images from V with the MLMVN: V ˜ { MLMVN ( v i ) v i V }
24:
         PSNR v c u r mean ( { PSNR ( v ˜ i , v i ) v ˜ i V ˜ , v i V } )
25:
        if  PSNR v c u r > = PSNR v  then
26:
           break
27:
        end if
28:
    end for
29:
until  PSNR v c u r > = PSNR v
30:
return A trained MLMVN

3. Results

3.1. Convolutional Kernels Resulting from Learning and Their Analysis

As mentioned above, we employ an MLMVN topology of k q k for filtering additive Gaussian noise and multiplicative noise in digital images. Here k is the number of input and output neurons and q is the number of neurons in a single hidden layer.
Let us consider a neuron s from a single hidden layer of our MLMVN. Let x = ( x 0 , x 1 , x 2 , , x k ) be the input to the neuron ( x 0 is a pseudo input corresponding to the bias and it always equals 1) and w = ( w 0 , w 1 , , w k ) be the neuron’s weights. The output of the neuron can be expressed as
O = f i = 0 k x i w i ,
or equivalently as
O = f x 0 w 0 + i = 1 k x i w i = f x 0 w 0 + R .
In this expression,
R = i = 1 k x i w i
Since the vector x results from the vectorization of an image fragment of dimensions m × n in the frequency domain, it can also be represented as a matrix X F , which is in its original shape. Similarly, the weights of the neuron (excluding the bias w 0 ) can be represented as a matrix W F . Consequently, the last formula can be rewritten as
R = X F W F ,
where the operator ⊙ denotes elementwise multiplication. According to the Convolution Theorem, this expression can be further expanded as
R = X F W F = F X s W s ,
where X s = F 1 ( X F ) and W s = F 1 ( W F ) , where ★ denotes the convolution operation. Based on this interpretation, the neurons in the hidden layer of MLMVN can be considered as convolutional processors (with their weights acting as convolutional kernels) and simultaneously as analyzers of the resulting convolutions. On the other hand, the neurons in the output layer act as ultimate processing units that integrate the outputs of neurons from the preceding layer, producing the result of filtering in the frequency domain.
Let us discover in detail how a neural network processes an image fragment. Each neuron in a single hidden layer actually performs filtering by utilization of the frequency domain convolution with a kernel K. Let I s be an image fragment in the spatial domain to be filtered, and I F is its representation in the frequency domain, that is, the Fourier transform of I s . Let O F be the output of MLMVN for input I F . The output O F is nothing but an approximation of the frequency domain representation of the processed image fragment I s . Its spatial domain representation should be obtained by applying the inverse Fourier transform as follows:
O s = F 1 ( O F ) .
If we define K F = O F I F and K s = F 1 ( K F ) , we can express the relationship as follows:
O F = I F K F = F ( I s ) F ( K s ) .
According to the Convolution Theorem,
F ( I s K s ) = F ( I s ) F ( K s ) .
Applying the inverse Fourier transform to both sides of the last equation, we obtain
I s K s = O s ,
where K s is the convolutional kernel in the spatial domain. To discover what kind of filters resulted from the learning process, we need to discover the convolutional kernels generated from this process. Let us consider the power spectra | K F | of some randomly selected convolutional kernels in the frequency domain resulting from the learning process as the weights of the respective neurons. These power spectra are just absolute values of the respective weights, which should be reshaped into matrices. They are shown in Figure 4.
Analysis of the power spectra shown in Figure 4 makes it possible to understand the frequency characteristics of filters resulting from the learning process. Figure 4a,b,d show that a neural network develops mostly various low-pass filters to reduce additive Gaussian noise. Basically Figure 4c also shows a low-pass filter, just a bit more sophisticated. It is clear from the high–low-frequency area “hills” in these examples that we are dealing with low-pass filters. However, it is important to mention that small “hills” appear in the medium- and high-frequency areas, which means that while these filters care about preserving low frequencies, they deal with high frequencies carefully, selectively (and, evidently, adaptively) preserving some of them. Analyzing Figure 4e–h, we may conclude that speckle noise filtering is a more sophisticated task, and a neural network adapts to this task in a different way. While the filters shown in Figure 4e,h are low-pass filters with some high-frequency areas carefully preserved, the filters shown in Figure 4f,g are much more sophisticated. They preserve five various-frequency areas each (this is clearly visible from the five “hills”), suppressing other frequencies at the same time. This means that a neural network adapts to specific noise by suppressing selective frequencies and preserving other frequencies (not only low frequencies, but preserving some specific frequencies across the bound), which might be important for preserving image details.

3.2. Batch Size and the Learning Results

The training dataset, as described earlier, consists of pairs of noisy and clean image fragments (patches) in the frequency domain. The learning process continued until a certain PSNR level was reached for the validation set, regardless of the learning error (that is, the error on the learning set) at that moment.
Another important preprocessing step is to perform mirroring (reflection padding) on a processed image. This step helps to avoid boundary effects while processing an image by ensuring that boundary intensities are processed in the same way as the rest of the intensities.
As was mentioned above, to organize the learning process, we used a batch learning algorithm [60,61]. In our work, we conduct experiments with various batch sizes. After performing many experiments with patches of various sizes, we found that a patch size of 8 × 8 is optimal as the one leading to better results. Thus, a patch size of 8 × 8 determines that our neural network should have 64 inputs and 64 outputs. To discover how a batch size influences the quality of filtering, the following experiment was conducted. MLMVN with a topology of 64-256-64 was trained to filter additive Gaussian noise with a standard deviation of 0.2 σ using different batch sizes. The network utilized a standard activation function for MVNs with continuous output in the hidden layer neurons, while an identity activation function was used in the output layer neurons. The learning epoch threshold was set to 3, while the iteration limit per batch was set to 100. The results of processing images from the test set with this MLMVN trained using various batch sizes are provided in Table 1. For convenience, we also provide the ideal images from the test dataset that were used for the evaluation of the MLMVN filtering capabilities in Figure 5. According to our observations, larger batch sizes allow for achieving better results. This can be explained by the well-known fact that LLS-based methods for solving overdetermined systems of linear algebraic equations, which are the key elements of batch learning, show more accurate results for systems that are more overdetermined to some level, that is, for the systems where the number of equations is significantly larger than the number of unknowns. The number of equations in this case is determined by the number of samples in a respective batch.
During our experiments, we considered various activation functions for the neurons in the hidden and output layers (Table 2). It is well known that the main role of an activation function is to limit the neuron’s output range. In our case, an expected output of MLMVN is a filtered image patch in the frequency domain, whose magnitude should be in the range [ 0 , 1 ] , as we work with normalized images and their normalized Fourier transforms. Therefore, the important role of an activation function in our case is to keep the magnitude of the resulting Fourier transform coefficients within the range [ 0 , 1 ] (Figure 6). The only exception is the linear (“identity”) activation function, which simply passes the computed weighted sums to the output without any modification.
When performing filtered image “synthesis” from a collection of filtered overlapping patches (Algorithm 2), an aggregation function should be used to process the overlapping pixels in the patches. In this work, we employed median and mean, both taken over the overlapping values as aggregation functions. According to the results of our experiments, the median aggregation function performs slightly better.
As part of our experiments, we trained several MLMVN models to filter additive Gaussian noise and speckle noise. As was already mentioned, we employed the MLMVN topology 64-1024-64 because it has shown the best results. This topology processes 8 × 8 image patches in the frequency domain. MLMVN-G-1, MLMVN-G-2, and MLMVN-G-3 were trained for denoising images with additive Gaussian noise, while MLMVN-S-1, MLMVN-S-2, and MLMVN-S-3 were trained for denoising images with multiplicative speckle noise. In all MLMVNs neurons in the hidden layer employed a standard continuous MVN activation function. The output layers of MLMVN-G-1 and MLMVN-S-1 employed the identity activation function, MLMVN-G-2 and MLMVN-S-2 employed the complex-valued sigmoid activation function, and MLMVN-G-3 and MLMVN-S-3 employed the complex hyperbolic tangent activation function. Figure 7 shows convergence curves for all these MLMVN models. These curves represents the mean PSNR over the validation set vs. the number of epochs/batches.

3.3. Simulation Results for Filtering Additive Gaussian Noise and Speckle Noise

Below, we provide the results of filtering noisy grayscale images corrupted by additive Gaussian noise (Table 3, Table 4 and Table 5) and multiplicative speckle noise (Table 6, Table 7 and Table 8) at various levels using MLMVN and popular filtering methods for comparison. For the BM3D filter, we considered multiple values of the noise standard deviation parameter because, in real-world tasks, this information is usually unknown. For both Lee and Frost filters, we used a filter window size of 3 × 3. The damping coefficient for the Frost filter was set to 2. We utilized MLMVN-G-1, MLMVN-G-2, and MLMVN-G-3, and MLMVN-S-1, MLMVN-S-2, and MLMVN-S-3 for filtering additive Gaussian noise and speckle noise, respectively. The evaluation of the MLMVN filtering capabilities was performed using images from the test dataset (Figure 5).
Figure 8 and Figure 9 present the results of filtering additive Gaussian noise and speckle noise, respectively, using the MLMVN and the BM3D filter. Images (a–d) illustrate the cases where the MLMVN achieved a higher PSNR than the BM3D filter. Images (e–h) show the cases where the PSNR values for both methods are approximately equal. Finally, images (i–l) depict the cases where the BM3D filter outperformed the MLMVN in terms of PSNR.

4. Discussion

Existing studies on the use of neural networks in image filtering have proven the effectiveness of this approach. Thus, MLPs, CNNs, AEs, and other types of neural networks have been utilized for image denoising and restoration. These results can be explained by the ability of neural networks to model complex nonlinear functional relationships between inputs and outputs.
In this paper, we considered the application of MLMVN as a framework for filtering additive Gaussian noise and multiplicative speckle noise in the frequency domain. The results obtained in our study demonstrate that MLMVN is effective for solving such tasks. MLMVN proved its efficiency in the development of frequency domain convolutional kernels generated from the learning process and processing overlapping fragments (patches) of images. This approach enables the use of a neural network with fewer neurons and parameters than, for example, in [35,44,45] and facilitates parallelism in programmatic implementation.
Our work also demonstrates the effectiveness of the batch learning algorithm in the training process. Noise filtering results obtained using MLMVN were presented and compared with other methods. The performance of MLMVN was found to be comparable to that of the BM3D, Lee, and Frost filters. Moreover, for some small detailed images our approach performed better because it better preserves small details. For additive Gaussian noise our approach works better for more heavily corrupted images. At the same time, our approach shows better results for speckle noise filtering than traditionally used filters. These findings indicate that MLMVN has potential for solving a wide range of image processing problems, especially those naturally suited to frequency domain analysis. We have also shown that multi-valued neurons may successfully employ the complex-valued logistic and the complex-valued hyperbolic tangent activation functions. This expands the MVNs’ functionality and makes it possible to utilize them for input/output mappings whose outputs are not necessarily located on the unit circle.
The results presented in this paper create a good background for further work on intelligent image processing in the frequency domain using MLMVN. For example, Poisson noise filtering and filtering of speckle noise with both multiplicative and additive components would be attractive subjects for further work. It would also be attractive and useful to discover possibilities of restoration of blurred images using a similar approach, that is, utilizing deconvolution in the frequency domain using MLMVN.

Author Contributions

Conceptualization, I.A.; methodology, I.A. and Y.T.; software, Y.T.; validation, I.A. and Y.T.; formal analysis, I.A. and Y.T.; investigation, I.A. and Y.T.; resources, I.A.; data curation, I.A. and Y.T.; writing—original draft preparation, I.A. and Y.T.; writing—review and editing, I.A. and Y.T.; visualization, Y.T.; supervision, I.A.; project administration, I.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The source code of the core components used across the experiments is available at URL https://drive.google.com/file/d/1xrJT8Tm7sogk7saxxCPU9-VgdAndpibx/view?usp=sharing (accessed on 23 July 2025). The experiments on learning and filtering of additive Gaussian noise and speckle noise are available at URL https://drive.google.com/file/d/1_TADbGTlsA3AvUZyZ_LMeQeDIiIk2bSy/view?usp=sharing (accessed on 23 July 2025). The archives ls_007_gaussian.tar https://drive.google.com/file/d/1Ez-49U6BLvFq2hoaMbjolqLjMGDohXif/view?usp=sharing (accessed on 23 July 2025) and ls_006_speckle.tar https://drive.google.com/file/d/1wfhiu_6d_o7hibqkUbaw-AWFaN6rMrWR/view?usp=sharing (accessed on 23 July 2025) contain the training and validation sets used for MLMVN training for the filtering of additive Gaussian and speckle noise, respectively. The test image sets used for the final performance evaluation are available at URLs https://drive.google.com/file/d/1sUyAHp_d6almxuz0UMmOG3Po5VS9IPEA/view?usp=sharing (accessed on 23 July 2025) and https://drive.google.com/file/d/1yf2NrGYUZXt80D5zWJUgWFZCXPM1CC7J/view?usp=sharing (accessed on 23 July 2025).

Acknowledgments

All experimental work presented in this paper was performed on the computers from the Kakos Center for Scientific Computing at the Kakos School of Arts and Sciences, Manhattan University.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bovik, A.C. Handbook of Image and Video Processing; Academic Press Series in Communications, Networking, and Multimedia; Academic Press: San Diego, CA, USA, 2000; pp. 325–335. ISBN 978-0-12-119790-2. [Google Scholar]
  2. Fine Art Photography. Understanding Film Grain and Digital Noise in Photography. Available online: https://kladoff.net/blog/film-grain-digital-noise.html?srsltid=AfmBOopX0m9HoWD-rt_S5TVjQpQTaUOm8tbhF7n7EtpAmoOV-qYRj9Dy (accessed on 23 July 2025).
  3. Goodman, J.W. Statistical Optics; Wiley Series in Pure and Applied Optics; J. Wiley & Sons: New York, NY, USA, 1985; pp. 347–351. ISBN 978-0-471-01502-4. [Google Scholar]
  4. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  5. Mafi, M.; Martin, H.; Cabrerizo, M.; Andrian, J.; Barreto, A.; Adjouadi, M. A Comprehensive Survey on Impulse and Gaussian Denoising Filters for Digital Images. Signal Process. 2019, 157, 236–260. [Google Scholar] [CrossRef]
  6. Yuan, J.; Wu, B.; Yuan, Y.; Huang, Q.; Chen, J.; Ren, L. Speckle Noise Reduction in SAR Images Ship Detection. In Proceedings of the SPIE, Remote Sensing of the Ocean, Sea Ice, Coastal Waters, and Large Water Regions 2012, Edinburgh, UK, 19 September 2012; SPIE: Bellingham, WA, USA, 2012; Volume 8532, p. 853210. [Google Scholar]
  7. Catté, F.; Lions, P.-L.; Morel, J.-M.; Coll, T. Image Selective Smoothing and Edge Detection by Nonlinear Diffusion. SIAM J. Numer. Anal. 1992, 29, 182–193. [Google Scholar] [CrossRef]
  8. Lin, Z.; Shi, Q. An Anisotropic Diffusion PDE for Noise Reduction and Thin Edge Preservation. In Proceedings of the 10th International Conference on Image Analysis and Processing, Venice, Italy, 27–29 September 1999; IEEE Computer Society: Washington, DC, USA, 1999; pp. 102–107. [Google Scholar]
  9. Prasath, V.B.S.; Singh, A. Well-Posed Inhomogeneous Nonlinear Diffusion Scheme for Digital Image Denoising. J. Appl. Math. 2010, 2010, 763847. [Google Scholar] [CrossRef]
  10. Chao, S.-M.; Tsai, D.-M. An Improved Anisotropic Diffusion Model for Detail- and Edge-Preserving Smoothing. Pattern Recognit. Lett. 2010, 31, 2012–2023. [Google Scholar] [CrossRef]
  11. Wang, Y.Q.; Guo, J.; Chen, W.; Zhang, W. Image Denoising Using Modified Perona–Malik Model Based on Directional Laplacian. Signal Process. 2013, 93, 2548–2558. [Google Scholar] [CrossRef]
  12. Xu, J.; Jia, Y.; Shi, Z.; Pang, K. An Improved Anisotropic Diffusion Filter with Semi-Adaptive Threshold for Edge Preservation. Signal Process. 2016, 119, 80–91. [Google Scholar] [CrossRef]
  13. Chen, Y.; Pock, T. Trainable Nonlinear Reaction Diffusion: A Flexible Framework for Fast and Effective Image Restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1256–1272. [Google Scholar] [CrossRef]
  14. Chen, Y.; Wei, Y.; Pock, T. On Learning Optimized Reaction Diffusion Processes for Effective Image Restoration. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 5261–5269. [Google Scholar]
  15. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear Total Variation Based Noise Removal Algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  16. Schmidt, U.; Roth, S. Shrinkage Fields for Effective Image Restoration. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 2774–2781. [Google Scholar]
  17. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted Nuclear Norm Minimization with Application to Image Denoising. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 2862–2869. [Google Scholar]
  18. Lee, J.-S. Digital Image Enhancement and Noise Filtering by Use of Local Statistics. IEEE Trans. Pattern Anal. Mach. Intell. 1980, PAMI-2, 165–168. [Google Scholar] [CrossRef]
  19. Frost, V.S.; Stiles, J.A.; Shanmugan, K.S.; Holtzman, J.C. A Model for Radar Images and Its Application to Adaptive Digital Filtering of Multiplicative Noise. IEEE Trans. Pattern Anal. Mach. Intell. 1982, PAMI-4, 157–166. [Google Scholar] [CrossRef]
  20. Astola, J.; Kuosmanen, P. Fundamentals of Nonlinear Digital Filtering, 1st ed.; CRC Press: Boca Raton, FL, USA, 2020; ISBN 978-1-003-06783-2. [Google Scholar]
  21. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 4th ed.; Pearson Education: New York, NY, USA, 2018; pp. 317–368. ISBN 978-1-292-22304-9. [Google Scholar]
  22. Pitas, I.; Venetsanopoulos, A.N. Order Statistics in Digital Image Processing. Proc. IEEE 1992, 80, 1893–1921. [Google Scholar] [CrossRef]
  23. Kim, V.; Yaroslavskii, L. Rank Algorithms for Picture Processing. Comput. Vis. Graph. Image Process. 1986, 35, 234–258. [Google Scholar] [CrossRef]
  24. Bovik, A.; Huang, T.; Munson, D. A Generalization of Median Filtering Using Linear Combinations of Order Statistics. IEEE Trans. Acoust. Speech Signal Process. 1983, 31, 1342–1350. [Google Scholar] [CrossRef]
  25. Zamperoni, P. Some Adaptive Rank Order Filters for Image Enhancement. Pattern Recognit. Lett. 1990, 11, 81–86. [Google Scholar] [CrossRef]
  26. Khammar, M.; Ashraff, M.; Bhasha, M.Y.; Rakesh, K.V. Visual Intelligence: Machine Learning Approaches to Image Filtering and Identification. Int. J. Sci. Res. Eng. Manag. 2024, 8, 1–12. [Google Scholar] [CrossRef]
  27. Jiang, B.; Li, J.; Lu, Y.; Cai, Q.; Song, H.; Lu, G. Eficient Image Denoising Using Deep Learning: A Brief Survey. Inf. Fusion 2025, 118, 103013. [Google Scholar] [CrossRef]
  28. Ilesanmi, A.E.; Ilesanmi, T.O. Methods for Image Denoising Using Convolutional Neural Network: A Review. Complex Intell. Syst. 2021, 7, 2179–2198. [Google Scholar] [CrossRef]
  29. Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C.-W. Deep Learning on Image Denoising: An Overview. Neural Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef]
  30. Bodhale, V.; Vijayalakshmi, M.; Chopra, S. An Efficient Image Denoising Using Convolutional Neural Network. In Proceedings of the International Conference on Communication and Computational Technologies, ICCCT 2024, Jaipur, India, 8–9 January 2024; Lecture Notes in Networks and Systems. Springer: Singapore, 2025; pp. 15–26. [Google Scholar] [CrossRef]
  31. Feng, X.; Huang, Q.; Li, X. Ultrasound Image De-Speckling by a Hybrid Deep Network with Transferred Filtering and Structural Prior. Neurocomputing 2020, 414, 346–355. [Google Scholar] [CrossRef]
  32. Kokil, P.; Sudharson, S. Despeckling of Clinical Ultrasound Images Using Deep Residual Learning. Comput. Methods Prog. Biomed. 2020, 194, 105477. [Google Scholar] [CrossRef]
  33. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed]
  34. Levesque, J.M.; Merritt, E.C.; Flippo, K.A.; Rasmus, A.M.; Doss, F.W. Neural Network Denoising of X-Ray Images from High-Energy-Density Experiments. Rev. Sci. Instruments 2024, 95, 063508. [Google Scholar] [CrossRef] [PubMed]
  35. Burger, H.C.; Schuler, C.J.; Harmeling, S. Image Denoising: Can Plain Neural Networks Compete with BM3D? In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 2392–2399. [Google Scholar]
  36. Karaoğlu, O.; Bilge, H.Ş.; Uluer, İ. Removal of Speckle Noises from Ultrasound Images Using Five Different Deep Learning Networks. Eng. Sci. Technol. Int. J. 2022, 29, 101030. [Google Scholar] [CrossRef]
  37. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  38. Chen, Z.; Zeng, Z.; Shen, H.; Zheng, X.; Dai, P.; Ouyang, P. DN-GAN: Denoising Generative Adversarial Networks for Speckle Noise Reduction in Optical Coherence Tomography Images. Biomed. Signal Process. Control 2020, 55, 101632. [Google Scholar] [CrossRef]
  39. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. ISBN 978-3-319-24573-7. [Google Scholar]
  40. Xie, J.; Xu, L.; Chen, E. Image Denoising and Inpainting with Deep Neural Networks. In Proceedings of the 26th International Conference on Neural Information Processing Systems (NIPS 2012), Lake Tahoe, NV, USA, 3–6 December 2012; Curran Associates, Inc.: Red Hook, NY, USA, 2012; pp. 341–349. [Google Scholar]
  41. Portilla, J.; Strela, V.; Wainwright, M.J.; Simoncelli, E.P. Image Denoising Using Scale Mixtures of Gaussians in the Wavelet Domain. IEEE Trans. Image Process. 2003, 12, 1338–1351. [Google Scholar] [CrossRef]
  42. Mairal, J.; Elad, M.; Sapiro, G. Sparse Representation for Color Image Restoration. IEEE Trans. Image Process. 2008, 17, 53–69. [Google Scholar] [CrossRef]
  43. Wang, G.; Gao, Z. Noise Suppressed Image Reconstruction for Quanta Image Sensors Based on Transformer Neural Networks. J. Imaging 2025, 11, 160. [Google Scholar] [CrossRef]
  44. Burger, H.C.; Schuler, C.J.; Harmeling, S. Image Denoising with Multi-Layer Perceptrons, Part 1: Comparison with Existing Algorithms and with Bounds. arXiv 2012, arXiv:1211.1544. [Google Scholar] [CrossRef]
  45. Burger, H.C.; Schuler, C.J.; Harmeling, S. Image Denoising with Multi-Layer Perceptrons, Part 2: Training Trade-Offs and Analysis of Their Mechanisms. arXiv 2012, arXiv:1211.1552. [Google Scholar] [CrossRef]
  46. Aizenberg, I. Complex-Valued Neural Networks with Multi-Valued Neurons; Studies in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2011; Volume 353, pp. 55–131. ISBN 978-3-642-20352-7. [Google Scholar]
  47. Aizenberg, I.; Moraga, C. Multilayer Feedforward Neural Network Based on Multi-Valued Neurons (MLMVN) and a Backpropagation Learning Algorithm. Soft Comput. 2007, 11, 169–183. [Google Scholar] [CrossRef]
  48. Chen, J.-P.; Wu, S.-F.; Lee, S.-J. Modified Learning for Discrete Multi-Valued Neuron. In Proceedings of the 2013 International Joint Conference on Neural Networks (IJCNN), Dallas, TX, USA, 4–9 August 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1–6. [Google Scholar]
  49. Grasso, F.; Luchetta, A.; Manetti, S. A Multi-Valued Neuron Based Complex ELM Neural Network. Neural Process. Lett. 2018, 48, 389–401. [Google Scholar] [CrossRef]
  50. Pavaloiu, I.B.; Dragoi, G.; Vasile, A. Gradient-Descent Training for Phase-Based Neurons. In Proceedings of the 2014 18th International Conference on System Theory, Control and Computing (ICSTCC), Sinaia, Romania, 17–19 October 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 874–878. [Google Scholar]
  51. Wu, S.-F.; Lee, S.-J. Multi-Valued Neuron with New Learning Schemes. In Proceedings of the 2013 International Joint Conference on Neural Networks (IJCNN), Dallas, TX, USA, 4–9 August 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1–7. [Google Scholar]
  52. Aizenberg, I.; Vasko, A. Frequency-Domain and Spatial-Domain MLMVN-Based Convolutional Neural Networks. Algorithms 2024, 17, 361. [Google Scholar] [CrossRef]
  53. Bindi, M.; Luchetta, A.; Lozito, G.M.; Carobbi, C.F.M.; Grasso, F.; Piccirilli, M.C. Frequency Characterization of Medium Voltage Cables for Fault Prevention Through Multi-Valued Neural Networks and Power Line Communication Technologies. IEEE Trans. Power Deliv. 2023, 38, 3227–3237. [Google Scholar] [CrossRef]
  54. Bindi, M.; Piccirilli, M.C.; Luchetta, A.; Grasso, F.; Manetti, S. Testability Evaluation in Time-Variant Circuits: A New Graphical Method. Electronics 2022, 11, 1589. [Google Scholar] [CrossRef]
  55. Fink, O.; Zio, E.; Weidmann, U. Predicting Component Reliability and Level of Degradation with Complex-Valued Neural Networks. Reliab. Eng. Syst. Saf. 2014, 121, 198–206. [Google Scholar] [CrossRef]
  56. Grasso, F.; Manetti, S.; Piccirilli, M.C.; Reatti, A. A Laplace Transform Approach to the Simulation of DC-DC Converters. Int. J. Numer. Model. 2019, 32, e2618. [Google Scholar] [CrossRef]
  57. Luchetta, A.; Manetti, S.; Piccirilli, M.C.; Reatti, A.; Corti, F.; Catelani, M.; Ciani, L.; Kazimierczuk, M.K. MLMVNNN for Parameter Fault Detection in PWM DC–DC Converters and Its Applications for Buck and Boost DC–DC Converters. IEEE Trans. Instrum. Meas. 2019, 68, 439–449. [Google Scholar] [CrossRef]
  58. Nedjah, N.; Galindo, J.D.L.; Mourelle, L.D.M.; Oliveira, F.D.V.R.D. Fault Diagnosis in Analog Circuits Using Swarm Intelligence. Biomimetics 2023, 8, 388. [Google Scholar] [CrossRef]
  59. Ronghua, J.; Shulei, Z.; Lihua, Z.; Qiuxia, L.; Saeed, I.A. Prediction of Soil Moisture with Complex-Valued Neural Network. In Proceedings of the 2017 29th Chinese Control And Decision Conference (CCDC), Chongqing, China, 28–30 May 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1231–1236. [Google Scholar]
  60. Aizenberg, E.; Aizenberg, I. Batch Linear Least Squares-Based Learning Algorithm for MLMVN with Soft Margins. In Proceedings of the 2014 IEEE Symposium on Computational Intelligence and Data Mining (CIDM), Orlando, FL, USA, 9–12 December 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 48–55. [Google Scholar]
  61. Aizenberg, I.; Luchetta, A.; Manetti, S. A Modified Learning Algorithm for the Multilayer Neural Network with Multi-Valued Neurons Based on the Complex QR Decomposition. Soft Comput. 2012, 16, 563–575. [Google Scholar] [CrossRef]
Figure 1. (a) Ideal image. (b) Image with additive Gaussian noise. (c) Magnitude of ideal image Fourier transform. (d) Magnitude of noisy image Fourier transform. (e) Magnitude of noisy image Fourier transform with low frequencies suppressed. (f) Magnitude of noisy image Fourier transform with high frequencies suppressed. (g) Inverse Fourier transform of noisy image with low frequencies suppressed. (h) Inverse Fourier transform of noisy image with high frequencies suppressed.
Figure 1. (a) Ideal image. (b) Image with additive Gaussian noise. (c) Magnitude of ideal image Fourier transform. (d) Magnitude of noisy image Fourier transform. (e) Magnitude of noisy image Fourier transform with low frequencies suppressed. (f) Magnitude of noisy image Fourier transform with high frequencies suppressed. (g) Inverse Fourier transform of noisy image with low frequencies suppressed. (h) Inverse Fourier transform of noisy image with high frequencies suppressed.
Algorithms 18 00461 g001
Figure 2. Process of extracting overlapping patches from image.
Figure 2. Process of extracting overlapping patches from image.
Algorithms 18 00461 g002
Figure 3. Image noise filtering process.
Figure 3. Image noise filtering process.
Algorithms 18 00461 g003
Figure 4. Examples of magnitudes of the convolution kernels in the frequency domain generated by the MLMVN learning process for filtering additive Gaussian noise (ad) and multiplicative speckle noise (eh). These examples are shown for 4 randomly selected hidden neurons for both kinds of noise.
Figure 4. Examples of magnitudes of the convolution kernels in the frequency domain generated by the MLMVN learning process for filtering additive Gaussian noise (ad) and multiplicative speckle noise (eh). These examples are shown for 4 randomly selected hidden neurons for both kinds of noise.
Algorithms 18 00461 g004
Figure 5. Ideal (clean) versions of the test images used for evaluating the MLMVN filtering capabilities.
Figure 5. Ideal (clean) versions of the test images used for evaluating the MLMVN filtering capabilities.
Algorithms 18 00461 g005
Figure 6. Effect of an activation function: (a) a standard continuous activation function for MVNs; (b) the complex-valued sigmoid activation function; (c) the complex-valued hyperbolic tangent activation function. Red crosses represent MVN weighted sum values and blue circles represent respective neuron outputs after applying a respective activation function to the weighted sums.
Figure 6. Effect of an activation function: (a) a standard continuous activation function for MVNs; (b) the complex-valued sigmoid activation function; (c) the complex-valued hyperbolic tangent activation function. Red crosses represent MVN weighted sum values and blue circles represent respective neuron outputs after applying a respective activation function to the weighted sums.
Algorithms 18 00461 g006
Figure 7. Convergence curves. The mean PSNR over the validation set during the learning process, measured after learning of each batch is completed: (a) MLMVN-G-1, (b) MLMVN-G-2, (c) MLMVN-G-3, (d) MLMVN-S-1, (e) MLMVN-S-2, (f) MLMVN-S-3. The notation E: X, B: Y denotes Epoch X and Batch Y.
Figure 7. Convergence curves. The mean PSNR over the validation set during the learning process, measured after learning of each batch is completed: (a) MLMVN-G-1, (b) MLMVN-G-2, (c) MLMVN-G-3, (d) MLMVN-S-1, (e) MLMVN-S-2, (f) MLMVN-S-3. The notation E: X, B: Y denotes Epoch X and Batch Y.
Algorithms 18 00461 g007
Figure 8. (a) Ideal image. (b) Noisy image. Gaussian noise 0.2 σ . PSNR 28.9904 dB. (c) Image filtered using MLMVN. PSNR 34.2611 dB. (d) Image filtered using BM3D filter. SD parameter is 0.5 σ . PSNR 31.2659 dB. (e) Ideal image. (f) Noisy image. Gaussian noise 0.2 σ . PSNR 25.5342 dB. (g) Image filtered using MLMVN. PSNR 29.32347 dB. (h) Image filtered using BM3D filter. SD parameter is 1 σ . PSNR 29.8181 dB. (i) Ideal image. (j) Noisy image. Gaussian noise 0.2 σ . PSNR 25.5579 dB. (k) Image filtered using MLMVN. PSNR 31.701 dB. (l) Image filtered using BM3D filter. SD parameter is 1 σ . PSNR 33.7363 dB.
Figure 8. (a) Ideal image. (b) Noisy image. Gaussian noise 0.2 σ . PSNR 28.9904 dB. (c) Image filtered using MLMVN. PSNR 34.2611 dB. (d) Image filtered using BM3D filter. SD parameter is 0.5 σ . PSNR 31.2659 dB. (e) Ideal image. (f) Noisy image. Gaussian noise 0.2 σ . PSNR 25.5342 dB. (g) Image filtered using MLMVN. PSNR 29.32347 dB. (h) Image filtered using BM3D filter. SD parameter is 1 σ . PSNR 29.8181 dB. (i) Ideal image. (j) Noisy image. Gaussian noise 0.2 σ . PSNR 25.5579 dB. (k) Image filtered using MLMVN. PSNR 31.701 dB. (l) Image filtered using BM3D filter. SD parameter is 1 σ . PSNR 33.7363 dB.
Algorithms 18 00461 g008
Figure 9. (a) Ideal image. (b) Noisy image. Speckle noise 0.3 σ . PSNR 27.8073 dB. (c) Image filtered using MLMVN. PSNR 32.2338 dB. (d) Image filtered using BM3D filter. SD parameter is 1 σ . PSNR 29.1584 dB. (e) Ideal image. (f) Noisy image. Speckle noise 0.3 σ . PSNR 27.7893 dB. (g) Image filtered using MLMVN. PSNR 31.4565 dB. (h) Image filtered using BM3D filter. SD parameter is 1.5 σ . PSNR 31.1511 dB. (i) Ideal image. (j) Noisy image. Speckle noise 0.3 σ . PSNR 25.929 dB. (k) Image filtered using MLMVN. PSNR 32.6948 dB. (l) Image filtered using BM3D filter. SD parameter is 1 σ . PSNR 33.8770 dB.
Figure 9. (a) Ideal image. (b) Noisy image. Speckle noise 0.3 σ . PSNR 27.8073 dB. (c) Image filtered using MLMVN. PSNR 32.2338 dB. (d) Image filtered using BM3D filter. SD parameter is 1 σ . PSNR 29.1584 dB. (e) Ideal image. (f) Noisy image. Speckle noise 0.3 σ . PSNR 27.7893 dB. (g) Image filtered using MLMVN. PSNR 31.4565 dB. (h) Image filtered using BM3D filter. SD parameter is 1.5 σ . PSNR 31.1511 dB. (i) Ideal image. (j) Noisy image. Speckle noise 0.3 σ . PSNR 25.929 dB. (k) Image filtered using MLMVN. PSNR 32.6948 dB. (l) Image filtered using BM3D filter. SD parameter is 1 σ . PSNR 33.8770 dB.
Algorithms 18 00461 g009
Table 1. The PSNR of the noisy images (additive Gaussian noise with SD = 0.2 σ ) from the test dataset processed by MLMVN trained using different batch sizes. The best results are highlighted in bold.
Table 1. The PSNR of the noisy images (additive Gaussian noise with SD = 0.2 σ ) from the test dataset processed by MLMVN trained using different batch sizes. The best results are highlighted in bold.
Number of samples in a batchTrain stationMusic hallBig BenCambridge bldgGaudi bldg
100029.849329.261227.116429.929828.7228
200029.675729.320226.999530.029728.8767
300029.628629.286127.017430.014728.8323
400030.118629.579927.366330.20729.1725
500030.252729.62127.476430.160129.2054
600030.233129.627.419230.24429.1891
10,00030.416629.600227.598630.076529.128
20,00029.848529.386127.309830.123628.9371
Number of samples in a batchBarcelona plazaSailboatManhattanAirplaneFighter-jet
100028.343233.360931.486933.767733.9199
200028.449633.642231.6334.028834.2254
300028.434733.533831.510833.799234.2069
400028.685333.820431.817934.449634.448
500028.71733.810131.811434.377634.5142
600028.727333.884931.87634.503134.5963
10,00028.765133.551731.704634.153934.4115
20,00028.596133.143431.326833.458333.9033
Table 2. Activation functions used in output layer MVNs in this work.
Table 2. Activation functions used in output layer MVNs in this work.
NameExpression
Identity activation function f ( z ) = z
Classical activation function for MVN with a continuous output f ( z ) = z | z |
Complex-valued sigmoid activation function f ( z ) = z , | z | 1 1 1 + e | z | e i arg ( z ) , | z | > 1
Complex-valued hyperbolic tangent activation function f ( z ) = tanh ( | z | ) e i arg ( z )
Table 3. Results of filtering grayscale images affected by additive Gaussian noise with a standard deviation of 0.1 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Table 3. Results of filtering grayscale images affected by additive Gaussian noise with a standard deviation of 0.1 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Train stationMusic hallBig BenCambridge bldgGaudi bldg
Noisy image PSNR31.453432.411129.628534.188131.8786
BM3D 0.5 σ 33.398733.614431.248534.950832.9469
BM3D 1 σ 37.431935.607334.225836.691134.6627
BM3D 1.5 σ 36.11934.209633.103735.502433.7332
MLMVN-G-135.424832.304931.600732.030831.7697
MLMVN-G-235.404732.234731.574232.005131.7367
MLMVN-G-335.459432.327631.616732.136231.8193
Barcelona plazaSailboatManhattanAirplaneFighter-jet
Noisy image PSNR31.464334.989933.438135.081735.1862
BM3D 0.5 σ 32.351836.764835.118137.152336.7276
BM3D 1 σ 33.865640.136638.255641.718739.8033
BM3D 1.5 σ 32.40338.328936.9441.161738.9479
MLMVN-G-131.55535.524634.30637.752337.2053
MLMVN-G-231.528935.526334.297237.728637.2716
MLMVN-G-331.600135.581934.389337.850437.2804
Table 4. Results of filtering grayscale images affected by additive Gaussian noise with a standard deviation of 0.2 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Table 4. Results of filtering grayscale images affected by additive Gaussian noise with a standard deviation of 0.2 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Train stationMusic hallBig BenCambridge bldgGaudi bldg
Noisy image PSNR25.557926.38523.999228.194325.9918
BM3D 0.5 σ 28.154428.13326.3529.49127.6905
BM3D 1 σ 33.736331.213530.130732.452130.6029
BM3D 1.5 σ 32.588629.583728.418730.803228.7687
MLMVN-G-131.70130.154728.596330.675829.6646
MLMVN-G-231.689630.112628.597130.663729.6592
MLMVN-G-331.682230.142828.587130.7429.678
Barcelona plazaSailboatManhattanAirplaneFighter-jet
Noisy image PSNR25.534228.990427.427129.080729.1782
BM3D 0.5 σ 27.083631.265929.657331.646531.345
BM3D 1 σ 29.818135.877234.28838.226336.4391
BM3D 1.5 σ 28.081133.892932.645837.352635.4323
MLMVN-G-129.324734.261132.453535.72935.685
MLMVN-G-229.321834.221232.429335.660635.669
MLMVN-G-329.332634.268232.470635.749735.6836
Table 5. Results of filtering grayscale images affected by additive Gaussian noise with a standard deviation of 0.3 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Table 5. Results of filtering grayscale images affected by additive Gaussian noise with a standard deviation of 0.3 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Train stationMusic hallBig BenCambridge bldgGaudi bldg
Noisy image PSNR22.161922.909820.860124.697222.5851
BM3D 0.5 σ 25.157825.033723.819926.347624.7072
BM3D 1 σ 31.580328.852527.695230.213228.2913
BM3D 1.5 σ 30.59927.294326.095628.678826.4801
MLMVN-G-127.896527.426325.487328.73426.979
MLMVN-G-227.915827.433225.508128.745227.0003
MLMVN-G-327.86527.401825.471428.763126.9621
Barcelona plazaSailboatManhattanAirplaneFighter-jet
Noisy image PSNR22.145625.49323.916925.587925.6751
BM3D 0.5 σ 24.164328.063526.429428.392428.1542
BM3D 1 σ 27.756133.76332.167735.97834.4722
BM3D 1.5 σ 26.284832.127230.600234.836833.4473
MLMVN-G-126.610431.72529.658232.290132.6759
MLMVN-G-226.628531.720529.673632.284132.6955
MLMVN-G-326.587831.700929.644232.280532.6411
Table 6. Results of filtering grayscale images affected by multiplicative speckle noise with a standard deviation of 0.2 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Table 6. Results of filtering grayscale images affected by multiplicative speckle noise with a standard deviation of 0.2 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Train stationMusic hallBig BenCambridge bldgGaudi bldg
Noisy image PSNR32.675332.115930.349835.335431.2367
BM3D 0.5 σ 36.730134.645333.313537.006133.7141
BM3D 1 σ 34.625232.04331.321633.629831.8137
BM3D 1.5 σ 33.155729.841228.720331.033128.8251
Frost29.810925.829825.236124.397726.0143
Lee31.461128.238526.790427.465327.7698
MLMVN-S-136.258533.113934.776735.089433.4302
MLMVN-S-235.739832.829534.424234.885333.0994
MLMVN-S-335.946832.951134.628734.989433.2681
Barcelona plazaSailboatManhattanAirplaneFighter-jet
Noisy image PSNR31.204634.608633.720635.748531.8837
BM3D 0.5 σ 33.386240.326937.950840.78635.6958
BM3D 1 σ 30.486836.267135.331740.206737.5485
BM3D 1.5 σ 28.237934.012432.993538.152635.7779
Frost25.347130.410327.766230.128831.1936
Lee27.081432.794530.094534.232734.3365
MLMVN-S-134.018338.725537.657739.576237.1491
MLMVN-S-233.717338.470337.259938.698237.1038
MLMVN-S-333.879238.702537.498739.120937.1174
Table 7. Results of filtering grayscale images affected by multiplicative speckle noise with a standard deviation of 0.3 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Table 7. Results of filtering grayscale images affected by multiplicative speckle noise with a standard deviation of 0.3 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Train stationMusic hallBig BenCambridge bldgGaudi bldg
Noisy image PSNR29.154828.629627.370431.828827.7893
BM3D 0.5 σ 34.203731.762331.026433.956431.1511
BM3D 1 σ 32.57629.415428.464430.918128.8088
BM3D 1.5 σ 31.359127.532226.557829.00526.6459
Frost29.301725.584624.999824.358125.8356
Lee30.735127.826726.437927.361627.5166
MLMVN-S-133.424831.286531.891733.688231.4565
MLMVN-S-232.911530.938131.714233.354631.1229
MLMVN-S-333.121231.085531.79833.509531.2845
Barcelona plazaSailboatManhattanAirplaneFighter-jet
Noisy image PSNR27.725231.078530.22932.294328.3538
BM3D 0.5 σ 30.496837.190135.255438.517332.8427
BM3D 1 σ 28.188733.938332.906537.770335.4343
BM3D 1.5 σ 26.597132.403430.971935.519533.7933
Frost25.212830.189627.631329.956930.7026
Lee26.886532.421229.852633.763333.3033
MLMVN-S-131.825936.631335.311437.480435.4004
MLMVN-S-231.537735.975634.715836.338135.0165
MLMVN-S-331.67736.327135.007536.85135.1973
Table 8. Results of filtering grayscale images affected by multiplicative speckle noise with a standard deviation of 0.4 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Table 8. Results of filtering grayscale images affected by multiplicative speckle noise with a standard deviation of 0.4 σ , where σ is the standard deviation of the ideal image. The best results are shown in bold, and the best results using our approach are highlighted in color.
Train stationMusic hallBig BenCambridge bldgGaudi bldg
Noisy image PSNR26.700526.186825.199329.345425.3672
BM3D 0.5 σ 32.497929.861629.281731.948929.4405
BM3D 1 σ 31.129627.722526.752429.429727.1516
BM3D 1.5 σ 30.080526.148225.256127.873325.5813
Frost28.689425.26724.670924.305225.6198
Lee29.874127.294825.96227.230527.2076
MLMVN-S-130.88629.40229.354332.155129.4495
MLMVN-S-230.516729.115929.376431.795829.2278
MLMVN-S-330.662729.227629.362731.953929.3293
Barcelona plazaSailboatManhattanAirplaneFighter-jet
Noisy image PSNR25.274728.58427.732729.838925.858
BM3D 0.5 σ 28.634535.065833.373236.87930.7953
BM3D 1 σ 26.919132.71931.397735.991333.9965
BM3D 1.5 σ 25.637931.521729.804633.813132.5234
Frost25.007629.904627.436629.758230.1304
Lee26.552931.917529.493833.193832.1024
MLMVN-S-129.507234.374332.96335.226333.2604
MLMVN-S-229.359233.672532.42134.179632.8563
MLMVN-S-329.430334.001132.664734.621333.0585
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aizenberg, I.; Tovt, Y. Intelligent Frequency Domain Image Filtering Based on a Multilayer Neural Network with Multi-Valued Neurons. Algorithms 2025, 18, 461. https://doi.org/10.3390/a18080461

AMA Style

Aizenberg I, Tovt Y. Intelligent Frequency Domain Image Filtering Based on a Multilayer Neural Network with Multi-Valued Neurons. Algorithms. 2025; 18(8):461. https://doi.org/10.3390/a18080461

Chicago/Turabian Style

Aizenberg, Igor, and Yurii Tovt. 2025. "Intelligent Frequency Domain Image Filtering Based on a Multilayer Neural Network with Multi-Valued Neurons" Algorithms 18, no. 8: 461. https://doi.org/10.3390/a18080461

APA Style

Aizenberg, I., & Tovt, Y. (2025). Intelligent Frequency Domain Image Filtering Based on a Multilayer Neural Network with Multi-Valued Neurons. Algorithms, 18(8), 461. https://doi.org/10.3390/a18080461

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop