Next Article in Journal
Deep UV Formation of Long-Term Stable Optical Bragg Gratings in Epoxy Waveguides and Their Biomedical Sensing Potentials
Previous Article in Journal
Climate-Aware and IoT-Enabled Selection of the Most Suitable Stone Fruit Tree Variety
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of a Generative Adversarial Network in Image Reconstruction of Magnetic Induction Tomography

1
Key Laboratory of Data Analytics and Optimization for Smart Industry, Northeastern University, Shenyang 110819, China
2
Key Laboratory of Infrared Optoelectric Materials and Micro-Nano Devices, Shenyang 110819, China
3
College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
4
College of Computer Science and Engineering, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(11), 3869; https://doi.org/10.3390/s21113869
Submission received: 28 April 2021 / Revised: 29 May 2021 / Accepted: 31 May 2021 / Published: 3 June 2021
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Image reconstruction of Magnetic induction tomography (MIT) is an ill-posed problem. The non-linear characteristics lead many difficulties to its solution. In this paper, a method based on a Generative Adversarial Network (GAN) is presented to tackle these barriers. Firstly, the principle of MIT is analyzed. Then the process for finding the global optimum of conductivity distribution is described as a training process, and the GAN model is proposed. Finally, the image was reconstructed by a part of the model (the generator). All datasets are obtained from an eight-channel MIT model by COMSOL Multiphysics software. The voltage measurement samples are used as input to the trained network, and its output is an estimate for image reconstruction of the internal conductivity distribution. The results based on the proposed model and the traditional algorithms were compared, which have shown that average root mean squared error of reconstruction results obtained by the proposed method is 0.090, and the average correlation coefficient with original images is 0.940, better than corresponding indicators of BPNN and Tikhonov regularization algorithms. Accordingly, the GAN algorithm was able to fit the non-linear relationship between input and output, and visual images also show that it solved the usual problems of artifact in traditional algorithm and hot pixels in L2 regularization, which is of great significance for other ill-posed or non-linear problems.

1. Introduction

Magnetic induction tomography (MIT) is a kind of electromagnetic imaging technology based on the eddy current testing principle [1]. In general applications, many groups of excitation–detection combination coils are used to obtain induced magnetic field signals near the imaging body [2], and then the conductivity distribution of the range to be measured is calculated by the reconstruction algorithm [3,4]. Because of its non-invasive and non-contact characteristics, MIT is suitable for geological exploration [5], industrial flaw detection [6], impurity detection [7], and medical imaging [8]. The research on it can be roughly divided into a positive problem and an inverse problem [9]. The former is to calculate the edge detection signals or the change of magnetic field according to the existing electrical characteristics of the object [10], while the latter is to use the detection signals to restore the conductivity distribution map (reconstruction). Due to the soft field effect of electromagnetic field and the limitation of detector, the inverse problem has to recover more complex electrical characteristics from very few signals, which is also the key problem of MIT where underdetermination and nonlinearity coexist.
Since MIT and other electromagnetic tomography methods entered the field of vision of scholars, many reconstruction algorithms for it have been continuously proposed. They can be divided into two categories namely traditional or intelligent algorithms. The general idea of traditional ones is to discretize the imaging volume and approximate the nonlinear soft field with linear equations. The Gauss-Newton [11], Landweber iteration [12], conjugate gradient (CG) [13], Tikhonov regularization method, etc. were proposed in the early stages, and current traditional algorithms are basically their superposition or improvement. In 2015, Han min et al. [14] proposed an iterative NR algorithm based on weighted matrix and L1 norm regularization to reduce the instability of classical NR. In 2017, Liu et al. [15] proposed a novel Tikhonov regularized conjugate gradient (CG) algorithm to enhance the imaging quality. In 2019, Xiao et al. [16] reconstructed three-dimensional images of the human head model with different volumes and locations of intraaxial hemorrhaging using single-step Tikhonov regularization. Although most of these methods are relatively easy, in practical application, they can only contain limited prior information, and the calculation of an objective function gradient or Hesse matrix may be time or memory consuming [17]. Realistic needs expect more abundant prior knowledge and learn imaging features autonomously [18], to construct a complex mapping from low to high dimension.
Neural networks have even lower computational demands in dealing with nonlinear problems [19]. Some models have better real-time performance than iterative algorithms and many of them have been put forward before the popularity of deep learning. Hopfield neural network [20], radial basis function (RBF) neural network [21], and back projection network [22] have been applied earlier. In 2012, Li et al. [23] used simulation software to generate capacitance data and trained a faster RBF neural network. The effectiveness of the proposed method is verified by image reconstruction of typical flow patterns. With the popularity of convolutional neural networks (CNNs), methods based on it are applied in Electrical Impedance Tomography (EIT) image reconstruction at first [24]. In 2017, Klosowski et al. [25] proposed two reconstruction methods based on fully-connected neural network (FCNN) and CNN, respectively, which proved that CNN can be directly used between the input and the output. In 2018, Zheng et al. [26,27] developed an auto-encoder method to achieve complicated reconstruction in electrical capacitance tomography. Feng et al. [28] investigated the feasibility of a back-propagation neural network (BPNN) to reestablish the distribution of optical properties in a diffuse optical tomography (DOT) problem. Their method of evaluating resolution is also used for reference by other scholars. The stacked auto encoder (SAE) is an old method similar to one-dimensional convolution before CNN, also proved to be an effective approach to enhance images lately [29]. In 2019, Chen et al. [30] proposed an algorithm based on an SAE neural network to reconstruct the anomaly in biological tissues. Nevertheless, SAE has sensitivity to measurement, resulting in random hot pixels that appear in the images, and some scholars began to seek a breakthrough in training methods beyond the network structure. In 2020, Ren et al. [31] proposed a two-stage deep learning (TSDL) method to achieve high resolution and modeling error robust EIT image, and provided a new idea for the training mode of deep model.
Generative Adversarial Networks (GAN) uses an neural network to supervise another one. Its various variants, such as SARA-GAN and PGGAN et al., are widely used in the field of image reconstruction, including medical image enhancement [32], sample expansion [33], region of interest segmentation [34], and lesion development prediction [35]. In 2020, Liu et al. [36] used multi-stage trained GAN to reconstruct multispectral images from RGB images. The results show that GAN is helpful to solve the underdetermined problem. Yuan et al. [37] Used GAN to construct the image details in Magnetic Resonance Imaging (MRI) results in order to shorten the necessary imaging time and reduce the pain of patients. In 2021, Chen et al. [38] used generative adversarial network (GAN) to enhance MIT images calculated by complex CNN firstly, and achieved good results, yet it needs hundreds of detection signals. It should be noted that all of the above image reconstructions refer to the acquisition of higher quality images by existing images, i.e., several kinds of image to image translation process, instead of reconstructing the image directly from the detected signal. It is true that MIT measurement equipment with fewer channels can reduce the measurement cost. However, due to various limitations, the existing algorithms still cannot complete the reconstruction task based on low dimensional input well. Specifically, in the case of extremely low conductivity difference, the algorithm which only detects less than hundred electrical signals to ensure real-time performance and can clearly distinguish two anomalies has not been developed.
In view of the shortcomings of existing MIT image reconstruction methods, this paper proposed a novel MIT imaging algorithm based on generative adversarial networks. The GAN with low dimensional input was established, and trained by the data obtained from an eight-channel MIT model, where its whole training is divided into two stages, which is the improvement of the training process in this paper, referring to the TSDL method. The generator is pre-trained with its hyperparameter adjusted by back propagation with L2 regularization at first, where the problem of MIT image reconstruction was regarded as a complex nonlinear multivariate function mapping problem. Then, the generator and discriminator are combined to form a complete GAN model for joint training. Its parameters of each layer of the generator are finally separated to be used to reconstruct high-quality MIT images. Four evaluation indexes are set to compare it with BPNN and Tikhonov regularization. Experiment results show that the proposed GAN has better accuracy and robustness than others even in the case of noise. Specifically, it has smaller error, less hot pixels and artifacts, which provides a new way to solve the similar ill-posed imaging problems.
The rest of this paper is organized as follows: Section 2 presents the improved method and its principles of mathematics; Section 3 introduces the eight-channel MIT model, the process of data set building and GAN model training; Section 4 analyzes and discusses the experimental results compare with that of other algorithms; Section 5 concludes the paper, meanwhile the future works is put forward.

2. Methodology

2.1. Principles of MIT Problem

MIT is based on the principle of electromagnetic induction and eddy current testing. The relationship between the eddy current density and the electrical characteristics of imaging object is described as Equation (1) according to Maxwell equations as follows:
J e = σ + j ω ε E
where eddy current J e is composed of two parts, ε is the permittivity that can be ignored in the inverse problem and eddy current is mainly generated by the conductivity σ in the biological object, with physical modeling and finite element model (FEM) discretization, the deterministic MIT measurement model can be expressed as:
V = F σ
where V is a vector of the measured phase of boundary voltage, F is the forward model map of the internal conductivity distribution, which is usually linearized to the sensitivity matrix in traditional algorithms as V = S σ . The purpose of the MIT image reconstruction (an ill-posed, nonlinear inverse problem) is to obtain the conductivity distributions by solving the model in Equation (2). The inverse problem is generally expressed as σ = F 1 V or σ = S 1 V .
Furthermore, the sensitivity matrix cannot accurately describe the corresponding relationship between the measured value and conductivity due to the soft field effect of electromagnetic field, which increases the difficulty of solving MIT problem. This is also a reason why the traditional algorithms are not always satisfactory.

2.2. Principles of ANN

Artificial neural network is a kind of mathematical structure formed by amounts of data processing units (or neurons) connected by directional operations. It is an digital abstraction of the mechanism of the brain. Specifically, each neuron represents a mapping relation called activation function. Connections between neurons represent the weights for signal passing through these connections, equivalent to network memory. The output of an ANN varies according to its different connection structures, weight values, and activation functions. This kind of output mode is called forward propagation in studies, as shown below:
a l + 1 = f W l + 1 , l a l + b l
where a i represents the values of neurons in layer i , and f represents the activation function of this layer. W and b are the weights and bias (matrix) to be optimized, collectively referred to as parameters matrix θ , and ANNs are able to fit various complex nonlinear functions by modifying θ . With the deepening of layers, they can describe increasingly complex mapping relationships.
Back propagation neural network (BPNN) is a fully connected ANN trained by error back propagation algorithm, which is based on the chain derivative rule. It transfers the error between the predicted result and the real value along the opposite direction of output mode, so as to adjust parameters matrix and optimize the fitting effect. It has the advantages of being simple and fast, with less super parameters, easy training, and convergence. The following GAN uses two such networks with different layers to achieve the goal of image reconstruction.

2.3. The GAN for MIT Image Reconstruction

Generative adversarial network (GAN) is a class of machine learning frameworks. Two neural networks contest with each other in a game (in the form of a zero-sum game). This technique learns to generate new data with the same statistics as the training set. Its key idea is based on the “indirect” training through the discriminator, which basically means that the generator is not only trained to minimize the distance to a specific image, but also to fool the discriminator, similar to supervised maximum likelihood estimation. The underdetermined and non-linear problems can be solved by it, and the over-fitting caused by direct feedback can be avoided.
To build a GAN model for an MIT problem, this paper pre-trains the generator firstly using L2 regularization, takes phase difference as input, adjusts the network architecture to complete simple imaging tasks, then connects it with a discriminator to form a counter network, and adds a cross verification set to monitor its convergence. When the generator error is less than the discriminator error and the verification set error is less than 0.05, the training is stopped. The general training process is shown in Figure 1. Its main mathematical logic will be introduced next.
The essence of the generator model here is a maximum likelihood estimate, which maximizes the likelihood of conductivity reconstructed by the generator based on the true distribution.
L G , V , σ = i = 1 M P G V σ i ; θ g
where the target is to find the appropriate generator parameter θ g * , where g represents the generator, V represents the voltage, σ represents the conductivity, and P represents the discrete distribution function.
θ g * = argmax θ g i = 1 M P G V σ i ; θ g Θ g * = argmax θ g log i = 1 M P G V σ i ; θ g Θ g * = argmax θ g i = 1 m log P G V σ i ; θ g Θ g * = argmax θ g   E σ [ log P G V σ i ; θ g ] Θ g * = argmax θ g σ   p σ x log p G V x ; θ g dx σ   p σ x log p σ x dx
where E represents the mathematical expectation and p represents the continuous distribution function. The term σ   p σ x log p σ x dx subtracted from the above formula, which is independent of θ g , shall not affect the solution of the maximum point. The purpose is to construct the following Kullback–Leibler divergence, a measure of probability distribution similarity in statistics.
θ g * = argmax θ g σ   p σ x log p x ; θ g p σ x dx Θ g * = argmin θ g   KL P σ x P G V x ; θ g
This K–L divergence minimization problem cannot be solved directly with maximum likelihood, so the discriminator loss is defined as follows to optimize the generator parameters, that is, to calculate instead of maximum likelihood estimation:
Value D , G = σ   p σ x log D x dx + V   p V x log 1 D G x dx
According to the Radon–Nikodym theorem:
Value D , G = σ   ( p σ x log D x + p G x log 1 D x ) dx
where p σ x and p G x are independent of the discriminator D x . Let the integrand of the above formula take the derivative of D x and make it equal to 0:
p σ x D x + p G x D x 1 = 0 D x = p σ x p σ x + p G x
It can be proved that if and only if p σ x = p G x , D x = 0.5 , which is the only optimal solution to this problem, means the reconstructed conductivity calculated by the generator is exactly consistent with the real conductivity so that the discriminator can’t tell whether the image is from generator or real samples. Independent back-propagation procedures are applied to both networks so that the generator produces better images while the discriminator becomes more skilled at flagging synthetic images. The generator and discriminator of GAN used in this paper are both BPNN to adapt to voltage and conductivity vector. The objective function for this GAN model is:
min θ g   max θ d   Value D , G = E σ logD σ + E V log 1 D G V
Due to the co-evolution of generators and discriminators, it performs better than general regularization or supervised training algorithms and are less prone to fall into local optimum. The optimization process is as following Algorithm 1.
Algorithm 1 Gradient descent optimization of GAN
for number of training iterations do
for k steps do
Sample batch of M n u m samples from voltage distributions V 1 , , V M .
Sample batch of M n u m samples from conductivity distributions σ 1 , , σ M .
Update the discriminator by ascending its stochastic gradient:
θ d 1 m i = 1 M n u m logD σ i + log 1 D G V i .
end for
Sample batch of M n u m samples from voltage distributions V 1 , , V M .
Update the generator by descending its stochastic gradient:
θ g 1 m i = 1 M n u m log 1 D G V i .
end for

2.4. Data Set Preparing and Network Training

The needed data sets are obtained by solving the MIT forward problem with finite element method before the training. Each sample consists of two parts: the conductivity distribution of the sampling grid and the computed phase difference of the measured boundary voltage. The process of obtaining experimental samples is shown in Algorithms 2 and 3.
Algorithm 2 Construct training samples
The sampling grid and 8-coil MIT model are established.
for number of needed samples do
Load the model.
for  ρ i , θ i in coordinates of anomalies do
for j in coordinates of number of coils (8) anomalies do
Assign ρ i , θ i to change the position of anomaly (anomalies) in the model.
The j-th coil acts as the excitation and the others as the receiver.
Re geometric modeling and mesh generation. Run the model.
7 voltage data are obtained from the receiving coils.
3409 conductivity data are obtained from the sampling grid.
end for
end for
Vec-operator the collected voltage data and the conductivity data.
Save as a corresponding sample V n u m , σ n u m .
end for
Algorithm 3 Construct training samples
Load the exiting dataset.
for number of expansion times do
Build a square lattice same with sampling grid.
for  ρ i , θ i and V i , σ i in i-th sample do
Convert polar coordinate ρ i , θ i + n u m 45 ° to rectangular coordinate x i , y i .
Set points x , y meets ( x x i ) 2 + ( y y i ) 2 < radius 2 in square lattice as anomalies.
Read the conductivity signal σ j in sequence again.
Move the data in V i forward 7 units in sequence to form V j .
end for
Save these V 1 , σ 1 , , V j , σ j , , V n u m , σ n u m as corresponding new samples.
end for
Firstly, the generator is pre-trained, and then it is connected to the discriminator for GAN training. Having obtained the voltages V 1 , V 2 , , V M with conductivities σ 1 , σ 2 , , σ M where M is the number of training samples, V k 0 , 1 n and σ k 0 , 1 m ( m is the number of phase differences of each detected sequence while n equals to the number of pixel of conductivity images), the BPNN denoted by G θ can be trained several epochs. In the first stage, given the output σ ^ i = G θ V i , we seek to minimize L2-error between that to real conductivity vector:
Loss σ , σ ^ = σ σ ^ 2 2
The above purpose is to determine the hyperparameters of BPNN as a generator to improve the performance in the subsequent training of GAN. In the second stage, the pre-trained network is as the initialization of the generator. Its outputs are as the inputs of discriminator with real conductivity distributions as labels. According to activation function of all layers, the gradient descent algorithm, based on the Adam optimizer, adjusts every parameter as follows:
θ θ α · Loss θ
where α is learning rate. After hyperparameters selection finished, a needed cross-validation set as ( V 1 , σ 1 ) , V 2 , σ 2 , , ( V N , σ N ) , a small number of untrained samples selected from the total data set, is used for supervised learning. Because the generator and discriminator of the GAN confront each other, which leads to the instability of their gradient descent and have to be monitored to determine the conditions of cessation in real time. A needed stronger generator can be selected when cross-validation signified its loss are smaller. The effect of cross validation is evaluated by MSE as follows:
MSE y , y ^ = 1 m i = 1 m y i y ^ i 2
which is the form of vector L2 error averaging to each dimension.

3. Experiments

3.1. MIT Signal Detection Model

In this paper, an 8-channel MIT measurement model was established with an anomaly placed in the circular imaging range with 8 coils around. After the excitation current was applied to one of these coils, the induced voltage of other detection coils was extracted. The general structure is shown in Figure 2.
The cylinder in the model represents the measured target tissue, with a height and a radius of 0.1 m, which is also the imaging radius. Eight identical 30 turn coils with a radius of 0.025 m were fixed on the surface circle of its longitudinal center, and their centers were 0.11 m away from the tissue center. Meanwhile, the eight coils were numbered counterclockwise, from coil 1 to coil 8. The conductivity of anomaly was 0.05 S/m to simulate the electrical characteristics of hematoma, and that of the cylinder model is 0.25 S/m to simulate the electrical characteristics of normal tissues. The alternating current with root mean square of 1 A and frequency of 10 MHz was introduced into each coil and received by other coils, a voltage vector with dimension of 56 can be measured in turn. At the same time, conductivity values with equal interval distribution can be extracted according to the dense square array within the circular range of imaging as the standard value of post reconstruction evaluation.

3.2. Data Set Preparation

In order to prepare for the optimization of artificial neural network model, needed data sets have to be generated for its training and testing. A widely used method was to generate samples with random parameters. To solve the inherent disadvantage of the MIT inverse problem, the poor sensitivity in the central region of imaging volume, the sample generation method moves the anomaly according to the polar coordinates, generate samples in turn, and then shuffle the order. In samples of a single target, a sphere anomaly with a radius of 2 cm was used to move from the center to the edge of the imaging region in turn. The angle step of each movement was 9°, and the radial step was 0.1 cm until the anomaly intersects the imaging boundary. The approximate method was also used in samples with two anomalies. Then, more samples were copied by rotating and symmetry. The measured voltage vector V and conductivity vector σ were stored as a matrix. The inputs correspond to the outputs, and they were scaled to the range of 0 to 1, which is normalization in deep learning. A total of 36,540 sets of data were obtained, including 2840 single-target and 33,664 two-anomaly samples. The following experiments were based on this identical database. The data set is generated by COMSOL multiphysics 5.5 with MATLAB.

3.3. Training and Reconstruction

Data sets were divided into two parts for training and testing. They were input into the GAN network as shown in Figure 1. One thousand epochs of the generator were trained to reduce loss using the Adam optimizer. The number of iterations and the learning rate have a serious effect on the accuracy of a DNN model. To minimize its loss, the generator structure of the proposed GAN was finally {56 128 256 512 1024 2048 4096 3409}. The activation function of its first four layers was ReLU, the fifth to seventh layers were tanh, and the last layer was sigmoid to satisfy the classification output between 0 and 1 (Figure 3).
Other hyperparameters were set as follows: In the pre-training phase of generator, the epoch was 1000 and the batch-size was 50. In the GAN fine-tuning phase, the batch-size was equal to 4000. In addition, control experiments were designed based on the same data set, including the classical and other intelligent algorithms.
In this paper, the image is reconstructed with 3409 pixels in a circular object field, with a 67 × 67 square grid. Algorithmic programming and image reconstruction were implemented under the Python 3.6 with Tensorflow 1.14.0 on a PC configured for 2.50 GHz CPU, 12.0 GB of memory.

4. Results and Discussion

4.1. Evaluating Indicators

Based on the requirements of image reconstruction, the following five evaluation indexes are proposed to measure the effect of model reconstruction.

4.1.1. Root Mean Squared Error (RMSE)

The sum of the square of the deviation between the reconstructed value, the true value, and the square root of the ratio of the total conductivity of the reconstructed sample m, which is used to measure the deviation between the reconstructed conductivity value and the true value. The RMSE is calculated as follows:
RMSE y , y ^ = 1 m i = 1 m y i y ^ i 2

4.1.2. Structural Similarity (SSIM)

Structural similarity is an index to measure the similarity of two images. SSIM uses two images, one is an uncompressed image without distortion, and the other is a distorted image. It is here used to evaluate the degree of image distortion after reconstruction:
SSIM y , y ^ = 2 μ y μ y ^ + c 1 2 σ y y ^ + c 2 μ y 2 + μ y ^ 2 + c 1 σ y 2 + σ y ^ 2 + c 2
where μ and σ 2 represent mean and variance, and σ y y ^ represents covariance. c 1 and c 2 are constants and are calculated as:
c 1 = ( k 1 L ) 2
c 2 = ( k 2 L ) 2
where k 1 is equal to 0.01 and k 2 is equal to 0.03. L is the range of the output value.

4.1.3. Peak Signal to Noise Ratio (PSNR)

Peak signal to noise ratio is the most common and widely used objective measure to assess picture quality, avoiding the subjective impact of direct observation (e.g., the human eye is more sensitive to differential contrast with lower spatial frequency). The PSNR is calculated as follows:
PSNR y , y ^ = 10 log 10 2 n 1 2 MSE y , y ^
where n is the binary digit of the image and MSE is the mean squared error (Equation (13)).

4.1.4. Correlation Coefficient (CC)

A correlation coefficient is a numerical measure of some type of correlation, meaning a statistical relationship between two variables. Several types of correlation coefficient exist, each with their own definition and own range of usability and characteristics. Here, we use:
CC y , y ^ = σ y y ^ σ y σ y ^
It assumes values in the range from −1 to +1, where ±1 indicates the strongest possible agreement and 0 the strongest possible disagreement.

4.2. Results and Analysis

Here we give the average value of 500 newly generated random test samples under four evaluation indices, including that with the different amplitude noises added, which is measured by the signal-to-noise ratio (SNR).
With the increase of noise amplitude, the index change of each model is shown in Figure 4. Figure 4a means that their reconstruction effect is worse as the noise increases. The RMSE of GAN increases from 0.1 to 0.2, more than doubled. In contrast, Tikhonov and BPNN are much more stable, about 0.5 and 0.35, respectively, but both are larger than GAN. Figure 4b illustrates an SSIM comparison, GAN is also far ahead, reaching more than 0.9 when noiseless. Figure 4c shows the effect of noise on PSNR. Its value range of GAN is (13, 22). Tikhonov and BPNN are stable at about 6 and 8, respectively. However, the BPNN without noise is not even as good as GAN with SNR of 20 dB. This signifies that GAN can filter noises in the signal as much as possible to avoid mapping to the image. Figure 4d shows that all three reconstruction methods can make the image positively correlated with the true value. The GAN algorithm is the best, with a mean up to (0.6, 0.9). What makes a discriminator superior to regularization is that it is only responsible for avoiding poor reconstruction results rather than a restriction on the range of parameter values.
In order to intuitively prove and reflect their advantages and disadvantages in robustness and resolution, we also give comparative experiments of imaging state and evaluation index of representative samples under different noise and different algorithms. In addition to the noiseless state, the SNR is 80, 40, and 20 dB, respectively; the minimum resolution distances of the two anomalies are set as 0, R, 2R, and 3R, respectively, where R is the radius of the anomalies. The pictures are shown in Figure 5 and Figure 6, and the relevant data are listed in Table 1, Table 2, Table 3 and Table 4.
It can be seen from the figures that although the influence of noise on GAN is slightly greater than other models, it is still better than BPNN and Tikhonov at any SNR. It also performs well in distinguishing two anomalies even they are tight together. In contrast, the BPNN model has more noise, while Tikhonov regularized image has a large area of artifacts. This also confirms the fact that GAN can filter noises.
From the data in the table, it appears that GAN exceeds 0.8 in both the SSIM and CC, which almost perfectly represents the true distribution of conductivity. RMSE can also be reduced below 0.2, although slightly higher in multi-target recognition, it is still satisfactory. From the point of view of signal restoration, its SSIM is over 15, which is much better than the other two algorithms.

5. Conclusions and Future Works

According to the basic principle of image reconstruction for MIT and the basic ideal of artificial neural network training, an MIT image reconstruction algorithm based on GAN is proposed in this paper. Firstly, an eight-coil MIT detection model is built, and anomalies in different positions are added. COMSOL Multiphysics software is used to solve the forward problem, and 36,540 of training samples are achieved after expansion. Secondly, the GAN model is built, and its generator is pre-trained by data sets with L2 regularization, and the network that can initially complete the reconstruction task is obtained. Then the generator and the discriminator are connected for joint training, and the convergence is determined by the cross-validation set. The separated generator can reconstruct the high quality conductivity distribution image according to the voltage signal. Finally, the reconstruction results of 500 new models that have not been used for training show that the average root mean square error of the proposed method is 0.090 and the average correlation coefficient is 0.940, which is much better than the traditional reconstruction algorithm Tikhonov regularization and BPNN using L2 regularization. This method provides a more effective solution for ill-posed inverse problem in imaging field.
Moreover, on this basis we can try to expand the imaging range to three-dimensional when collecting the conductivity signal. Specifically, it is not only to establish the sampling grid on the plane with z = 0, but also to divide several layers equidistant in the z direction, for example, to represent a 3D image by stacking 10 2D images in sequence. In this case, the structure of the artificial neural network should be adjusted accordingly, such as changing the vector output to the matrix output. It is also likely that the training time and even the computer hardware system put forward higher requirements, so that this algorithm still has potentials for further imaging studies.

Author Contributions

Conceptualization, J.L. and X.W.; methodology, D.Y.; software, J.L.; validation, D.Y. and B.X.; investigation, J.L.; writing—original draft preparation, J.L. and Y.W.; writing—review and editing, X.W. and D.Y.; visualization, B.X.; supervision, D.Y.; funding acquisition, D.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China under Grant Nos. 51607029 and 61836011. The authors also thank the Fundamental Research Funds for the Central Universities 2020GFZD008 and 2020GFYD011, and the 111 Project (B16009).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wei, H.Y.; Soleimani, M. Theoretical and Experimental Evaluation of Rotational Magnetic Induction Tomography. IEEE Trans. Instrum. Meas. 2012, 61, 3324–3331. [Google Scholar] [CrossRef]
  2. Walker, D.W.; Kramer, S.C.; Biebl, F.R.; Ledger, P.D.; Brown, M. Accelerating magnetic induction tomography-based imaging through heterogeneous parallel computing. Concurr. Comput. Pract. Exp. 2019, 31, 1–17. [Google Scholar] [CrossRef]
  3. Feldkamp, J.R. Single-coil magnetic induction tomographic three-dimensional imaging. J. Med Imaging 2015, 2, 013502. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Ke, L.; Zu, W.; Du, Q.; Chen, J.; Ding, X. A bio-impedance quantitative method based on magnetic induction tomography for intracranial hematoma. Med Biol. Eng. Comput. 2020, 58, 857–869. [Google Scholar] [CrossRef]
  5. Zakaria, Z.; Mansor, M.S.B.; Rahim, R.A.; Balkhis, I.; Rahiman, M.H.F.; Rahim, H.A.; Yaacob, S. Magnetic induction tomography: A review on the potential application in agricultural industry of Malaysia. J. Agric. Sci. 2013, 5, 78–82. [Google Scholar] [CrossRef]
  6. Bevington, P.; Gartman, R.; Chalupczak, W.; Deans, C.; Marmugi, L.; Renzoni, F. Non-destructive structural imaging of steelwork with atomic magnetometers. Appl. Phys. Lett. 2018, 113, 063503. [Google Scholar] [CrossRef]
  7. Polydorides, N.; Georghiou, G.E.; Kim, D.H.; Won, C. Subspace constrained regularization for corrosion detection with magnetic induction tomography. Ndt E Inter. 2008, 41, 510–516. [Google Scholar] [CrossRef]
  8. Bevacqua, M.T.; Bellizzi, G.G.; Crocco, L.; Isernia, T. A method for quantitative imaging of electrical properties of human tissues from only amplitude electromagnetic data. Inverse Probl. 2019, 35, 025006. [Google Scholar] [CrossRef]
  9. Chen, R.; Huang, J.; Li, B.; Wang, J.; Wang, H. Technologies for magnetic induction tomography sensors and image reconstruction in medical assisted diagnosis: A review. Rev. Sci. Instrum. 2020, 91, 091501. [Google Scholar] [CrossRef] [PubMed]
  10. Chen, R.; Huang, J.; Song, Y.; Li, B.; Wang, J.; Wang, H. Deep learning algorithms for brain disease detection with magnetic induction tomography. Med. Phys. 2020, 48, 745–759. [Google Scholar] [CrossRef]
  11. Harikumar, R.; Prabu, R.; Raghavan, S. Electrical impedance tomography (EIT) and its medical applications: A review. Int. J. Soft Comput. Eng. 2013, 3, 193–198. [Google Scholar]
  12. Kim, B.S.; Khambampati, A.K.; Kim, S.; Kim, K.Y. Electrical resistance imaging of two-phase flow using direct Landweber method. In Flow Measurement and Instrumentation; Elsevier: Amsterdam, The Netherlands, March 2015; Volume 41, pp. 41–49. [Google Scholar]
  13. Zhao, B.; Wang, H.X.; Chen, X.Y.; Shi, X.L.; Yang, W.Q. Linearized solution to electrical impedance tomography based on the schur conjugate gradient method. Measurement Science and Technology 2007, 18, 3373. [Google Scholar] [CrossRef]
  14. Han, M.; Xue, Y.; Qin, P. An improved image reconstruction algorithm based on iteration NR in magnetic induction tomography. Chin. J. Biomed. Eng. 2015, 34, 190–197. [Google Scholar]
  15. Liu, X.; Liu, Z.; Zhu, S. A modified conjugate gradient reconstruction algorithm for electromagnetic tomography. In Proceedings of the IEEE International Conference on Imaging Systems and Techniques, Krakow, Poland, 16–18 October 2018. [Google Scholar]
  16. Xiao, Z.; Tan, C.; Dong, F. 3-D Hemorrhage Imaging by Cambered Magnetic Induction Tomography. IEEE Trans. Instrum. Meas. 2019, 68, 2460–2468. [Google Scholar] [CrossRef]
  17. Lin, Z.; Guo, R.; Zhang, K.; Li, M.; Yang, F.; Xu, S.; Abubakar, A. Neural network-based supervised descent method for 2D electrical impedance tomography. Physiol. Meas. 2020, 41, 074003. [Google Scholar] [CrossRef] [PubMed]
  18. Xiao, J.; Liu, Z.; Zhao, P.; Li, Y.; Huo, J. Deep Learning Image Reconstruction Simulation for Electromagnetic Tomography. IEEE Sens. J. 2018, 18, 3290–3298. [Google Scholar] [CrossRef]
  19. Hu, D.; Lu, K.; Yang, Y. Image reconstruction for electrical impedance tomography based on spatial invariant feature maps and convolutional neural network. In Proceedings of the 2019 IEEE International Conference on Imaging Systems & Techniques, Abu Dhabi, United Arab Emirates, 9–10 December 2019. [Google Scholar]
  20. Denisov, I.V.; Kulchin, Y.N.; Panov, A.V.; Rybalchenko, N.A. Neural network methods of reconstruction tomography problem solutions. Opt. Mem. Neural Netw. 2015, 14, 45–58. [Google Scholar]
  21. Michalikova, M.; Abedt, R.; Prauzek, M.; Koziorek, J. Image Reconstruction in Electrical Impedance Tomography Using Neural Network. In Proceedings of the 2014 Cairo International Biomedical Engineering Conference, Giza, Egypt, 11–13 December 2014; pp. 39–42. [Google Scholar]
  22. Miller, A.S.; Blott, B.H.; Hames, T.K. Neural networks for electrical impedance tomography image characterisation. Clin. Phys. Physiol. Meas. 1992, 13 (Suppl. A), 119–123. [Google Scholar] [CrossRef]
  23. Li, J.; Yang, X.; Wang, Y.; Pan, R. An image reconstruction algorithm based on RBF neural network for electrical capacitance tomography. In Proceedings of the IEEE 2012 Sixth International Conference on Electromagnetic Field Problems and Applications, Dalian, China, 19–21 June 2012; pp. 1–4. [Google Scholar]
  24. Tan, C.; Lv, S.; Dong, F.; Takei, M. Image reconstruction based on convolutional neural network for electrical resistance tomography. IEEE Sens. J. 2019, 19, 196–204. [Google Scholar] [CrossRef]
  25. Kłosowski, G.; Rymarczyk, T. Using neural networks and deep learning algorithms in electrical impedance tomography. Inform. Autom. Pomiary Gospod. Ochr. Sr. 2017. [Google Scholar] [CrossRef]
  26. Zheng, J.; Peng, L. An autoencoder-based image reconstruction for electrical capacitance tomography. IEEE Sens. J. 2018, 18, 5464–5474. [Google Scholar] [CrossRef]
  27. Zheng, J.; Li, J.; Li, Y.; Peng, L. A benchmark dataset and deep learning-based image reconstruction for electrical capacitance tomography. Sensors 2018, 18, 3701. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Feng, J.; Sun, Q.; Li, Z.; Sun, Z.; Jia, K. Back-propagation neural network-based reconstruction algorithm for diffuse optical tomography. J. Biomed. Opt. 2018, 24, 051407. [Google Scholar] [CrossRef] [PubMed]
  29. Li, X.; Lu, Y.; Wang, J.; Dang, X.; Wang, Q.; Duan, X.; Sun, Y. An Image Reconstruction Framework Based on Deep Neural Network for Electrical Impedance Tomography. In Proceedings of the IEEE International Conference on Image Processing, Beijing, China, 17–20 September 2017; pp. 3585–3589. [Google Scholar]
  30. Chen, R.; Huang, J.; Wang, H.; Li, B.; Zhao, Z.; Wang, J.; Wang, Y. A Novel Algorithm for High-Resolution Magnetic Induction Tomography Based on Stacked Auto-Encoder for Biological Tissue Imaging. IEEE Access 2019, 7, 185597–185606. [Google Scholar] [CrossRef]
  31. Ren, S.; Sun, K.; Tan, C.; Dong, F. A Two-Stage Deep Learning Method for Robust Shape Reconstruction with Electrical Impedance Tomography. IEEE Trans. Instrum. Meas. 2020, 69, 4887–4897. [Google Scholar] [CrossRef]
  32. Han, C.; Rundo, L.; Araki, R.; Furukawa, Y.; Mauri, G.; Nakayama, H.; Hayashi, H. Infinite brain MR images: PGGAN-based data augmentation for tumor detection. In Neural Approaches to Dynamics of Signal Exchanges. SIST; Springer: Singapore, 2020; Volume 151, pp. 291–303. [Google Scholar] [CrossRef] [Green Version]
  33. Hussein, S.A.; Tirer, T.; Giryes, R. Image-adaptive gan based reconstruction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 3121–3129. [Google Scholar]
  34. Xu, C.; Xu, L.; Ohorodnyk, P.; Roth, M.; Chen, B.; Li, S. Contrast agent-free synthesis and segmentation of ischemic heart disease images using progressive sequential causal GANs. Med. Image Anal. 2020, 62, 101668. [Google Scholar] [CrossRef]
  35. Han, C.; Rundo, L.; Murao, K.; Noguchi, T.; Shimahara, Y.; Milacski, Z.Á.; Satoh, S.I. MADGAN: Unsupervised medical anomaly detection GAN using multiple adjacent brain MRI slice reconstruction. BMC Bioinform. 2021, 22, 31. [Google Scholar] [CrossRef]
  36. Liu, X.; Gherbi, A.; Wei, Z.; Li, W.; Cheriet, M. Multispectral image reconstruction from color images using enhanced variational autoencoder and generative adversarial network. IEEE Access 2020. [Google Scholar] [CrossRef]
  37. Yuan, Z.; Jiang, M.; Wang, Y.; Wei, B.; Li, Y.; Wang, P.; Yang, G. SARA-GAN: Self-Attention and Relative Average Discriminator Based Generative Adversarial Networks for Fast Compressed Sensing MRI Reconstruction. Front. Neuroinform. 2020, 14. [Google Scholar] [CrossRef]
  38. Chen, Z.; Yuan, Q.; Song, X.; Chen, C.; Zhang, D.; Xiang, Y.; Xuan, Q. Mitnet: Gan enhanced magnetic induction tomography based on complex cnn. arXiv 2021, arXiv:2102.07911. [Google Scholar]
Figure 1. The training process of MIT-GAN model.
Figure 1. The training process of MIT-GAN model.
Sensors 21 03869 g001
Figure 2. The model of 8-channel MIT signal acquisition system.
Figure 2. The model of 8-channel MIT signal acquisition system.
Sensors 21 03869 g002
Figure 3. The hyperparameters and activation functions of GAN. The voltage data is input into the generator of the 8-layer structure, its results are compared with the real conductivity distribution by the discriminator.
Figure 3. The hyperparameters and activation functions of GAN. The voltage data is input into the generator of the 8-layer structure, its results are compared with the real conductivity distribution by the discriminator.
Sensors 21 03869 g003
Figure 4. The average value of evaluating indicators, under different algorithms and SNR, as follows: (a) RMSE, (b) SSIM, (c) PSNR, and (d) CC.
Figure 4. The average value of evaluating indicators, under different algorithms and SNR, as follows: (a) RMSE, (b) SSIM, (c) PSNR, and (d) CC.
Sensors 21 03869 g004
Figure 5. The image reconstruction under different noise and algorithm.
Figure 5. The image reconstruction under different noise and algorithm.
Sensors 21 03869 g005
Figure 6. The image reconstruction under different noise and algorithm.
Figure 6. The image reconstruction under different noise and algorithm.
Sensors 21 03869 g006
Table 1. The RMSE of two models under different noise and algorithm.
Table 1. The RMSE of two models under different noise and algorithm.
ModelModel 1Model 2
SNR/80 dB40 dB20 dB/80 dB40 dB20 dB
Tikhonov0.47280.48000.48480.47580.45720.45900.45470.4523
BPNN0.33580.33350.33930.33540.35010.34980.35200.3521
GAN0.07530.09390.10750.12960.17060.17760.16810.1726
Table 2. The SSIM of two models under different noise and algorithm.
Table 2. The SSIM of two models under different noise and algorithm.
ModelModel 1Model 2
SNR/80 dB40 dB20 dB/80 dB40 dB20 dB
Tikhonov0.46410.45470.45790.44820.43860.45900.44550.4500
BPNN0.67510.64320.59670.51470.70290.68880.64300.5163
GAN0.92160.88470.83560.73110.85170.84040.84270.8143
Table 3. The PSNR of two models under different noise and algorithm.
Table 3. The PSNR of two models under different noise and algorithm.
ModelModel 1Model 2
SNR/80 dB40 dB20 dB/80 dB40 dB20 dB
Tikhonov6.50616.37526.28796.45136.79876.76316.84616.8896
BPNN9.47769.53769.38769.48949.11619.12449.06989.0658
GAN22.459820.542819.365717.751015.356915.008315.486215.2568
Table 4. The CC of two models under different noise and algorithm.
Table 4. The CC of two models under different noise and algorithm.
ModelModel 1Model 2
SNR/80 dB40 dB20 dB/80 dB40 dB20 dB
Tikhonov0.57700.56990.58580.55380.58190.57850.58990.5890
BPNN0.79740.75810.70640.65420.82760.81600.77080.6663
GAN0.92540.88420.84100.75760.88070.87030.85860.8180
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, D.; Liu, J.; Wang, Y.; Xu, B.; Wang, X. Application of a Generative Adversarial Network in Image Reconstruction of Magnetic Induction Tomography. Sensors 2021, 21, 3869. https://doi.org/10.3390/s21113869

AMA Style

Yang D, Liu J, Wang Y, Xu B, Wang X. Application of a Generative Adversarial Network in Image Reconstruction of Magnetic Induction Tomography. Sensors. 2021; 21(11):3869. https://doi.org/10.3390/s21113869

Chicago/Turabian Style

Yang, Dan, Jiahua Liu, Yuchen Wang, Bin Xu, and Xu Wang. 2021. "Application of a Generative Adversarial Network in Image Reconstruction of Magnetic Induction Tomography" Sensors 21, no. 11: 3869. https://doi.org/10.3390/s21113869

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop