Next Article in Journal
K-Anonymity Privacy Protection Algorithm for Multi-Dimensional Data against Skewness and Similarity Attacks
Next Article in Special Issue
Effect of LED Lighting on Selected Quality Parameters of Electricity
Previous Article in Journal
A Two-Step Feature Selection Radiomic Approach to Predict Molecular Outcomes in Breast Cancer
Previous Article in Special Issue
Field Modeling of the Influence of Defects Caused by Bending of Conductive Textronic Layers on Their Electrical Conductivity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Machine Learning Algorithms to the Discretization Problem in Wearable Electrical Tomography Imaging for Bladder Tracking

by
Bartłomiej Baran
1,
Edward Kozłowski
2,
Dariusz Majerek
3,
Tomasz Rymarczyk
1,4,*,
Manuchehr Soleimani
5 and
Dariusz Wójcik
1,4
1
Research & Development Centre Netrix S.A., 20-704 Lublin, Poland
2
Faculty of Management, Lublin University of Technology, 20-618 Lublin, Poland
3
Faculty of Fundamentals of Technology, Lublin University of Technology, 20-618 Lublin, Poland
4
WSEI University, 20-209 Lublin, Poland
5
Department of Electronic and Electrical Engineering, University of Bath, Bath BA2 7AY, UK
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(3), 1553; https://doi.org/10.3390/s23031553
Submission received: 13 December 2022 / Revised: 27 January 2023 / Accepted: 29 January 2023 / Published: 31 January 2023

Abstract

:
The article presents the implementation of artificial intelligence algorithms for the problem of discretization in Electrical Impedance Tomography (EIT) adapted for urinary tract monitoring. The primary objective of discretization is to create a finite element mesh (FEM) classifier that will separate the inclusion elements from the background. In general, the classifier is designed to detect the area of elements belonging to an inclusion revealing the shape of that object. We show the adaptation of supervised learning methods such as logistic regression, decision trees, linear and quadratic discriminant analysis to the problem of tracking the urinary bladder using EIT. Our study focuses on developing and comparing various algorithms for discretization, which perfectly supplement methods for an inverse problem. The innovation of the presented solutions lies in the originally adapted algorithms for EIT allowing for the tracking of the bladder. We claim that a robust measurement solution with sensors and statistical methods can track the placement and shape change of the bladder, leading to effective information about the studied object. This article also shows the developed device, its functions and working principle. The development of such a device and accompanying information technology came about in response to particularly strong market demand for modern technical solutions for urinary tract rehabilitation.

1. Introduction

Lower urinary tract diseases are on the rise and reduce the quality of life of those affected. It is estimated that up to 50% of the population will suffer from various forms of incontinence at some stage in their lives [1,2]. In some cases, there will be a spontaneous regression, but approximately 70% of this group will develop persistent urinary incontinence of varying severity. Approximately 50% of people with incontinence use pads only to prevent leakage and soiling of underwear and do not attempt self-treatment. In 15% of patients with severe urinary incontinence, surgical treatment is resorted to, which improves the quality of life to varying degrees. However, according to the literature in the field of functional disorders of the urinary system [3,4], it is argued that the rehabilitation of the muscles responsible for the efficient emptying of the urinary tract should be performed both before and after a surgical procedure. Nowadays, the treatment of bladder dysfunction based on biofeedback and EMG therapy is a universal standard. However, the devices currently available on the market struggle with poor comfort and mobility, which does not allow for daily and regular rehabilitation. In contrast to static imaging with ultrasound, CT or nuclear magnetic resonance, EIT allows continuous (up to a dozen hours) diagnostics of urinary tract function. The great advantages of this method are its low invasiveness, long-term evaluation of urinary tract function and the relatively low price of performing the test [5,6,7,8,9,10].
To face the aforementioned problems, we developed a device for non-invasive monitoring and diagnosis of lower urinary tract functional disorders, as presented in Figure 1. The device will enable the measurement of muscle tension (EMG) with the possibility of electrostimulation and biofeedback-type therapy. At the same time, the device has implemented a system for visualization of the urinary tract based on EIT and computational intelligence algorithms. Due to its compact size, the device can be used in specialized healthcare centers and at home.
In this article, we briefly introduce the concept of the developed device equipped with the EIT imaging technique. Such a device can be equipped with optional EMG diagnostics of muscle and nerve function or electrostimulation to allow for the required muscle contractions. Firstly, we describe the principle of operation of the EIT system and the equipment required to create such a device. Then, we present and compare various mathematical algorithms for discretizing FEM elements. In principle, the main idea is to construct a classifier allowing for assigning the FEM components to a class consisting of the inclusion or background elements. Our study focuses on the reconstruction of the urinary bladder with EIT using experimental data and applying machine learning algorithms. Similar studies have been conducted in this field, but the results were obtained with a simulated data set [5,6,11], or reconstruction has been obtained with measurement data using the Gauss–Newton method [12], known for its low-quality results with experimental data frames.

2. Materials and Methods

EIT is an imaging technique that exploits the different electrical properties of materials [13,14,15,16,17,18,19,20,21,22,23,24,25]. In this method, a source of electric voltage is connected to an object, resulting in current flows through its interior and electrical potential distributions at the surface of that subject item. The gathered information is processed by an algorithm that reproduces the internal impedance distribution. Such a method has a relatively low spatial image resolution. The difficulty in obtaining high resolution is primarily due to the limited number of measurements, the nonlinear current flow through the studied medium and the insufficient sensitivity of the voltage measurement apparatus to changes in conductivity in a given area. Image reconstruction is very sensitive to pervasive modeling errors, which are caused by inaccurately derived auxiliary variables of the measurement model. In practice, first and foremost, if the shape of the object is inaccurately known, errors in its modeling cause divergence from the real object. From a mathematical point of view, EIT belongs to the category of inverse problems designed to study the distribution of electromagnetic fields. An inverse problem is a procedure of traceability, development or synthesis in which the parameters describing the electrical field are predicated based on having some information specific to it. Such problems are challenging and cumbersome to analyze. This is because there is usually not enough or a redundancy of information, which is sometimes contradictory or linearly dependent. The numerical examination of the physical problem based on EIT is solved using the finite element method (FEM).

2.1. Measurement System

The device presented in Figure 1b is equipped with 16 channels for measuring the surface potential of the body with an accuracy of 1 μ V. Each of the channels is independently programmable (measurement and stimulation parameters), allowing one to perform a current simulation arbitrarily. During a single measurement, two of the attached electrodes are used as charge injectors forcing the flow of AC current of constant amplitude through the object under examination, while the remaining electrodes work as measuring electrodes probing electrical voltages at points of contact with the object. After data are collected from one sampling, the injecting electrodes are switched to the next in the sequence, and the procedure repeats until data are collected from all specified combinations. The developed prototype is a compact, single-board EIT scanner capable of performing measurements on 16 electrodes in any desired configuration. In our setup, the electrodes are arranged in two rows of eight electrodes, which allow for 2D (8 sensors) or 3D (16 sensors) reconstruction by solving the inverse problem. The stimulation between injecting electrodes is performed with a current in m A range at 100 k Hz . The injection appears only between electrodes from the same row. As a result, the device records 256 different voltages in a single attempt.
In detail, the device consists of a current source, a current measurement module, a voltage measurement module, a set of multiplexers and a control unit. The central unit presented in Figure 2 is designed to fit into a modest ergonomic case and constructed with multilayer circuits and assembled with BGA soldering. The block structure of this solution is presented in Figure 3. The power source is connected to any electrode pair via an analog dual 16-channel multiplexer. It consists of two independent digital-to-analog converters where the voltage output of one of the converters generates the shape of the forced current waveform, and the voltage output of the other is the reference voltage. The converters are controlled from the FPGA via a serial bus. Such a solution allows precise control of the waveform shape and its amplitude. The current flow is tested on the measurement resistor using a high-end ADC along with the necessary signal conditioning circuits. The circuit has programmable gain, pre-filtering, and differential signal-forming capability. The ADC sends the data to the FPGA via a parallel bus. Based on the reading of the value from the converter, the logic regulates the value of the current obtained from the current source. The electrode voltages are tested with a measurement module consisting of a series of signal conditioning circuits and an analog-to-digital converter similar to current measurement. The electrodes are connected directly to a set of preamplifiers, followed by a 16-channel multiplexer. The multiplexed signal is transferred to the amplifier with a programmable enhancement, which is equipped with an additional module that enables the signal conversion to a differential signal. The next step transfers the signal to an analog-to-digital converter through a filter. The control unit in the device is an Altera Cyclone IV FPGA with 144 leads and 10,000 logic elements. It controls all processes in the measurement cycle and is used for data acquisition and transfer. The electronics and housing are designed to fulfill the standard of EMC requirements. It was ensured that the electronics did not have a negative impact on the surroundings and were not exposed to electrostatic discharges from the environment.
This high-end 16-bit digital-to-analog converter has a sampling rate of up to 25 Msps, a signal conditioning circuit containing a measurement amplifier with adjustable gain and a differential amplifier that adjusts the signal under test to the inputs of the ADC accordingly. The main component of the measurement part is the LTC2202 analog-to-digital converter, a 16-bit ADC with a sampling frequency of 10 Msps. This chip has a built-in PGA front-end that allows changing the input range 1× or 1.5×, is powered by a single 3.3 V voltage and has a high-speed parallel interface. A signal conditioning circuit consisting of a measurement amplifier and a differential amplifier has been used at the input of the transmitter, allowing 1–100× gain adjustment and signal adjustment to the differential input of the transmitter. Digital gain control is realized with a digital potentiometer in an adjustable resistor circuit with a resolution of 10-bit.
The device supports communication with a master device via Bluetooth or Wi-Fi and integrates with PCs running the Microsoft Windows operating system and mobile devices with Google Android or Apple iOS. The device is equipped with software for superior devices that performs the function of control and programming of devices as well as visualization and archiving of test results. An optional module with a touch screen allows you to control the operation of the device and change its settings without the need to use a computer. This solution will be equipped with mechanisms allowing for remote control of rehabilitation settings and control of its results.

2.2. Measurement Data Acquisition

The measurement data are collected and aggregated by the device, then transmitted as data packets to an external unit. Figure 4 shows an example of a measurement obtained on a patient under study. It shows voltages on the body surface (blue line) and corresponding electrical potential differences between injection and ground electrodes (green line). This frame presents measurement data recorded from 16 measurement electrodes. When the device is connected to a computer via the USB port, the delay in communication is close to 6 ms; using wireless communication, the delay in sending commands is around 50 ms. At the current stage of development, the designed device is able to perform 2.5 measurements per second.

3. Discriminant Algorithms

This section includes descriptions of various mathematical models used to solve the discrimination problem. We begin by defining the problem of the discrete dataset consisting of the mesh elements representing the torso cross section at the urinary bladder level. For the mesh, a dataset was prepared from the solution of the forward problem, assuming different shapes and positions of inclusions. As a result, the dataset contains 5000 different examples of simulated measurements.
In the further section, we concentrate on the description and theoretical aspects of the implemented methods. We present the algorithms based on the linear regression supported by elastic net regularization, the linear, quadratic and regularized discriminant analysis with the possibility of application of principal component analysis (PCA). Moreover, we include a short description of the decision tree method.

3.1. The Finite Element Mesh and Discretization Problem

The main purpose of discriminant analysis (pattern classification) is to find a classification rule [13,14]. This process involves identifying eligibility for a certain class based on observations of the independent variable. The decision on class membership is made based on knowledge of the distribution of the independent variable and the distribution of class a priori. In the case of EIT, the elements of the field of view are a collection of such a class. Figure 5 presents the finite element mesh composed of 848 nodes and 1555 triangles. This model uses eight line electrodes that are placed in front of the torso.
Such a discrete set of finite elements with different inclusion positions and shapes is a learning dataset for the models introduced below. This learning dataset can be written as D = x i , y i : x i R m , y i { 0 , 1 } , 1 i n . The components of the sequence x i 1 i n belong to two classes, where membership to the class is expressed as y i { 0 , 1 } for 1 i n . In the analyzed case of the presence of an inclusion for a finite element (a pixel in the field of view), we assume y i = 1 , while if the element does not represent an inclusion, we take y i = 0 . Analyzing the signal received from the sensors x R m , it is possible to classify each finite element. Logistic regression was used to create a classifier (finite element mapping) f : R m { 0 , 1 } .

3.2. Logistic Regression (LR)

Let Ω , F , P be a probabilistic space and Y a random variable with a discrete distribution Y : Ω 0 , 1 . Then, the ratio of the probability of success to the probability of failure can be defined as
θ X = P Y = 1 X 1 P Y = 1 X .
The main purpose of logistic regression (LR) is to evaluate the probability of success P Y = 1 X , where X denotes the realization of predictors [13,15,16]. It can be assumed that
P Y = 1 X = p X .
Knowing that the probability of success is p X 0 , 1 , from Equation (1) it follows that the chance of a success is θ X 0 , , while the log-odds (also known as logit) are ln θ X , .
Using the logistic regression method, we analyze the linear dependence of the logarithm on the independent variables X. For this purpose, we examine the correlation specified by the following expression
ln θ X = ln p ( X ) 1 p ( X ) = X β + ε ,
where ε is a random variable with a normal distribution N 0 , σ 2 and an estimator β = β 1 , β 2 , β 3 , . . . , β m R m .
Using Equation (3), we can derive the probability of success as
p ( β , X ) = d e f p ( X ) = e X β 1 + e X β .
In order to estimate β parameters, maximum-likelihood estimation (MLE) was used [17].

Logistic Regression with Elastic Net Regularization

Due to the collinearity of independent variables (measurements obtained from the sensors), certain regularization methods should be applied. In our work, we focus on elastic net regularization [18], which is a linear combination of LASSO (least absolute shrinkage and selection operator) regression and ridge regression, called Tikhonov regularization [19,20].
To determine the linear regression parameters in the model defined in Equation (3), one needs to solve the following relation
max β i = 1 N y i x ( i ) β ln 1 + e x ( i ) β λ P α β ,
where λ > 0 , 0 α 1 and P α represents the penalty given by the formula
P α β = α β L 1 + 1 α 2 β L 2 = j = 1 p α β j + 1 α 2 β j 2 .
The penalty P α β is the linear combination of the norms of the vector of estimators β in L 1 and L 2 spaces. For α = 0 , we have ridge regression, while for α = 1 we obtain LASSO regression. In order to adjust the regularization parameters, Equation (5) is solved for each finite element. First, the sequence of possible values of the λ parameter is determined. Then, for different values of the regularization parameter, β coefficients are determined using the K-fold cross-validation method. As λ parameter and β parameter estimators, we take such values for which the cross-validation estimate of error is smallest (in the case under consideration, we chose such values for which the cross-validation estimate of accuracy was highest).

3.3. Linear Discriminant Analysis (LDA)

For the given learning dataset D, the probabilistic space Ω , F , P and random variable Y, we can define the a priori distribution for Y as
π k = n k n ,
where n k = # i : y i = k , and π 0 + π 1 = 1 . For k 0 , 1 , we construct a decision rule on the basis of Bayes’ theorem
P Y = k | X = x = P X = x | Y = k π k j = 0 1 P X = x | Y = j π j .
Determining class affiliation based on Equation (8), we compare values P X = x | Y = k π k . A larger value of this product means a higher probability that the random variable Y will take the value k (i.e., the observation x belongs to the k class) [13,14,16]. Considering the linear discriminant analysis, we assume that the covariance matrix of the random variable X for each group is identical, i.e., Σ 0 = Σ k = Σ . The conditional distribution of random variable X belonging to the class k, k 0 , 1 is given as a formula
f k ( x ) = P X = x | Y = k = 1 ( 2 π ) m / 2 | Σ | exp 1 2 x μ k T Σ 1 x μ k .
To compare the probabilities of belonging to two different classes presented in Equation (8), it is enough to analyze the logarithm of the quotient of these probabilities, i.e.,
log P Y = 1 | X = x P Y = 0 | X = x = log π 1 log π 0 1 2 μ 1 T Σ 1 μ 1 + 1 2 μ 0 T Σ 1 μ 0 + x T Σ 1 μ 1 x T Σ 1 μ 0 .
Moreover, for class k 0 , 1 , we can introduce a linear discriminant function in the form
δ i ( x ) = log π i + x T Σ 1 μ i 1 2 μ i T Σ 1 μ i
and a definition of the plane separating the two classes as
H = x R m : P Y = 1 | X = x = P Y = 0 | X = x .
Combining Equations (10) and (11), we obtain the plane H
H = x R m : δ 1 ( x ) = δ 0 ( x ) ,
which splits the entire space R m into two separable sets, where the membership of an observed signal in a set is equivalent to membership in the corresponding class. Based on the introduced properties, the decision rule can be expressed in the form
Y ^ = 1 , δ 1 ( x ) δ 0 ( x ) , 0 , δ 1 ( x ) < δ 0 ( x ) = a r g m a x k { 0 , 1 } δ k ( x ) .
As the estimators of the unknown parameters of the distributions of observations for each class, we determine:
  • expected values
    μ ^ k = 1 n k i : y i = k x i
    for k 0 , 1 ;
  • covariance matrix
    Σ ^ = 1 n 2 k = 0 1 i : y i = k x i μ k ^ x i μ k ^ T .
It is important to remember that in the EIT experiment, the predictors are highly correlated, so to overcome the problem of the singularity of the Σ matrix, it is necessary to use regularization techniques [21,22].

3.4. Quadratic Discriminant Analysis (QDA)

Quadratic discriminant analysis (QDA) presents an alternative approach to defining the classification rule. Identical to the LDA, we assume that the conditional distribution f k ( x ) = P X = x | Y = k of a random variable X is a normal distribution N μ k , Σ k for k 0 , 1 . The difference is that for LDA we assume the identical covariance matrix of the random variable X for each group, i.e., Σ 0 = Σ 1 = Σ , while in the case of QDA, they are different. Therefore, for each class we determine the expected value vector and covariance matrix:
μ ^ k = 1 n k i : y i = k x i ,
Σ k ^ = 1 n k 1 i : y i = k x i μ k ^ x i μ k ^ T ,
where k 0 , 1 .
Comparing the probabilities of belonging to two different classes in Equation (8), we analyze the logarithm of the quotient of these probabilities, i.e.,
log P Y = 1 | X = x P Y = 0 | X = x = log π 1 1 2 x μ 1 T Σ 1 1 x μ 1 1 2 log det ( Σ 1 1 ) log π 0 + 1 2 x μ 0 T Σ 0 1 x μ 0 + 1 2 log det ( Σ 0 )
Just as previously for LDA, we can define a quadratic discriminant function δ i ( x ) for QDA and class k 0 , 1 in the following form
δ i ( x ) = log π k 1 2 x μ k T Σ k 1 x μ k 1 2 log det ( Σ k 1 ) .
For QDA, the decision rule is expressed by Equation (14).

3.5. Regularized Discriminant Analysis

When predictors are highly correlated with each other, the prediction with LDA and QDA models is unstable. Friedman proposed a compromise (in the sense of co-integration) between LDA and QDA that allows the covariance for QDA models to shrink toward LDA [23]. Such a technique is called regularization, while the method is referred to as RDA (regularized discriminant analysis). This method is very similar to ridge regression, allowing for shrinkage of covariances between characteristics, namely the covariance matrices for each class are referred to as
Σ k α = α Σ k + ( 1 α ) Σ
for k 0 , 1 and α 0 , 1 .
In practice, the α parameter is chosen so that the classification error when applying the model to validation data (or using cross-validation) is as small as possible. For LDA models, the identical regularization technique can also be used; the covariance matrix is determined by the formula
Σ γ = γ Σ + ( 1 γ ) σ 2 I ,
where γ 0 , 1 and I R m × m is an identity matrix. In the results section, the RDA was used in the discrimination, where Equation (20) was used to shrink the covariance matrix.

3.6. Principal Component Analysis (PCA)

As already described, our X signals can be highly correlated. Another method of eliminating redundancy could be principal component analysis. It relies on identifying the factors (components) present in a dataset by creating linear combinations of the original variables in such a way that the new components explain the largest part of the variation in the original space [13,24,25]. We call the coordinates of the new system loads of the created principal components. In the new auxiliary space, most variability is explained by the initial factors. PCA is often used to reduce the size of a statistical dataset by discarding the last factors [15].

3.7. Decision Trees

A decision tree is a hierarchical structure representing a classification or regression model. They are used especially often when the functional form of the correlation between predictors and the outcome variable is unknown or hard to determine. Each decision tree consists of a root, nodes and leaves. The root is called the initial node of the tree, from which subsequent descendant nodes are formed through divisions. The terminal nodes that do not undergo divisions are called leaves, and the lines connecting the nodes are called branches.
If the tree is used for classification tasks, the leaves contain information about which class in a given sequence of subdivisions is most likely to occur. On the other hand, if the tree is for regression purposes, the leaves contain conditional measures of the outcome’s central tendency (usually the mean). The condition represents a series of divisions leading to a given terminal node (leaf). In both cases (classification and regression), the tree tends to such a division that successive nodes and leaves are as homogeneous as possible regarding the outcome variable.
There are a variety of types of splitting rules used in decision trees. They are selected so that subsequent nodes are characterized by less impurity. In classification trees, the most commonly used measure of impurity is the Gini index
G i T n ( c | t ) = x R t | T n , t = r | | T n | G i T n , t = r ( c ) ,
where
G i T n , t = r ( c ) = d C P T n , t = r ( c = d ) · ( 1 P T n , t = r ( c = d ) ) = 1 d C P T n , t = r 2 ( c = d ) .
Among the biggest advantages of using decision trees are:
  • Easy to interpret;
  • Do not require tedious data preparation (no standardization, introduction of binary variables, allows for missing data);
  • Potential non-linearity of the relationship between the outcome variable and the predictors;
  • Robust to deviations from assumptions;
  • Allows for a quick analysis of large data sets.
The biggest disadvantage of single decision trees is low predictive power, especially in complex tasks. In such cases, it is recommended to use tree ensembles in the form of random forests, bagging or boosting.

3.8. Measures of Fit Assessment

For each element in the dataset D, based on the readings of X, we determine the probability of inclusion P Y = 1 | X based on model prediction, assuming
Belonging to the area = inclusion , P Y = 1 | X t background , P Y = 0 | X < t
for the threshold t 0 , 1 . The most commonly accepted threshold value is 0.5. The elementary terminology and factors describing the recognition of inclusions in the field of view are presented below. We take the lack of inclusion in a finite element location as a negative case (N), while the presence of inclusion is a positive case (P). The confusion matrix should be specified with the following values: TP (true positive), the number of finite elements for which inclusions were correctly recognized; TN (true negative), the number of finite elements for which the lack of inclusion was correctly identified; FP (false positive), the number of elements without inclusions, which are assigned to have inclusions (false alarm); FN (false negative), the number of finite elements with inclusions, for which they were considered to have no inclusions (see Table 1).
We use the standard definition of basic fit measures as follows [26]:
A c c u r a c y = T P + T N T P + T N + F P + F N ,
T r u e P o s i t i v e R a t e = S e n s i v i t y = T P T P + F N ,
S p e c i f i c i t y = 1 F a l s e P o s i t i v e R a t e = T N T N + F P ,
P o s i t i v e P r e d i c t i v e V a l u e = T P T P + F P ,
N e g a t i v e P r e d i c t i v e V a l u e = T N T N + F N ,
P r e v a l e n c e = T P + F N T P + T N + F P + F N ,
D e t e c t i o n R a t e = T P T P + T N + F P + F N ,
D e t e c t i o n P r e v a l e n c e = T P + F P T P + T N + F P + F N ,
B a l a n c e d A c c u r a c y = S e n s i v i t y + S p e c i f i c i t y 2 .
The use of so many measures of model fit was dictated by the fact that each measure exposes a different aspect of model fit. Using them complementarily helps assess the model’s performance.
In the EIT image reconstruction, it is also necessary to describe the ability to find inclusions in the field of view. To evaluate the ability of the classifier based on the logistic regression (see e.g., [26,27]), we determine the curve describing the receiver operating characteristic (ROC curve). This curve illustrates the relationship between sensitivity and specificity for different threshold levels. The diagonal in the ROC figure describes a strategy based on guessing inclusions during reconstruction. When the ROC is above the diagonal, it means that the recognition technique is clearly better than guessing. The area under the ROC curve in the literature is called AUC (area under ROC curve) and denotes a measure of predictability. This quantity is also included in the tables describing the reconstructions.
To determine the credibility of performed discretizations (consistency between the inclusion and prediction), we rely on Cohen’s ratio κ defined as follows:
κ = 2 ( T P · T N F N · F P ) ( T P + F P ) ( F P + T N ) + ( T P + F N ) ( F N + T N ) ,
where κ [ 0 , 1 ] . A larger value of κ determines greater consistency between inclusion and discretization results.
On the other hand, to verify an inconsistency between inclusion and discretization result, we use McNemar’s test defined as follows:
χ 2 = ( F P F N ) 2 F N + F P ,
This statistical test compares the sensitivity and specificity of discretization result.

4. Results

In this section, we present the results of the model classification and the tables with the measures of fit describing the performance of the acquired discretizations. For all of the specified methods, we present the example of the result obtained for different regularization approaches and compare the inclusion pattern corresponding to the used simulated data frame.
The single data frame consists of 32 elements for 2D simulated voltage at the edge of the model presented in Figure 5. Due to the fact that in EIT the measurements of the predictors are highly correlated, we present a covariance matrix in Figure 6 confirming this statement.
In the last part of this section, we present the application of prepared algorithms to the case of a healthy male in his 20s with an almost full bladder. We show a comparison of results for all introduced discretization methods.

4.1. Results for Logistic Regression

In Figure 7, we show the results obtained for logistic regression with applied regularization methods, where the higher brightness of the mesh element defines a bigger probability of such an element belonging to the inclusion. It can be noted that the result with ridge or LASSO regularization roughly defines the position and shape of the inclusion.
However, the result for ridge regularization is reasonably good and represents the object’s center. The better result gives a combination of both methods. We observe visible image quality enhancement. In this result, a constriction of the inclusion area towards the center of the inclusion can also be noted.
Figure 8 shows the ROC curve for the results in Figure 7 for the given regularization methods. It can be seen that the lines for elastic net and ridge regularization overlap relatively well, where the ROC curve for LASSO differs from the other two. This result shows that the best fit is obtained by ridge regularization, and it is the main component of the discretization classifier that defines its diagnostic ability.
Table 2 shows the values of the basic coefficients, Cohen’s ratio, and McNemar’s test for LR and every regularization method. The κ ratio shows that the highest credibility of performed classifications is obtained by the LR method with Elastic Net regularization. On the other hand, the smallest χ 2 inconsistency between the snapshot and discretization is also performed using this method.

4.2. Results for Linear and Quadratic Discriminant Methods

Figure 9 shows the results obtained for the discriminant methods without and with applied dimensionality reduction. The top left panel in Figure 9 represents the inclusion pattern. The next panels show the results for LDA, QDA and RDA, respectively. It can be noted that the predictions of those classifiers indicate the position of the inclusion very well. Additionally, the QDA method almost perfectly illustrates the shape of the predicted object.
The presence of noise can also be noted, with a randomly higher probability for single elements near the mesh boundary in the result for the LDA and RDA methods. The last two panels in Figure 9 represent the LDA and QDA methods with applied principal components analysis (PCA). In the result, we observe the improvement of the discretization quality for LDA, where the QDA application of the PCA does not significantly affect the final probability distribution. Comparing the results in Figure 9, one can notice that the QDA with PCA is the best match of the prediction with an assumed inclusion pattern qualitatively. A closer look at PCA, presented in Figure 10, reveals that the first eight principal components can fully explain the variance.
In Figure 11, we show the ROC curve for the results presented in Figure 7, including the dimensionality reduction method. Note that all of the used approaches result in the high-quality discretization of FEM components. This graph shows that the RDA has the lowest diagnostic ability compared with the other methods.
In Table 3, we present the fit measures for every discriminant method. The κ value indicates that QDA obtains the best result with the applied PCA. However, the lowest inconsistency χ 2 is provided by QDA without the application of PCA. This inconsistency is visible in Figure 9, where the modest difference between the result of QDA and QDA with applied PCA can be noted, resulting in extra mesh elements (classified as the object) at the bottom of the predicted inclusion.

4.3. Results for Classification and Regression Trees

In Figure 12, we present the use of regression trees in our classification problem. This result clearly shows the high alignment between reconstruction and inclusion pattern; we can observe almost a one-to-one mapping. In this example, we can notice particular elements incorrectly classified as pattern elements. However, comparing this graph with the results obtained by previous methods, it can be seen that the decision trees algorithm shows the best reconstruction quality. Since CART is robust to collinearity, none of the previously mentioned methods of regularization or dimensionality reduction are needed.
In detail, Figure 13 shows the ROC curve for the results presented in Figure 12. This result indicates high-quality discretization of FEM components.
In Table 4, we present the values of basic coefficients for the decision trees method. The value of the κ ratio confirms the highest accuracy of the reconstruction, and the small inconsistency defined by χ 2 is also achieved.

4.4. Reconstruction Performance

In order to assess the reconstruction performance, the following measures were used:
M S E = 1 n m i = 0 m 1 j = 0 n 1 I ( i , j ) K ( i , j ) 2 ,
M A E = 1 n m i = 0 m 1 j = 0 n 1 I ( i , j ) K ( i , j ) ,
P S N R = 10 · log 10 max x i 2 M S E ,
S S I M = ( 2 μ I μ K + c 1 ) ( 2 σ I K + c 2 ) ( μ 1 2 + μ 2 2 + c 1 ) ( σ 1 2 + σ 2 2 + c 2 ) ,
where μ I , μ K are means of I, K; σ 1 2 , σ 2 2 are variances of I, K; σ I K is the covariance matrix of I, K; c 1 , c 2 are some constants; and I i K are the images of N × N . Assuming that I ( i , j ) and K ( i , j ) are values of ( i , j ) pixel in the original and reconstructed image I, K, respectively. These are well-known metrics used in assessing reconstruction quality [28,29,30].
In Table 5, we present the values of measures for the introduced algorithms and compare them with the values for selected literature models. Reconstructions obtained using different machine learning techniques produce different results in the context of the adopted measures of reconstruction quality. The best reconstructions were obtained using decision trees (all measures have the best results). QDA and QDAPCA methods give slightly worse results in the context of all analyzed measures. The worst results were obtained using ridge regression and LASSO regularization models.
Comparing the results with those obtained in other works [31,32,33,34,35,36,37,38] in the field of EIT reconstruction using machine learning methods, it can be concluded that at least some of the models presented in this work (especially the CART model) dominate the published achievements in terms of the obtained measures of reconstruction quality.
The best of our models (CART) outperforms most models in the literature in terms of the PSNR measure, except for TN-Net, which has a slightly higher value. In contrast, none of the published models tops our models in terms of the SSIM measure. The less frequently used MSE measure, on the other hand, shows that some models such as DNN, CNN and TN-Net are better than most of our models except for the CART model, which also dominates in terms of this measure.

4.5. Application of the Algorithms to a Real Case

The comparison of the results obtained using the developed discretization methods with the actual measurement data is shown in Figure 14. The first row contains results for logistic regression with regularization methods, the second row contains LDA, LDA with PCA and RDA, and the last row contains data for QDA, QDA with PCA and CART. These results show that a finite element mesh represents the probability distribution of finding a bladder in the abdominal cavity.
In the results, it can be seen that some methods do not give sufficient results due to the discontinuity of the reconstructions and the irregularity of their shape. For example, the logistic regression method gives a reconstruction with an oval shape of the studied object, while elastic net and LASSO regularization result in a hollow area inside the bladder. Moreover, it can be noted that the object reconstructed by logistic regression is too close to the abdominal wall, which is biologically impossible due to the presence of abdominal tissue.
The second row of the results presented in Figure 14 shows that the LDA method does not give reliable reconstructions regardless of regularisation. We observe that the reconstruction does not create a uniform structure and is divided into many fragments. It discredits the algorithm as a reliable source of information. Similar behavior is observed with RDA.
The last line in Figure 14 contains the results for QDA, QDA with PCA and CART. The sensitivity of QDA methods allows to sufficiently capture the shape of the bladder. Due to its real shape, it can be expected that the probability distribution should be larger in the center of the object and smaller outside. This result is clearly visible in the reconstruction plots. Contrary to the good quality of reconstructions for QDA methods, the results obtained with the CART method are far from acceptable. We observe that the reconstruction obtained with CART yields several separate regions. It is quite surprising because the CART algorithm has a very high efficiency of reconstruction obtained from simulated measurements (see Table 4).

5. Discussion

The main objective of our study is to find the best classifier for discrete elements of the introduced FEM and real measurements. In this work, logistic regression, linear discriminant analysis, quadratic discriminant analysis and decision trees were used to reconstruct the inclusion image using the signal obtained from the simulation. In the next step, those models were used to reconstruct the image of the bladder using a real data frame. Since some of the methods are sensitive to redundancy, various types of regularization and dimensionality reductions were applied. To give an answer to the question of which method gives the best reconstruction, we show the results obtained for random simulated inclusion patterns and collected measurement data. In order not to rely only on visual data, many of the fit measures were estimated: accuracy, sensitivity, specificity, positive predictive value, negative predictive value, detection rate, and AUC. Furthermore, tests of the reliability of the reconstruction (Cohen’s kappa) and the discrepancy between the model and the reconstruction (McNemar’s test) were performed.
Comparing such characteristics as accuracy (the part of the field of view that the model correctly recognized as inclusion) and specificity (the part of the field of view belonging to the background), we see that their values are quite accurate for all of the studied methods. However, the visual analysis of the reconstructions indicates that reconstruction quality drastically differs between the methods. The differences above stand out when comparing positive predicted value, negative predicted value, precision, and F1 measures between models. The results for LR and LDA do not provide sufficient good reconstructions. On the other hand, QDA with dimensionality reduction and decision trees present a very good fit in this context.
To select the best qualitative results, we have to rely on more representative coefficients, i.e., Cohen’s coefficient and McNemar’s test. Comparing the results for κ and χ 2 , one can notice that the best results provide the QDA method and decision trees. Comparing models with regularization and dimensionality reduction with models without redundancy correction clearly shows that the former has a better fit. To determine the final reliability of the models, we compare their predictions gained from the real measurements. We observe a significant advantage in the quality of reconstructions obtained by the QDA method over decision trees.

6. Conclusions

The presented monitoring system is designed for automatic and unsupervised bladder tracking using EIT. Besides tracking, this solution allows for bladder imaging using the inverse problem solution. In this work, we focused on adapting statistical methods for the problem of bladder discretization together with FEM representing the abdomen cross section at the bladder level. We introduced several methods, such as logistic regression (LR), linear and quadratic discriminant methods (LDA, QDA) and decision trees. For LR, LDA and QDA, we implemented several regularization methods, e.g., ridge, LASSO and PCA.
The underlying reason for our research was the implementation of the algorithm for tracking the bladder, disregarding the problem of the impedance distribution in the interior of the studied object. The reliable algorithms for high prediction accuracy provide the ability to track the bladder, as well as the ability to monitor its filling. In addition, those algorithms could work in hybrid mode together with deterministic methods for the inverse problem to obtain the high-resolution impedance distribution in the studied case. In our work, we placed the measuring electrodes on only one side of the patient’s abdomen. This approach opens up new possibilities for creating friendlier and easier-to-use monitoring devices, which are more comfortable for the patient to wear.
In conclusion, we have shown that logistic regression is not sufficient for our task, despite the use of the regularization method. The results obtained by LR are strongly distorted. The reconstruction roughly determines the center of the object but completely loses information about the shape of the inclusion. The better performance was the discrimination result obtained by the LDA method, where the position and shape of the prediction are fairly well-defined. However, the LDA method does not achieve enough high accuracy required for bladder tracking. In our study, the finest and most trustworthy results are given by QDA, characterized by a high Cohen ratio and minor inconsistency χ 2 defined by McNemar’s test.
Thus, the presented study results contain significant information that may accelerate the development of bladder tracking methods in medical tomography. In addition, the research contributes to improving the accuracy of tomographic imaging. The presented algorithm can be used as a hybrid method for predicting an object’s electrical properties.

Author Contributions

Development of the concept of algorithms and implementation of the image reconstruction in EIT, E.K. and D.M.; development of the system concept, measurement methodology, and supervision, M.S. and T.R.; preparation of measurements, development of research methodology, preparation of measurement documentation B.B. and D.W.; literature review, formal analysis, general review, and editing of the manuscript, B.B., D.M. and T.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Maserejian, N.N.; Chen, S.; Chiu, G.R.; Wager, C.G.; Kupelian, V.; Araujo, A.B.; McKinlay, J.B. Incidence of lower urinary tract symptoms in a population-based study of men and women. Urology 2013, 82, 560–564. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Wang, J.Y.; Liao, L.; Liu, M.; Sumarsono, B.; Cong, M. Epidemiology of lower urinary tract symptoms in a cross-sectional, population-based study. Medicine 2018, 97, e11554. [Google Scholar] [CrossRef]
  3. Dumoulin, C.; Hay-Smith, J.; Habée-Séguin, G.M.; Mercier, J. Pelvic floor muscle training versus no treatment, or inactive control treatments, for urinary incontinence in women: A short version Cochrane systematic review with meta-analysis. Neurourol. Urodynamics 2015, 34, 300–308. [Google Scholar] [CrossRef] [Green Version]
  4. Okayama, H.; Ninomiya, S.; Naito, K.; Endo, Y.; Morikawa, S. Effects of wearing supportive underwear versus pelvic floor muscle training or no treatment in women with symptoms of stress urinary incontinence: An assessor-blinded randomized control trial. Int. Urogynecol. J. 2019, 30, 1093–1099. [Google Scholar] [CrossRef] [PubMed]
  5. Schlebusch, T.; Nienke, S.; Leonhardt, S.; Walter, M. Bladder volume estimation from electrical impedance tomography. Physiol. Meas. 2014, 35, 1813. [Google Scholar] [CrossRef] [PubMed]
  6. Leonhäuser, D.; Castelar, C.; Schlebusch, T.; Rohm, M.; Rupp, R.; Leonhardt, S.; Walter, M.; Grosse, J.O. Evaluation of electrical impedance tomography for determination of urinary bladder volume: Comparison with standard ultrasound methods in healthy volunteers. Biomed. Eng. Online 2018, 17, 95. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. He, W.; Ran, P.; Xu, Z.; Li, B.; Li, S.n. A 3D Visualization Method for Bladder Filling Examination Based on EIT. Comput. Math. Methods Med. 2012, 2012, 528096. [Google Scholar] [CrossRef] [Green Version]
  8. Duongthipthewa, O.; Uliss, P.; Pattarasritanawong, P.; Sukaimod, P.; Ouypornkochagorn, T. Analysis of Current Patterns to Determine the Bladder Volume by Electrical Impedance Tomography (EIT). In Proceedings of the 2020 10th International Conference on Biomedical Engineering and Technology, ICBET 2020, Tokyo, Japan, 15–18 September 2020; pp. 122–127. [Google Scholar] [CrossRef]
  9. Ma, G.; Soleimani, M. Spectral Capacitively Coupled Electrical Resistivity Tomography for Breast Cancer Detection. IEEE Access 2020, 8, 50900–50910. [Google Scholar] [CrossRef]
  10. Liu, K.; Bai, X.; Wei, J.; Liang, X.; Sun, J.; Xu, L. Investigation of Bladder Volume Measurement Based on Fringe Effect of Electrical Impedance Tomography Sensors. IEEE Open J. Instrum. Meas. 2022, 1, 1–10. [Google Scholar] [CrossRef]
  11. Konki, S.K.; Khambampati, A.K.; Sharma, S.K.; Kim, K.Y. A deep neural network for estimating the bladder boundary using electrical impedance tomography. Physiol. Meas. 2020, 41, 115003. [Google Scholar] [CrossRef]
  12. Rosa, B.M.G.; Yang, G.Z. Bladder Volume Monitoring Using Electrical Impedance Tomography With Simultaneous Multi-Tone Tissue Stimulation and DFT-Based Impedance Calculation Inside an FPGA. IEEE Trans. Biomed. Circuits Syst. 2020, 14, 775–786. [Google Scholar] [CrossRef] [PubMed]
  13. Wehrens, R. Chemometrics with R; Springer: Berlin/Heidelberg, Germany, 2011; Volume 3. [Google Scholar]
  14. Hastie, T.; Tibshirani, R.; Wainwright, M. Statistical learning with sparsity. Monogr. Stat. Appl. Probab. 2015, 143, 143. [Google Scholar]
  15. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning; Springer Series in Statistics; Springer: New York, NY, USA, 2009. [Google Scholar] [CrossRef]
  16. Ziegler, A. An Introduction to Statistical Learning with Applications. R. G. James, D. Witten, T. Hastie, and R. Tibshirani (2013). Berlin: Springer. 440 pages, ISBN: 978-1-4614-7138-7. Biom. J. 2016, 58, 715–716. [Google Scholar] [CrossRef]
  17. Young, G.A. Mathematical Statistics: An Introduction to Likelihood Based Inference Richard J. Rossi John Wiley & Sons, 2018, xv + 422 pages, £85.99, ebook ISBN: 978-1-118-77104-4, LCCN 2018010628 (ebook). Int. Stat. Rev. 2019, 87, 178–179. [Google Scholar] [CrossRef]
  18. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
  19. Tibshirani, R. Regression Shrinkage and Selection via the Lasso. J. R. Stat. Soc. Ser. B (Methodol.) 1996, 58, 267–288. [Google Scholar] [CrossRef]
  20. Friedman, J.H.; Hastie, T.; Tibshirani, R. Regularization Paths for Generalized Linear Models via Coordinate Descent. J. Stat. Softw. 2010, 33, 1–22. [Google Scholar] [CrossRef] [Green Version]
  21. Guo, Y.; Hastie, T.; Tibshirani, R. Regularized linear discriminant analysis and its application in microarrays. Biostatistics 2006, 8, 86–100. [Google Scholar] [CrossRef] [Green Version]
  22. Pan, Y.; Mai, Q.; Zhang, X. TULIP: A Toolbox for Linear Discriminant Analysis with Penalties. R J. 2021, 12, 61–81. [Google Scholar] [CrossRef]
  23. Friedman, J.H. Regularized Discriminant Analysis. J. Am. Stat. Assoc. 1989, 84, 165–175. [Google Scholar] [CrossRef]
  24. Yan, X.; Su, X.G. Linear Regression Analysis; World Scientific: Singapore, 2009. [Google Scholar] [CrossRef]
  25. Kuhn, M.; Johnson, K. Applied Predictive Modeling; Springer: New York, NY, USA, 2013. [Google Scholar] [CrossRef]
  26. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  27. Hand, D.J.; Till, R.J. A Simple Generalisation of the Area Under the ROC Curve for Multiple Class Classification Problems. Mach. Learn. 2001, 45, 171–186. [Google Scholar] [CrossRef]
  28. Faragallah, O.S.; El-Hoseny, H.; El-Shafai, W.; El-Rahman, W.A.; El-Sayed, H.S.; El-Rabaie, E.S.M.; El-Samie, F.E.A.; Geweid, G.G.N. A Comprehensive Survey Analysis for Present Solutions of Medical Image Fusion and Future Directions. IEEE Access 2021, 9, 11358–11371. [Google Scholar] [CrossRef]
  29. Chervyakov, N.; Lyakhov, P.; Nagornov, N. Analysis of the Quantization Noise in Discrete Wavelet Transform Filters for 3D Medical Imaging. Appl. Sci. 2020, 10, 1223. [Google Scholar] [CrossRef] [Green Version]
  30. Rehman, A.; Zeng, K.; Wang, Z. Display device-adapted video quality-of-experience assessment. In Human Vision and Electronic Imaging XX; Rogowitz, B.E., Pappas, T.N., de Ridder, H., Eds.; International Society for Optics and Photonics, SPIE: Bellingham, WA, USA, 2015; Volume 9394, p. 939406. [Google Scholar] [CrossRef]
  31. Li, X.; Lu, Y.; Wang, J.; Dang, X.; Wang, Q.; Duan, X.; Sun, Y. An image reconstruction framework based on deep neural network for electrical impedance tomography. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3585–3589. [Google Scholar] [CrossRef]
  32. Chen, Z.; Ma, G.; Jiang, Y.; Wang, B.; Soleimani, M. Application of Deep Neural Network to the Reconstruction of Two-Phase Material Imaging by Capacitively Coupled Electrical Resistance Tomography. Electronics 2021, 10, 1058. [Google Scholar] [CrossRef]
  33. Chen, Z.; Xiang, J.; Bagnaninchi, P.O.; Yang, Y. MMV-Net: A Multiple Measurement Vector Network for Multifrequency Electrical Impedance Tomography. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–12. [Google Scholar] [CrossRef]
  34. Liu, Z.; Yang, Y. Multimodal Image Reconstruction of Electrical Impedance Tomography Using Kernel Method. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  35. Liu, Z.; Zhao, R.; Anderson, G.; Bagnaninchi, P.; Yang, Y. Enhanced Multi-Scale Feature Cross-Fusion Network for Impedance-optical Dual-modal Imaging. IEEE Sens. J. 2022. [Google Scholar] [CrossRef]
  36. Ren, S.; Guan, R.; Liang, G.; Dong, F. RCRC: A Deep Neural Network for Dynamic Image Reconstruction of Electrical Impedance Tomography. IEEE Trans. Instrum. Meas. 2021, 70, 1–11. [Google Scholar] [CrossRef]
  37. Yi, Z.; Chen, Z.; Yang, Y. A Learning-Based 3D EIT Image Reconstruction Method. arXiv 2022, arXiv:2208.14449. [Google Scholar] [CrossRef]
  38. Zhang, X.; Wang, Z.; Fu, R.; Wang, D.; Chen, X.; Guo, X.; Wang, H. V-Shaped Dense Denoising Convolutional Neural Network for Electrical Impedance Tomography. IEEE Trans. Instrum. Meas. 2022, 71, 1–14. [Google Scholar] [CrossRef]
Figure 1. Panel (a) shows a visualization of an innovative device for measuring EIT and EMG with a module for muscle electrostimulation. Panel (b) presents a prototype of a wearable EIT measurement system with skin-safe sensors and a comfortable to-carry recording device.
Figure 1. Panel (a) shows a visualization of an innovative device for measuring EIT and EMG with a module for muscle electrostimulation. Panel (b) presents a prototype of a wearable EIT measurement system with skin-safe sensors and a comfortable to-carry recording device.
Sensors 23 01553 g001
Figure 2. The main board with a programmable logic controller (PLC) used in the EIT device.
Figure 2. The main board with a programmable logic controller (PLC) used in the EIT device.
Sensors 23 01553 g002
Figure 3. The block diagram of a solution containing a field-programmable gate array (FPGA).
Figure 3. The block diagram of a solution containing a field-programmable gate array (FPGA).
Sensors 23 01553 g003
Figure 4. An example of a measurement data frame. The data frame consists of 256 components of the recorded EIT voltages and the applied voltage for the injection current (marked with a green line).
Figure 4. An example of a measurement data frame. The data frame consists of 256 components of the recorded EIT voltages and the applied voltage for the injection current (marked with a green line).
Sensors 23 01553 g004
Figure 5. The finite element mesh of the torso cross section at the urinary bladder level.
Figure 5. The finite element mesh of the torso cross section at the urinary bladder level.
Sensors 23 01553 g005
Figure 6. The heatmap of the covariance matrix showing correlations between predictors.
Figure 6. The heatmap of the covariance matrix showing correlations between predictors.
Sensors 23 01553 g006
Figure 7. The top left panel shows the shape of the inclusion. The other panels show the results obtained by logistic regression with regularization methods as shown in the graphs.
Figure 7. The top left panel shows the shape of the inclusion. The other panels show the results obtained by logistic regression with regularization methods as shown in the graphs.
Sensors 23 01553 g007
Figure 8. ROC analysis for results shown in Figure 7.
Figure 8. ROC analysis for results shown in Figure 7.
Sensors 23 01553 g008
Figure 9. The top left panel shows the shape of the inclusion. The other panels show the results obtained by the discriminant method with/without dimensionality reduction.
Figure 9. The top left panel shows the shape of the inclusion. The other panels show the results obtained by the discriminant method with/without dimensionality reduction.
Sensors 23 01553 g009
Figure 10. Principal component analysis explained variance plots. The (left) panel shows the percentage of explained variance by particular components. The (right) panel presents the cumulative variance explained by the first k components.
Figure 10. Principal component analysis explained variance plots. The (left) panel shows the percentage of explained variance by particular components. The (right) panel presents the cumulative variance explained by the first k components.
Sensors 23 01553 g010
Figure 11. ROC analysis for results shown in Figure 9.
Figure 11. ROC analysis for results shown in Figure 9.
Sensors 23 01553 g011
Figure 12. The first panel presents the inclusion shape. The second panel shows reconstruction obtained by CART.
Figure 12. The first panel presents the inclusion shape. The second panel shows reconstruction obtained by CART.
Sensors 23 01553 g012
Figure 13. ROC analysis for results shown in Figure 12.
Figure 13. ROC analysis for results shown in Figure 12.
Sensors 23 01553 g013
Figure 14. Comparison of results obtained for real measurement data using introduced discretization methods. As a reference, we show the USG image for the bladder superimposed on the finite element mesh with the proportions maintained.
Figure 14. Comparison of results obtained for real measurement data using introduced discretization methods. As a reference, we show the USG image for the bladder superimposed on the finite element mesh with the proportions maintained.
Sensors 23 01553 g014
Table 1. Confusion matrix.
Table 1. Confusion matrix.
PositiveNegative
Positive PredictionTPFP
Negative PredictionFNTN
Table 2. Table of fit measures obtained for the reconstruction shown in Figure 7.
Table 2. Table of fit measures obtained for the reconstruction shown in Figure 7.
Elastic NetRidgeLASSO
Accuracy0.9000.8890.857
Sensitivity1.0001.0001.000
Specificity0.8980.8870.854
Pos Pred Value0.1700.1570.126
Neg Pred Value1.0001.0001.000
Precision0.1700.1570.126
F10.2910.2710.224
Prevalence0.0210.0210.021
Detection Rate0.0210.0210.021
Detection Prevalence0.1210.1310.163
Balanced Accuracy0.9490.9440.927
AUC0.9960.9970.985
κ 0.2650.2440.194
χ 2 154.006170.006220.005
Table 3. Table of fit measures obtained for the reconstruction shown in Figure 9.
Table 3. Table of fit measures obtained for the reconstruction shown in Figure 9.
LDAQDARDALDAPCAQDAPCA
Accuracy0.9430.9950.9470.9470.995
Sensitivity1.0000.9291.0001.0001.000
Specificity0.9420.9960.9460.9460.995
Pos Pred Value0.2410.8130.2520.2520.778
Neg Pred Value1.0000.9991.0001.0001.000
Precision0.2410.8120.2520.2520.778
F10.3890.8670.4030.4030.875
Prevalence0.0180.0180.0180.0180.018
Detection Rate0.0180.0170.0180.0180.018
Detection Prevalence0.0750.0210.0710.0710.023
Balanced Accuracy0.9710.9620.9730.9730.997
AUC0.9860.9980.9850.9881.000
κ 0.3710.8640.3850.3850.872
χ 2 86.0111.12581.01281.0126.125
Table 4. Table of coefficients obtained for the reconstruction shown in Figure 12.
Table 4. Table of coefficients obtained for the reconstruction shown in Figure 12.
CART
Accuracy0.999
Sensitivity0.947
Specificity1.000
Pos Pred Value1.000
Neg Pred Value0.999
Precision1.000
F10.973
Prevalence0.023
Detection Rate0.022
Detection Prevalence0.022
Balanced Accuracy0.973
AUC0.982
κ 0.972
χ 2 2
Table 5. Table of measures that determine the quality of the reconstructions.
Table 5. Table of measures that determine the quality of the reconstructions.
ModelMAEMSESSIMPSNR
Elastic0.110.051.0012.53
Ridge0.150.051.0010.89
LASSO0.140.101.009.99
LDA0.060.051.0013.22
QDA0.010.011.0022.93
RDA0.060.051.0013.40
LDA PCA0.070.031.0014.84
QDA PCA0.010.011.0023.00
CART0.000.001.0028.92
DNN [31]-0.0083--
CNN [32]-0.01400.901118.5387
MMV-Net [33]-0.0490.935423.7423
Kernel method [34]--0.7822-
En-MSFCF-Net [35]--0.9862-
RCRC [36]-33.1770.68-
TN-Net [37]-0.00580.965730.709
VDD-Net [38]-0.941--
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Baran, B.; Kozłowski, E.; Majerek, D.; Rymarczyk, T.; Soleimani, M.; Wójcik, D. Application of Machine Learning Algorithms to the Discretization Problem in Wearable Electrical Tomography Imaging for Bladder Tracking. Sensors 2023, 23, 1553. https://doi.org/10.3390/s23031553

AMA Style

Baran B, Kozłowski E, Majerek D, Rymarczyk T, Soleimani M, Wójcik D. Application of Machine Learning Algorithms to the Discretization Problem in Wearable Electrical Tomography Imaging for Bladder Tracking. Sensors. 2023; 23(3):1553. https://doi.org/10.3390/s23031553

Chicago/Turabian Style

Baran, Bartłomiej, Edward Kozłowski, Dariusz Majerek, Tomasz Rymarczyk, Manuchehr Soleimani, and Dariusz Wójcik. 2023. "Application of Machine Learning Algorithms to the Discretization Problem in Wearable Electrical Tomography Imaging for Bladder Tracking" Sensors 23, no. 3: 1553. https://doi.org/10.3390/s23031553

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop