Next Article in Journal
Histopathological Analysis for Detecting Lung and Colon Cancer Malignancies Using Hybrid Systems with Fused Features
Next Article in Special Issue
AI Model for Detection of Abdominal Hemorrhage Lesions in Abdominal CT Images
Previous Article in Journal
Effects of a Graphene Heating Device on Fatigue Recovery of Biceps Brachii
Previous Article in Special Issue
Optimization System Based on Convolutional Neural Network and Internet of Medical Things for Early Diagnosis of Lung Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Diffuse Optical Tomography Using Extreme Gradient Boosting and Genetic Programming

by
Ami Hauptman
1,†,
Ganesh M. Balasubramaniam
2,*,† and
Shlomi Arnon
2
1
Department of Computer Science, Sapir Academic College, Sderot 7915600, Israel
2
Department of Electrical and Computer Engineering, Ben-Gurion University of the Negev, Be’er Sheva 8441405, Israel
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Bioengineering 2023, 10(3), 382; https://doi.org/10.3390/bioengineering10030382
Submission received: 20 February 2023 / Revised: 18 March 2023 / Accepted: 20 March 2023 / Published: 21 March 2023
(This article belongs to the Special Issue Artificial Intelligence in Advanced Medical Imaging)

Abstract

:
Diffuse optical tomography (DOT) is a non-invasive method for detecting breast cancer; however, it struggles to produce high-quality images due to the complexity of scattered light and the limitations of traditional image reconstruction algorithms. These algorithms can be affected by boundary conditions and have a low imaging accuracy, a shallow imaging depth, a long computation time, and a high signal-to-noise ratio. However, machine learning can potentially improve the performance of DOT by being better equipped to solve inverse problems, perform regression, classify medical images, and reconstruct biomedical images. In this study, we utilized a machine learning model called “XGBoost” to detect tumors in inhomogeneous breasts and applied a post-processing technique based on genetic programming to improve accuracy. The proposed algorithm was tested using simulated DOT measurements from complex inhomogeneous breasts and evaluated using the cosine similarity metrics and root mean square error loss. The results showed that the use of XGBoost and genetic programming in DOT could lead to more accurate and non-invasive detection of tumors in inhomogeneous breasts compared to traditional methods, with the reconstructed breasts having an average cosine similarity of more than 0.97 ± 0.07 and average root mean square error of around 0.1270 ± 0.0031 compared to the ground truth.

1. Introduction

Imaging with randomly scattered light is a significant challenge with a pressing need in non-invasive biomedical diagnosis [1,2,3,4,5,6]. One of the more significant applications is a non-invasive functional imaging technique called diffuse optical tomography (DOT), which uses near-infrared (NIR) light to map in 3D the optical characteristics of tissue by penetrating it deeply [7,8,9,10,11,12,13]. Soft tissue, including the breast and the brain, can be penetrated several centimeters by diffuse light in the NIR wavelength range. Measurements from reflected or transmitted light at the tissue surface are used to reconstruct DOT images using inverse problems [14,15,16]. Moreover, DOT has many advantages, including the non-ionizing nature of the light spectra used for imaging tissues and the non-invasive nature of the technology. As a result, DOT has found critical applications, especially in breast tissue imaging [8,15,17] and tissue property estimation [10,18,19,20,21]. However, many challenges persist. Although near-infrared (NIR) photons can travel several centimeters into the tissue to enable non-invasive biomedical imaging, these photons scatter widely and migrate in random directions before escaping or being absorbed by the medium, making imaging and detection tasks challenging [20,22].
Moreover, DOT reconstruction is an ill-defined and inadequately framed topic that calls for an inversion regularization to enhance convergence [15,23]. Inaccurate DOT reconstruction can also be caused by model inadequacies, including poor boundary conditions, the chest wall beneath the breast tissue, and inaccurate background tissue estimation [8,24]. Due to the above reasons, DOT for breast cancer imaging has previously been investigated but found to be facing various challenges when considering large-scale medical deployment. Since then, significant technological advancements have been made, including improved computational efficiency that stimulated the creation of novel deep learning algorithms and a deeper comprehension of how light travels through tissues. Therefore, a less complicated yet computationally effective approach is needed to detect anomalies and tumors present in thick tissues.
Recent research has attempted to address some of the challenges posed due to the diffuse nature of photons and ill-posed inverse problems by employing deep learning (DL) [15,17,20,25,26] and shallow machine learning (ML) techniques [8,27]. The usage of ML algorithms regularly beats the analytical solution technique in all of the articles previously mentioned. Compared to traditional analytical models, ML and DL algorithms are better posed, faster, and provide a better convergence while determining tissues’ optical properties or detecting locations of tumors and anomalies in thick tissues. For instance, Murad et al. [28] experimentally demonstrated the simultaneous reconstruction of the absorption and scattering coefficient of tissue mimicking using a 1D convolution neural network (1D-CNN). Here, simple batch normalization (BN) layers were used to significantly improve the accuracy and reduce the time consumed for DOT image reconstruction. Yoo et al. [15] developed a novel DL algorithm that accurately detects anomalies in beast tissues by inverting the Lippman–Schwinger equation, achieving significant results. In a more recent study, Mozumder et al. [29] employed a model-based DL approach to improve the estimation of the absorption and scattering coefficient of diffuse media. It was shown in this study that the proposed DL method also significantly reduces computation time. Moreover, several excellent reviews [30] and tutorials [14] exist that provide valuable insights regarding the use of DL for DOT image reconstruction.
However, despite the improvements in accuracy obtained in DOT reconstruction by DL methods, there is still a pressing need for algorithms that provide similarly good results with smaller datasets and reduce the computational load on the system. ML algorithms may provide solutions to some of these issues as they work well on smaller datasets, can function without a GPU, and are generally computationally quicker than DL methods. Recently, Zou et al. [8] used a machine learning model with physical constraints to reconstruct DOT images. Yun Zou et al. proposed that the ML algorithm significantly reduced the DOT image reconstruction error, especially in high-contrast samples, compared to the Born gradient descent analytic reconstruction method. The significant observations and the results obtained from this work clearly show the potential of machine learning algorithms where DOT reconstruction problems are concerned.
Extreme gradient boosting or XGBoost is a widely used machine learning algorithm known for its high performance and accuracy in various applications such as regression, classification, and ranking problems [31]. XGBoost shows high scalability, robustness, and speed compared to other models [32,33]. Therefore, in this article, for the first time, to the best of our knowledge, we use XGBoost to detect tumors located in compressed breast tissues. We later apply an optimization method called genetic programming (GP) [34,35,36] to improve our results further. GP is an evolutionary algorithm that uses natural selection to evolve solutions to optimization problems. It can automate the design and optimization of complex systems, generate novel solutions, find robust and interpretable solutions, and be more efficient than other optimization algorithms. The proposed machine learning diffuse optical tomography (ML-DOT) method requires short training times and delivers high accuracy in detecting tumors even with small datasets, and thus significantly reduces the computational load.
The article is designed as follows: The imaging geometry used in this study and the data generation are described in Section 2.1, and the proposed technique for tumor detection in compressed breasts is described in Section 2.2 and Section 2.3, along with the obtained results. Section 3 discusses the results of the proposed algorithm, and we finish with the conclusions derived from this study in Section 4.

2. Materials and Methods

Image reconstruction in DOT has traditionally been achieved using inverse problems, which take the form [37]:
F i n d x X   f r o m   d a t a   y = A ( x ) + δ ,   y Y
where the parameter space is defined by X , Y is the measurement space, A is the propagation model of photons convolved with the optical component response, and δ is the noise mechanism in the system. The scarcity of ballistic photons coupled with the loss of imaging data as a result of repeated scattering events, however, results in a non-linear, ill-posed inverse problem, needing an effective solution to accomplish this challenging task. In order to simplify the DOT system and precisely locate cancers embedded in the compressed breast tissue, approaches based on extreme gradient boosting and genetic programming are proposed in this section. This section includes a brief introduction to the proposed machine learning techniques, the proposed methodology, and the results obtained from this study.

2.1. Simulating Photon Migration in Digital Breast Phantoms to Generate Dataset

The compressed breast geometry is chosen for this study. This particular type of geometry is used as it considerably reduces movement artifacts, evenly spreads the different layers of tissue, and also helps considerably reduce the amount of tissue to be imaged [16,38]. The dataset used to train, validate, and test the algorithm is created using the finite element method (FEM) [39]. The evaluation of DOT models is usually conducted through computer simulations, as it allows for easy comparison to the work of other scientists, creating a larger dataset and providing a better-controlled environment. Additionally, simulations allow for the avoidance of costly and wasteful fabrication of clinical prototype systems that may possess inherent engineering problems. Furthermore, it is both complicated and expensive to create an inhomogeneous breast phantom for experimental purposes [14,38]. The compressed breast mesh itself is loaded from the “DigiBreast” [40] digital breast phantom modeled accurately using the optical properties of the skin, breast tissues, and the chest wall [17,41]. The digital breast phantom includes a realistic 3D glandularity map, which is more detailed than the representation of breast tissue in conventional numerical breast phantoms. These traditional phantoms use piecewise-constant regions to represent the fibroglandular and adipose tissue, which can result in a loss of information. In contrast, the digital breast phantom uses statistical or fuzzy segmentation methods to create spatially varying tissue volume fraction maps, which preserve more detailed information about the breast tissue.
The dimensions of the compressed breast are 220.8 mm, 102.9 mm, and 23.7 mm in the X, Y, and Z directions, respectively. A parallel-plate-based measurement scheme is employed to set a 48 × 54 source–detector arrangement on either side of the breast. The simulated compressed breast mesh and the source–detector arrangement are shown in Figure 1.
The simulations are then performed in the transmission geometry using the Toast++ [39] forward solver to simulate light intensity measurements at the detectors. As tumors often have a high blood concentration, in this study, they are modeled as oxyhemoglobin spheres with radii ranging from 2 mm to 15 mm randomly located inside a compressed breast mesh. The algorithm used 5000 simulated breast measurements taken in the transmission geometry as input, and the x, y, and z coordinates, the absorption coefficient matrix, and the tumor’s radius were the labels. These simulated measurements were performed in the continuous wave regime at a wavelength of 833 nm, and a 2% gaussian noise was added to mimic errors that might occur if the measurements were taken from an experimental procedure.

2.2. Detecting Tumors in the Compressed Breast Using Extreme Gradient Boosting

Following the creation of the dataset, the collected data are used to solve the inverse problem and detect the location of tumors inside the compressed breast. A machine learning algorithm called extreme gradient boosting (XGBoost) is employed to solve the inverse problem. XGBoost is a distributed, scalable gradient-boosted decision tree (GBDT) machine learning framework with several applications that use regression [32] and classification [33]. Here, gradient boosting refers to “boosting” or strengthening one weak model by fusing it with several other weak models to create a more robust model. As an extension of boosting, gradient boosting formalizes the process of additively creating weak models as a gradient descent method over a fitness function. Gradient boosting algorithms create a model that predicts the label by evaluating a tree of logical feature questions and determining the minimum number of questions required to assess the probability of obtaining a correct decision. To reduce errors, gradient boosting establishes desired outcomes for the upcoming model. The gradient of the error concerning the prediction determines the targeted outcomes for each case, hence the name “gradient boosting”.
Created by Tianqi Chen and contributions from numerous developers, “Extreme Gradient Boosting (XGBoost)” is one of the most important and successful implementations of gradient boosting machines [31]. XGBoost is a scalable and highly accurate gradient-boosting algorithm that pushes the limits of computing power for boosted tree algorithms. It was created primarily to enhance machine learning models’ performance and computational speed. Unlike GBDTs, where decision trees are generated sequentially, the trees in XGBoost are built in parallel. The XGBoost method takes a level-wise approach, scanning over multiple gradient values and using the partial sum generated from the gradient values to evaluate the performance of the splits for each possible split in the training dataset.
As described earlier, the dataset is created using 5000 different measurements at the transmission boundaries of the compressed breast. The log of the measured signal at each detector with respect to different sources (log(d)) is used as the input to the XGBoost algorithm. The “x”, “y”, and “z” coordinates, the absorption coefficient, and the radius are the labels. The algorithm is employed to find an accurate model that establishes a precise relation between the simulated measurements and the tumor locations. The root mean squared error (RMSE) loss function is used to assess the algorithm’s performance, and the accuracy of the predicted outcome is measured using the RMSE and the cosine similarity (CS) metrics [42]. The RMSE loss function is given by:
RMSE = i = 1 N ( Y i X i ) 2 N  
and the cosine similarity is given by:
CS = i = 1 N X i Y i i = 1 N X i 2 i = 1 N Y i 2  
where Y i and X i are the predictions and the labels, respectively, and N is the total number of datasets. These metrics are specifically used primarily because the R2 metrics perform relatively poorly when the prediction window is narrow, as is the case with the radius of the spheres. Furthermore, the RMSE and the CS metrics are better suited to solving inverse problems where the prediction interval is narrow and continuous [8,43]. A computer with an i9 series 9900k processor and two NVIDIA GeForce RTX 2080Ti graphics processors were used to build and train the XGBoost algorithm. Each GPU contains 11 GB of VRAM, and the two GPUs are connected through an NVlink. The network is by Adam Optimizer with an initial learning rate of 0.0001. For the XGB predictions, 50% of the total dataset is used, out of which 60% is used for training and validation, and 40% of the data is used for testing. The XGBoost algorithm learns the features from the input and the labels and accurately predicts the relationship between the input (measurement data) and the labels (tumor location) by regression. Due to variations in the dimensions of the compressed breast and the nature of the coordinate system used to model the compressed breast, the model prediction is made separately for the three different coordinates, the absorption coefficient, and the radius. The best results obtained by the XGBoost algorithm are RMSE values of 0.1862 ± 0.0018 for the x coordinate, 0.1678 ± 0.0042 for the y coordinate, 0.1505 ± 0.0009 for the z coordinate, 0.1131 ± 0.0091 for the absorption coefficient, and 0.2157 ± 0.0103 for the radius. Some of the predictions of the XGBoost algorithm are shown below in Figure 2.
From the results obtained by the XGBoost algorithm, it is undoubtedly well-posed to detect tumors to a certain degree of accuracy using the simulated measurements. The predictions by the XGBoost algorithm obtained a very high average cosine similarity value of 0.9564 ± 0.0076 compared to the ground truth. However, the predicted size and locations of the embedded tumors could still be improved, especially the predictions of the radius and the z coordinate, which is also the mean direction of propagation. Therefore, the following section applies a post-processing algorithm using genetic programming (GP) to improve the model predictions.

2.3. Enhancing Tumor Detection Capabilities Using Genetic Programming

Genetic programming (GP) is a nature-inspired hyper-heuristic search algorithm that optimizes a set (or a population) of solutions (or individuals), embodied as computer programs, which are typically represented as tree-like structures, using a predefined loss function (or a fitness function) for a given task [35,44]. In other words, GP starts with a high-level declaration of “what needs to be done” and constructs a computer program to solve the problem automatically. As described in Figure 3, GP begins with a primordial soup of a large pool of randomly generated computer programs (initial guess of the model prediction), and this population of programs evolves throughout generations. The evolutionary search employs the Darwinian concept of natural selection (survival of the fittest) and analogs of many naturally occurring events, such as crossover (sexual recombination), mutation, gene duplication, and gene deletion. Moreover, genetic algorithms, by means of fitness-based selections and applying genetic operators over time, can optimize solutions automatically during the course of simulated evolution. Thus, GPs’ adaptive nature, free of human prejudices or biases, may often provide answers better than the most exemplary human efforts.
Genetic or evolutionary algorithms have a wide range of applications, particularly in domains where an exact structure of the solution is unknown in advance or when finding an approximate solution is considered suitable [45]. Genetic programming is frequently used in conjunction with other types of machine learning because it is useful for performing symbolic regressions and feature classifications [46].
The execution of a GP algorithm is described below:
  • Randomly generate an initial population of solutions called individuals. Each individual is generated as a random tree of limited depth, consisting of nodes taken from the terminal set and the function set. The terminal set contains constants and variables, and the function set consists of various operators, for example, mathematical operations, logical operators, etc.
  • While the termination criterion is not fulfilled, the following sub-steps are repeated:
    a.
    Evaluate the individuals in the current population according to the fitness function, which outputs a numerical value representing the quality of the individual as a solution.
    b.
    Select individuals from the population using a selection method, where the probability for selection is related to fitness values, for producing the next set of individuals.
    c.
    Apply the following genetic operators to produce new individuals with predetermined probabilities:
    I.
    Reproduction: clone an individual selected by the sub-step ”b” to the population.
    II.
    Crossover: randomly recombine two selected individuals to produce two new offspring.
    III.
    Mutation: randomly alter one selected individual to produce one new offspring.
  • Output the best individual found during the run as the output.
To implement the steps described above, we employed Koza-style [47] genetic programming using the DEAP Python environment [48]. We now describe the details of our implementation:
Essentially, each GP individual represents an function, which receives a single real-valued coordinate as input (namely, x ,   y ,   z , or r a d i u s ) outputted by XGBoost in the previous stage), and outputs a single real-value, as an estimate for the actual coordinate ( x ,   y , z, or r a d u s ). The function set for each individual (as can be seen in the pseudo-code above) is comprised of binary arithmetic operations. The terminal set is comprised of a single real-valued coordinate and integer ephemeral random constants (ERCs) [35,44,47] in the range [ 5 ,   5 ] . For example, if the coordinate in question is x , it is possible to form an expression such as add (multiply (x, x), negate (multiply (x, 5.0))) representing the function: x 2 + ( 5 x ) . Typically, GP trees can represent expressions that are far more complex.
For assessing our individuals, fitness was calculated by randomly selecting 10% of the dataset (with the given single label) and applying the individual to each of the selected examples. The fitness score was determined by the RMSE measure between the actual values for the labels in the dataset and those predicted by the individual. We used tournament selection as the selection mechanism. Additional parameters of the GP algorithm are given in Table 1, and the entire workflow of the proposed method, including the XGBoost and GP, is shown in Figure 3.
The remaining 50% of the total data after the XGBoost is used for the GP algorithm. The data split for the GP algorithm’s training, validation, and testing is similar to the data split used for the XGBoost, and the algorithm is also run on the same computational hardware. After applying the algorithm for 30 runs with 100 generations each, the final results obtained after XGBoost and GP are RMSE values of 0.1808 ± 0.0014 for the x coordinate, 0.1539 ± 0.0057 for the y coordinate, 0.1340 ± 0.0032 for the z coordinate, 0.0975 ± 0.0065 for the absorption coefficient, and 0.2017 ± 0.0126 for the radius. The reconstructed images from the predictions are shown in Figure 4.
As seen in Figure 4, predictions significantly improve when the post-processing GP algorithm is employed. The most significant enhancement is observed in the predictions of the z coordinate of the embedded tumors. Considering that the z-axis is the mean direction of the photon propagation due to the forward scattering anisotropy of the media, this is a commendable result, as minimal errors in the prediction of the z coordinate will allow us to detect anomalies and tumors at greater depths. The results obtained using the XGBoost and the enhancement due to GP are summarized in Table 2.

3. Discussion

As seen in Figure 4, the predicted outcomes are very close to the ground truth locations and the radii. When the CS metric is used to validate the model’s performance, an average similarity of 0.9720 ± 0.0062 is observed when the predicted labels are compared to the ground truth. From the RMSE and CS values, it is apparent that the prediction is very robust. The algorithm is also computationally fast, as training consumes less than 60 s. It must also be noted that the algorithms do not have any prior information or knowledge regarding the inverse solution to DOT. Our study found that our predictions for the x and y coordinates were significantly more accurate than those for the z coordinate and radius. This discrepancy may be due to the fact that the range of the z coordinate and radius was limited compared to the range of the x and y coordinates. This limited range not only adversely affected the algorithm’s performance but also negatively impacted the simulated measurements. The small changes within this limited range were not as clearly reflected in the measurement matrix, resulting in a less accurate regression model for predicting the tumor location and radius.
Moreover, our study’s absorption coefficient reconstruction is still not very accurate. The XGBoost algorithm underestimates the absorption coefficient, while the GP regularly overestimates the absorption coefficient. This is likely due to the inhomogeneous distribution of tissue within the breast and the varying contrasts between the anomalies (such as tumors) and the surrounding tissue. Inhomogeneity in the tissue distribution can introduce variations in the optical properties of the breast, leading to challenges in accurately reconstructing the absorption coefficient. Similarly, the variability in contrast between the anomalies and the background can make it difficult to accurately distinguish between the two, leading to errors in the reconstruction. These factors likely contributed to the limitations in the accuracy of the absorption coefficient reconstruction in our study.
Despite these limitations, it is worth noting that the algorithm was still able to learn an accurate model for solving the inverse problem and identifying tumors based on a relatively small, simulated dataset. This demonstrates the robustness and effectiveness of the algorithm, which effectively processed and analyzed the data despite the challenges presented by the limited range of the z coordinate and radius. Overall, our results show that the combination of XGBoost and genetic programming (GP) can be a powerful tool for solving inverse problems in diffuse optical tomography and may have potential applications in medical imaging and other fields.
XG boost and genetic programming (GP) are highly effective algorithms for solving diffuse optical tomography (DOT) problems. XG boost, a decision tree-based algorithm, is particularly well-suited for handling highly non-linear data and can be used for classification and regression tasks. GP is versatile and can be used for a variety of purposes, including regression, classification, and optimization. These algorithms have been shown to outperform other approaches, such as conventional analytical inverse problems and convolutional neural networks (CNNs), in many complex applications due to their ability to accurately process and analyze the data. Therefore, XGBoost and GP are the preferred choice for solving DOT problems.
Moreover, the primary issue with employing neural networks to solve DOT inverse problems is the near impossibility of collecting massive amounts of experimental data to train the algorithm, leading to trade-offs between precision and accuracy in actual investigations. This study addresses this issue as shallow machine learning algorithms such as XGBoost do not require substantial amounts of data for accurate model prediction. The comparison of the effectiveness of our algorithm with that of the existing state-of-the-art algorithms is shown in Table 3. A more detailed comparison to other imaging modalities and algorithms using different metrics can be found in excellent review articles [14,30,49].
As seen in Table 3, the proposed method is well-suited and outperforms other algorithms that solve a similar problem. Furthermore, transfer learning can close the gap between simulations and real tissue imaging. Additionally, to improve performance, the proposed methods can be trained using a more complex and realistic dataset with different geometries and source–detector configurations, paving the way for the realization of large-scale medical applications involving non-invasive and radiation-free diffuse optical techniques that can provide an accurate diagnosis.

4. Conclusions

In this study, an extreme gradient boosting algorithm was used in conjunction with a genetic programming algorithm to detect embedded tumors or anomalies in inhomogeneous compressed breast tissues. A dataset of 5000 compressed breasts with anomalies was simulated for the study, and simulated light measurements were used to determine the location and size of the tumors. After applying the proposed method, the obtained RMSE and cosine similarity values showed that the algorithm was highly accurate and robust when the predicted tumor locations and sizes were compared to the ground truth. Moreover, the issues regarding large datasets and expensive computational costs were addressed to a certain extent as only a few breast samples were used, and the tumor location and radius predictions consume a very minimal amount of time. The results from the proposed method to accurately determine the anomalies in breast tissues show that extreme gradient boosting and genetic programming algorithms can be employed to detect tumors in compressed breast tissues with excellent accuracy compared to previously existing methods.

Author Contributions

A.H.: conceptualization, visualization, validation, methodology, and writing (initial draft and review). G.M.B.: conceptualization, visualization, investigation, validation, light migration simulations, dataset creation, writing (initial draft, review, and editing), and figures. S.A.: project administration, supervision, conceptualization, investigation, validation, and writing (review). All authors have read and agreed to the published version of the manuscript.

Funding

The authors thank the Kreitman School of Advanced Graduate Studies and the Ben-Gurion University of the Negev for funding and fellowships to continue the research.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The simulation and machine learning codes presented in this paper are not publicly available but may be obtained from the authors upon reasonable request.

Acknowledgments

The authors thank S.T. for their generous donation towards breast cancer research at BGU.

Conflicts of Interest

The authors declare no conflict of interests.

References

  1. Shrestha, S.; Deshpande, A.; Farrahi, T.; Cambria, T.; Quang, T.; Majeski, J.; Na, Y.; Zervakis, M.; Livanos, G.; Giakos, G.C. Label-free discrimination of lung cancer cells through mueller matrix decomposition of diffuse reflectance imaging. Biomed. Signal Process. Control 2018, 40, 505–518. [Google Scholar] [CrossRef]
  2. Jeeva, J.B.; Singh, M. Reconstruction of optical scanned images of inhomogeneities in biological tissues by Monte Carlo simulation. Comput. Biol. Med. 2015, 60, 92–99. [Google Scholar] [CrossRef] [PubMed]
  3. Perkins, G.A.; Eggebrecht, A.T.; Dehghani, H. Multi-modulated frequency domain high density diffuse optical tomography. Biomed. Opt. Express 2022, 13, 5275–5294. [Google Scholar] [CrossRef] [PubMed]
  4. Marcos-Vidal, A.; Ripoll, J. Recent advances in optical tomography in low scattering media. Opt. Lasers Eng. 2020, 135, 10619. [Google Scholar] [CrossRef]
  5. Yoon, S.; Kim, M.; Jang, M.; Choi, Y.; Choi, W.; Kang, S.; Choi, W. Deep optical imaging within complex scattering media. Nat. Rev. Phys. 2020, 2, 141–158. [Google Scholar] [CrossRef] [Green Version]
  6. Bertolotti, J.; Katz, O. Imaging in complex media. Nat. Phys. 2022, 18, 1008–1017. [Google Scholar] [CrossRef]
  7. Ferrari, M.; Culver, J.P.; Hoshi, Y.; Wabnitz, H. Special Section Guest Editorial: Clinical near-infrared spectroscopy and imaging. J. Biomed. Opt. 2016, 21, 091301. [Google Scholar] [CrossRef] [Green Version]
  8. Zou, Y.; Zeng, Y.; Li, S.; Zhu, Q. Machine learning model with physical constraints for diffuse optical tomography. Biomed. Opt. Express 2021, 12, 5720–5735. [Google Scholar] [CrossRef]
  9. Dileep, B.P.V.; Dutta, P.K.; Prasad, P.M.K.; Santhosh, M. Sparse recovery based compressive sensing algorithms for diffuse optical tomography. Opt. Laser Technol. 2020, 128, 106234. [Google Scholar] [CrossRef]
  10. Zhang, M.; Li, S.; Zou, Y.; Zhu, Q. Deep learning-based method to accurately estimate breast tissue optical properties in the presence of the chest wall. J. Biomed. Opt. 2021, 26, 106004. [Google Scholar] [CrossRef]
  11. Paul, R.; Murali, K.; Varma, H.M. High-density diffuse correlation tomography with enhanced depth localization and minimal surface artefacts. Biomed. Opt. Express 2022, 13, 6081–6099. [Google Scholar] [CrossRef] [PubMed]
  12. Bowden, A.K.; Durr, N.J.; Erickson, D.; Ozcan, A.; Ramanujam, N.; Jacques, P.V. Optical Technologies for Improving Healthcare in Low-Resource Settings: Introduction to the feature issue. Biomed. Opt. Express 2020, 11, 3091–3094. [Google Scholar] [CrossRef] [PubMed]
  13. Li, C.; Cheung, M.R. The utility of a marched absorbing layer boundary condition in the finite element analysis of diffuse photon density wave propagation in tissues relevant to breast imaging. Comput. Biol. Med. 2009, 39, 934–939. [Google Scholar] [CrossRef] [PubMed]
  14. Balasubramaniam, G.M.; Wiesel, B.; Biton, N.; Kumar, R.; Kupferman, J.; Arnon, S. Tutorial on the Use of Deep Learning in Diffuse Optical Tomography. Electronics 2022, 11, 305. [Google Scholar] [CrossRef]
  15. Yoo, J.; Sabir, S.; Heo, D.; Kim, K.H.; Wahab, A.; Choi, Y.; Lee, S.I.; Chae, E.Y.; Kim, H.H.; Bae, Y.M.; et al. Deep Learning Diffuse Optical Tomography. IEEE Trans. Med. Imaging 2020, 39, 877–887. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Balasubramaniam, G.M.; Arnon, S. Deep-learning algorithm to detect anomalies in compressed breast: A numerical study. In Proceedings of the Bio-Opitcs: Design and Application, Washington, DC, USA, 12–16 April 2021. [Google Scholar]
  17. Balasubramaniam, G.M.; Arnon, S. Regression-based neural network for improving image reconstruction in diffuse optical tomography. Biomed. Opt. Express 2022, 13, 2006–2017. [Google Scholar] [CrossRef] [PubMed]
  18. Nguyen, M.H.; Zhang, Y.; Wang, F.; De La Garza Evia Linan, J.; Markey, M.K.; Tunnell, J.W. Machine learning to extract physiological parameters from multispectral diffuse reflectance spectroscopy. J. Biomed. Opt. 2021, 26, 052912. [Google Scholar] [CrossRef]
  19. Fredriksson, I.; Larsson, M.; Strömberg, T. Machine learning for direct oxygen saturation and hemoglobin concentration assessment using diffuse reflectance spectroscopy. J. Biomed. Opt. 2020, 25, 112905. [Google Scholar] [CrossRef]
  20. Sabir, S.; Cho, S.; Kim, Y.; Pua, R.; Heo, D.; Kim, K.H.; Choi, Y.; Cho, S. Convolutional neural network-based approach to estimate bulk optical properties in diffuse optical tomography. Appl. Opt. 2020, 59, 1461–1470. [Google Scholar] [CrossRef]
  21. Hokr, B.H.; Bixler, J.N. Machine learning estimation of tissue optical properties. Sci. Rep. 2021, 11, 6561. [Google Scholar] [CrossRef]
  22. He, J.; Li, C.L.; Wilson, B.C.; Fisher, C.J.; Ghai, S.; Weersink, R.A. A Clinical Prototype Transrectal Diffuse Optical Tomography (TRDOT) System for in vivo Monitoring of Photothermal Therapy (PTT) of Focal Prostate Cancer. IEEE Trans. Biomed. Eng. 2020, 67, 2119–2129. [Google Scholar] [CrossRef] [PubMed]
  23. Arridge, S.R.; Hebden, J.C. Optical imaging in medicine: II. Modelling and reconstruction. Phys. Med. Biol. 1997, 42, 841–853. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Uddin, K.M.S.; Zhang, M.; Anastasio, M.; Zhu, Q. Optimal breast cancer diagnostic strategy using combined ultrasound and diffuse optical tomography. Biomed. Opt. Express 2020, 11, 2722–2737. [Google Scholar] [CrossRef] [PubMed]
  25. Zhao, Y.; Deng, Y.; Bao, F.; Peterson, H.; Istfan, R.; Roblyer, D. Deep learning model for ultrafast multifrequency optical property extractions for spatial frequency domain imaging. Opt. Lett. 2018, 43, 5669–5672. [Google Scholar] [CrossRef]
  26. Kazanci, H.O.; Oral, O. Improving image quality in diffuse optical tomography. Opt. Quantum Electron. 2022, 54, 655. [Google Scholar] [CrossRef]
  27. Fogarty, M.; Tripathy, K.; Svoboda, A.M.; Schroeder, M.L.; Rafferty, S.; Mansfield, P.; Ulbrich, R.; Booth, M.; Richter, E.J.; Smyser, C.D.; et al. Machine Learning Feature Extraction in Naturalistic Stimuli for Human Brain Mapping Using High-Density Diffuse Optical Tomography. In Proceedings of the SPIE BiOS, San Francisco, CA, USA, 22–27 January 2022. [Google Scholar]
  28. Murad, N.; Pan, M.-C.; Hsu, Y.-F. Reconstruction and Localization of Tumors in Breast Optical Imaging via Convolution Neural Network Based on Batch Normalization Layers. IEEE Access 2022, 10, 57850–57864. [Google Scholar] [CrossRef]
  29. Mozumder, M.; Hauptmann, A.; Nissila, I.; Arridge, S.R.; Tarvainen, T. A Model-Based Iterative Learning Approach for Diffuse Optical Tomography. IEEE Trans. Med. Imaging 2022, 41, 1289–1299. [Google Scholar] [CrossRef]
  30. Ben Yedder, H.; Cardoen, B.; Hamarneh, G. Deep learning for biomedical image reconstruction: A survey. Artif. Intell. Rev. 2021, 54, 215–251. [Google Scholar] [CrossRef]
  31. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  32. Zhang, X.; Yan, C.; Gao, C.; Malin, B.A.; Chen, Y. Predicting Missing Values in Medical Data Via XGBoost Regression. J. Healthc. Inform. Res. 2020, 4, 383–394. [Google Scholar] [CrossRef]
  33. Liew, X.Y.; Hameed, N.; Clos, J. An investigation of XGBoost-based algorithm for breast cancer classification. Mach. Learn. Appl. 2021, 6, 100154. [Google Scholar] [CrossRef]
  34. Langdon, W.B.; Poli, R. Foundations of Genetic Programming; Springer International Publishing: Cham, Switzerland, 2002. [Google Scholar]
  35. Koza, J.R.; Bennett, F.H.; Andre, D.; Keane, M.A. Genetic programming III: Darwinian invention and problem solving [Book Review]. IEEE Trans. Evol. Comput. 2005, 3, 251–253. [Google Scholar] [CrossRef]
  36. Hauptman, A.; Elyasaf, A.; Sipper, M.; Karmon, A. GP-rush: Using genetic programming to evolve solvers for the rush hour puzzle. In Proceedings of the 11th Annual conference on Genetic and Evolutionary Computation (GECCO ‘09), Montreal, QC, Canada, 8–12 July 2009; Association for Computing Machinery: New York, NY, USA, 2009; pp. 955–962. [Google Scholar]
  37. Bertero, M.; Boccacci, P. Introduction to Inverse Problems in Imaging, 2nd ed.; CRC Press: Boca Raton, FL, USA, 2020. [Google Scholar]
  38. Pogue, B.W.; McBride, T.O.; Osterberg, U.L.; Paulsen, K.D. Comparison of imaging geometries for diffuse optical tomography of tissue. Opt. Express 1999, 4, 270–286. [Google Scholar] [CrossRef] [PubMed]
  39. Schweiger, M.; Arridge, S. The Toast++ software suite for forward and inverse modeling in optical tomography. J. Biomed. Opt. 2014, 19, 040801. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Deng, B.; Brooks, D.H.; Boas, D.A.; Lundqvist, M.; Fang, Q. Characterization of structural-prior guided optical tomography using realistic breast models derived from dual-energy x-ray mammography. Biomed. Opt. Express 2015, 6, 2366–2379. [Google Scholar] [CrossRef] [Green Version]
  41. Jacques, S.L. Optical properties of biological tissues: A review. Phys. Med. Biol. 2013, 58, 37–61. [Google Scholar] [CrossRef]
  42. Luo, C.; Zhan, J.; Xue, X.; Wang, L.; Ren, R.; Yang, Q. Cosine normalization: Using cosine similarity instead of dot product in neural networks. In Proceedings of the Artificial Neural Networks and Machine Learning-ICANN 2018, Rhodes, Greece, 4–7 October 2018; pp. 382–391. [Google Scholar]
  43. Takamizu, Y.; Umemura, M.; Yajima, H.; Abe, M.; Hoshi, Y. Deep Learning of Diffuse Optical Tomography based on Time-Domain Radiative Transfer Equation. arXiv 2020, arXiv:2011.125020. [Google Scholar] [CrossRef]
  44. Koza, J.R. Genetic programming as a means for programming computers by natural selection. Stat. Comput. 1994, 4, 87–112. [Google Scholar] [CrossRef]
  45. Ahvanooey, M.T.; Li, Q.; Wu, M.; Wang, S. A survey of genetic programming and its applications. KSII Trans. Internet Inf. Syst. 2019, 13, 1765–1794. [Google Scholar] [CrossRef]
  46. Al-Sahaf, H.; Bi, Y.; Chen, Q.; Lensen, A.; Mei, Y.; Sun, Y.; Tran, B.; Xue, B.; Zhang, M. A survey on evolutionary machine learning. J. R. Soc. N. Z. 2019, 49, 205–228. [Google Scholar] [CrossRef]
  47. Koza, J.R.; Poli, R. Genetic programming. In Search Methodologies: Introductory Tutorials in Optimization and Decision Support Techniques; Springer International Publishing: Cham, Switzerland, 2005; pp. 127–164. [Google Scholar]
  48. Fortin, F.A.; De Rainville, F.M.; Gardner, M.A.; Parizeau, M.; Gagńe, C. DEAP: Evolutionary algorithms made easy. J. Mach. Learn. Res. 2012, 13, 2171–2175. [Google Scholar]
  49. Mudeng, V.; Ayana, G.; Zhang, S.-U.; Choe, S. Progress of Near-Infrared-Based Medical Imaging and Cancer Cell Suppressors. Chemosensors 2022, 10, 471. [Google Scholar] [CrossRef]
Figure 1. Simulated compressed breast mesh with various components and source–detector geometry.
Figure 1. Simulated compressed breast mesh with various components and source–detector geometry.
Bioengineering 10 00382 g001
Figure 2. The reconstructed tumor locations using predictions from the XGBoost machine learning algorithm. (a) showing the reconstructed tumor locations at (z = 5.325 mm), and (b) showing the reconstructed tumor locations at (z = 11.289 mm), are two different examples of the XGBoost predictions. The color bar shows the absorption coefficient in ( cm 1 ).
Figure 2. The reconstructed tumor locations using predictions from the XGBoost machine learning algorithm. (a) showing the reconstructed tumor locations at (z = 5.325 mm), and (b) showing the reconstructed tumor locations at (z = 11.289 mm), are two different examples of the XGBoost predictions. The color bar shows the absorption coefficient in ( cm 1 ).
Bioengineering 10 00382 g002
Figure 3. Schematic block diagram showing the workflow of the proposed methodology. The black and blue boxes show the working of the XGBoost and GP algorithms, respectively.
Figure 3. Schematic block diagram showing the workflow of the proposed methodology. The black and blue boxes show the working of the XGBoost and GP algorithms, respectively.
Bioengineering 10 00382 g003
Figure 4. The reconstructed tumor locations using predictions from the GP post-processing algorithm. (a) showing the reconstructed tumor locations at (z = 14.375 mm), (b) showing the reconstructed tumor locations at (z = 5.325 mm), and (c) showing the reconstructed tumor locations at (z = 11.289), are different examples of the XGBoost and GP predictions. The color bar shows the absorption coefficient in ( cm 1 ).
Figure 4. The reconstructed tumor locations using predictions from the GP post-processing algorithm. (a) showing the reconstructed tumor locations at (z = 14.375 mm), (b) showing the reconstructed tumor locations at (z = 5.325 mm), and (c) showing the reconstructed tumor locations at (z = 11.289), are different examples of the XGBoost and GP predictions. The color bar shows the absorption coefficient in ( cm 1 ).
Bioengineering 10 00382 g004
Table 1. Parameters for the GP algorithm.
Table 1. Parameters for the GP algorithm.
ParameterRange of Values
Population sizeBetween 10,000 and 15,000
Generation countBetween 100 and 250
Reproduction probability0.35
Crossover probability0.5
Mutation probability0.15 (including ERC)
Tree depthBetween 2 and 6
Tournament size4
Table 2. Performance of the proposed model in terms of RMSE.
Table 2. Performance of the proposed model in terms of RMSE.
Values (Units)RMSE (After XGBoost)RMSE (After GP)
X coordinate (mm)0.1862 ± 0.00180.1808 ± 0.0014
Y coordinate (mm)0.1678 ± 0.00420.1539 ± 0.0057
Z coordinate (mm)0.1505 ± 0.00090.1340 ± 0.0032
Radius (mm)0.2157 ± 0.01030.2017 ± 0.0126
μ a   (mm−1)0.1131 ± 0.00910.0975 ± 0.0065
Table 3. Comparison of the XGBoost-GP algorithm with existing algorithms in terms of RMSE.
Table 3. Comparison of the XGBoost-GP algorithm with existing algorithms in terms of RMSE.
S. No.ArticleResearch TypeBackground TypeRMSE
P.Proposed algorithm (XGBoost + GP)SimulationInhomogeneous background mesh (DigiBreast [40])0.12
1.Jaejun Yoo et al. [15] (Neural network for inverting Lippman–Schwinger equation)Simulation and ExperimentHomogeneous background mesh (breast mesh and full body rat mesh)0.66
2.Yun Zou et al. [8](ML-PC model)Simulation and ExperimentHomogeneous background mesh0.30
3.GM. Balasubramaniam et al. [17] (Cascaded feed-forward neural network)SimulationHomogeneous background mesh0.17
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hauptman, A.; Balasubramaniam, G.M.; Arnon, S. Machine Learning Diffuse Optical Tomography Using Extreme Gradient Boosting and Genetic Programming. Bioengineering 2023, 10, 382. https://doi.org/10.3390/bioengineering10030382

AMA Style

Hauptman A, Balasubramaniam GM, Arnon S. Machine Learning Diffuse Optical Tomography Using Extreme Gradient Boosting and Genetic Programming. Bioengineering. 2023; 10(3):382. https://doi.org/10.3390/bioengineering10030382

Chicago/Turabian Style

Hauptman, Ami, Ganesh M. Balasubramaniam, and Shlomi Arnon. 2023. "Machine Learning Diffuse Optical Tomography Using Extreme Gradient Boosting and Genetic Programming" Bioengineering 10, no. 3: 382. https://doi.org/10.3390/bioengineering10030382

APA Style

Hauptman, A., Balasubramaniam, G. M., & Arnon, S. (2023). Machine Learning Diffuse Optical Tomography Using Extreme Gradient Boosting and Genetic Programming. Bioengineering, 10(3), 382. https://doi.org/10.3390/bioengineering10030382

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop