Next Article in Journal
Cardiovascular Tissue Engineering Models for Atherosclerosis Treatment Development
Previous Article in Journal
Biological Prosthesis (Hollow 3D-Printed Titanium Custom-Made Prosthesis and Bone Graft) for Humeral Reconstruction in Pediatric Oncologic Patients: Surgical Indications and Results
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

BlobCUT: A Contrastive Learning Method to Support Small Blob Detection in Medical Imaging

1
School of Computing and Augmented Intelligence, Arizona State University, Tempe, AZ 85281, USA
2
Division Nephrology, Department of Pediatrics, University of Virginia, Charlottesville, VA 22903, USA
3
Department of Radiology, Washington University, St. Louis, MO 63130, USA
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(12), 1372; https://doi.org/10.3390/bioengineering10121372
Submission received: 16 October 2023 / Revised: 19 November 2023 / Accepted: 27 November 2023 / Published: 29 November 2023
(This article belongs to the Section Biosignal Processing)

Abstract

:
Medical imaging-based biomarkers derived from small objects (e.g., cell nuclei) play a crucial role in medical applications. However, detecting and segmenting small objects (a.k.a. blobs) remains a challenging task. In this research, we propose a novel 3D small blob detector called BlobCUT. BlobCUT is an unpaired image-to-image (I2I) translation model that falls under the Contrastive Unpaired Translation paradigm. It employs a blob synthesis module to generate synthetic 3D blobs with corresponding masks. This is incorporated into the iterative model training as the ground truth. The I2I translation process is designed with two constraints: (1) a convexity consistency constraint that relies on Hessian analysis to preserve the geometric properties and (2) an intensity distribution consistency constraint based on Kullback-Leibler divergence to preserve the intensity distribution of blobs. BlobCUT learns the inherent noise distribution from the target noisy blob images and performs image translation from the noisy domain to the clean domain, effectively functioning as a denoising process to support blob identification. To validate the performance of BlobCUT, we evaluate it on a 3D simulated dataset of blobs and a 3D MRI dataset of mouse kidneys. We conduct a comparative analysis involving six state-of-the-art methods. Our findings reveal that BlobCUT exhibits superior performance and training efficiency, utilizing only 56.6% of the training time required by the state-of-the-art BlobDetGAN. This underscores the effectiveness of BlobCUT in accurately segmenting small blobs while achieving notable gains in training efficiency.

1. Introduction

Imaging biomarkers are measurable features or characteristics that are detected using medical imaging technologies, such as magnetic resonance imaging (MRI), computed tomography (CT), positron emission tomography (PET), or ultrasound. These biomarkers provide objective and quantitative measures of disease or injury and are often used in medical research and clinical practice to diagnose, monitor, and evaluate the progression or treatment response of the disease. In particular, some important imaging biomarkers are derived from small, densely distributed structures (e.g., cell nuclei, kidney glomeruli) known as “blobs” in vivo. The characterization of blobs based on features such as shape, size, and quantity has led to their use as imaging biomarkers to aid in diagnosis, prognosis, and treatment planning [1,2,3,4]. Thus, the detection and segmentation of blobs in medical images are critical for accurately assessing and interpreting disease states.
Similar to its versatile applications in various fields like manufacturing [5,6] and pool boiling [7,8,9], Artificial Intelligence (AI) has emerged as a transformative force in the field of medicine as well, offering numerous advantages that enhance patient care and streamline healthcare processes. One of the key benefits is the ability of AI to analyze vast amounts of medical data quickly and accurately, aiding in early disease detection and diagnosis. Machine learning algorithms can identify patterns in medical images, such as X-rays and MRIs, enabling healthcare professionals to make more informed decisions [10,11]. Concurrently, the escalating demand for AI support in the medical domain serves as a catalyst for advancements in various realms of computer science, notably in the domain of computer vision.
However, detecting and segmenting blobs is a difficult computer vision task for several reasons. Firstly, acquiring labeled data, a prerequisite for numerous deep learning algorithms, which is costly for blob-like data, proves to be an exceedingly challenging task. And blobs have varying and often irregular shapes, which complicates the segmentation process. Additionally, medical images are often characterized by low resolution, image noise, and other challenges that hinder the accurate segmentation of small, intricate blobs. Moreover, the detection and recognition of blobs’ boundaries are challenging, especially when blobs overlap. As a result, sophisticated algorithms are necessary to address these challenges and accurately detect and segment blobs in medical images.
Current methods of segmentation often struggle to handle the intricacies of blob detection, particularly in medical imaging. Traditional approaches, such as U-Net [12] and U-Net-like models [13,14], represent a class of fully supervised models commonly employed in various image segmentation tasks. However, their efficacy is contingent upon the availability of annotated data, a prerequisite that becomes particularly challenging and costly, especially in scenarios involving the detection of small blobs. Although leveraging the advantages of transfer learning techniques, which aim to capitalize on knowledge gained from related tasks, can mitigate the problem, it may face challenges in adapting to domain shifts, and the pre-trained knowledge may not always be perfectly aligned with the target task, potentially leading to suboptimal performance in certain scenarios. Moreover, U-Net and U-Net-like models often depend on thresholding techniques applied to probability maps. The problem with this method is that under-segmentation arises when small blobs or objects are not properly segmented due to the thresholding process. When segmenting small blobs, setting a fixed threshold may lead to the exclusion of parts of the object that fall below the threshold, even if they are part of the object of interest. This can result in incomplete segmentation and missing small but crucial details, leading to a loss of information and accuracy in the segmentation task, as demonstrated by Xu et al. [14].
The importance of efficient small blobs segmentation techniques holds paramount importance in the realm of medical imaging and diagnostics. Specifically, it is evident in medical applications such as identifying glomeruli from kidney Cationic-ferritin-enhanced MR (CFE-MR) images which are roughly spherical and have blob-like shapes in 3D space. Glomeruli, being the functional units of the kidney, play a crucial role in the filtration of blood and the formation of urine. Accurate identification and segmentation of glomeruli in three-dimensional space contribute to a better understanding of renal function, aiding in the diagnosis and monitoring of kidney-related disorders. The 3D nature of glomeruli, with their roughly spherical and blob-like shapes, poses a unique challenge for segmentation techniques. The ability to precisely detect and segment these structures can provide valuable insights into the health of the kidneys.
In light of the limitations in current segmentation methods and the impact of addressing these limitations on the medical imaging and diagnosis field, the proposed work aims to overcome these challenges by introducing a novel approach that allows unpaired image-to-image translation from medical images to label masks. The proposed model, named BlobCUT, is motivated by the CUT model [15] and specifically designed for the detection and segmentation of glomeruli. This innovative method involves a multi-step process to achieve robust segmentation. First, a 3D elliptical Gaussian function is utilized to synthesize 3D blobs and their masks, simulating the geometric properties of glomeruli. This generated data serves as the clean data domain (target domain) for training, to which the noisy glomeruli data needs to be translated. Second, a modification to the CUT model is introduced, incorporating a convexity consistency constraint to minimize the impact of noise and ensure the preservation of blobs’ locations, shapes, and data distributions. Third, a distribution consistency constraint based on Kullback–Leibler divergence is proposed to maintain the same distribution between denoised and synthetic blob images, leveraging information at the voxel level. Finally, the model is capable of denoising glomerular images and deriving the Hessian convexity mask to facilitate blob identification.
To validate the performance of BlobCUT, two studies were conducted. The first study is conducted on a 3D simulated blobs dataset where the locations of blobs are known. The study included six methods from the literature for comparasion, namely UH-DoG [13], BTCAS [14], CUT [15], UVCGAN [16], EGSDE [17], and the state-of-the-art BlobDetGAN [18]. The second study was performed on a set of 3D images of mouse kidneys from CFE-MRI, where BlobCUT was compared against UH-DoG, BTCAS, UVCGAN, EGSDE, BlobDetGAN, and stereology [19]. This comprehensive assessment aims to demonstrate the efficacy and superiority of BlobCUT in addressing the challenges posed by current segmentation methods in the context of glomeruli identification. In this context, the proposed 3D BlobCUT model becomes not only a technical innovation but also a vital tool for advancing our understanding of renal physiology and pathology, making it a crucial application in the field of medical image segmentation.
In summary, this research presents three significant contributions:
  • A novel small blob detector: the proposed BlobCUT leverages Generative Adversarial Network (GAN) and Contrastive Learning to address the challenge of limited labeled data. By incorporating the assumption of geometric properties, BlobCUT offers a novel, effective and efficient solution for detecting small blobs.
  • Novel constraints for improved performance: This work introduces a novel convexity consistency constraint, based on Hessian analysis, and a unique blob distribution consistency constraint, based on Kullback–Leibler divergence. These constraints are designed to preserve the geometry property and intensity distribution of blobs, resulting in enhanced segmentation performance.
  • Comprehensive performance evaluation: Extensive comparisons were conducted with six state-of-the-art methods, spanning both simulated and real MRI datasets. The outcomes of these comparisons affirm the superior performance of BlobCUT, underlining its effectiveness in small blob detection compared to existing methodologies.
The rest of the paper is organized as follows. Section 2 illustrates the advantages and disadvantages of some related works. Section 3 presents our methodology to detect and segment small blobs. Section 4 describes our experiment setup and results evaluation. Section 5 discusses the potential to apply the proposed method to another medical scenario and other areas. Section 6 concludes contributions and results followed by discussing limitations and future work.

2. Literature Review

Several works have been introduced to address the small blob detection problem. Kong et al. proposed a generalized Laplacian of Gaussian (gLoG) [20] filter for detecting general elliptical blob structures in images. The gLoG detector is able to accurately locate the blob centroids and estimate the sizes, shapes, and orientations of the detected blobs. Zhang et al. developed a series of blob detectors including the Hessian-based Laplacian of Gaussian (HLoG) [21] and the Hessian-based Difference of Gaussian (HDoG) [22]. The advantage of the Hessian analysis is that it can automatically extract the regional features, but the problem is that it makes the unsupervised clustering-based post-pruning more sensitive to noise than the traditional threshold-based post-pruning [23] commonly used in most imaging detectors [13]. Moreover, blobs’ background in medical images is often heterogeneous (e.g., uneven light of background), leading to the over-detection problem [14].
Recent studies have demonstrated the effectiveness of deep-learning-based segmentation methods in detecting and segmenting large objects in medical images [1,3,4,24,25]. However, due to the small size and dense distribution of blobs in the images, the performance of such methods on blob segmentation is questionable [13]. Thresholding on the probability map from commonly used deep learning methods, such as U-net, has been shown to lead to under-segmentation problems when segmenting small blobs, as demonstrated by Xu et al. [14]. Instead, researchers have explored using deep learning models’ denoising ability to support blob detection. U-Net-based methods have been proposed for blob detection, such as the UH-DoG blob detector [13], which was developed primarily for detecting glomeruli in Cationic-ferritin-enhanced MRI (CFE-MRI) images. The UH-DoG model was pre-trained on a nuclei dataset, and the blobs’ probability map in the CFE-MR images was derived. It was then combined with a Hessian convexity map to identify true glomeruli. To improve the computational efficiency and segmentation accuracy of glomeruli detection and segmentation, BTCAS [14] was developed as an extension of the UH-DoG detector. This approach transforms blobs to local optimum DoG space adaptively by utilizing the bounded scales from U-Net.
Essentially, methods based on U-Net are supervised and heavily dependent on the quality of the training dataset. However, the scarcity and costliness of labeled medical data impose constraints on the applicability of U-Net-based approaches for blob segmentation. To overcome the issues of noisy medical data and the rarity of labels, researchers started utilizing unpaired image-to-image (UI2I) translation models which are label-free. UI2I translation models are machine learning algorithms used to transform images from one domain to another without the need for paired data. One popular type of unpaired image-to-image translation model is based on generative adversarial network (GAN) [26], which consists of two neural networks: a generator and a discriminator. The generator network learns to create new images similar to the target domain, while the discriminator network learns to distinguish between the generated images and real images from the target domain. The two networks are trained together in a game-like process, where the generator tries to create more realistic images and the discriminator tries to identify the generated images. This process continues until the generator is able to create images that are indistinguishable from real images in the target domain. These models can be used for a variety of tasks, such as style transfer [27], image colorization [28], and image super-resolution [29]. Utilizing the power of GAN, Zhao et al. [17] proposed EGSDE which uses energy-guided stochastic differential equations to help training, and Torbunov et al. [16] further integrated a CycleGAN [30] with U-Net-ViT [31], achieving state-of-the-art performance in UI2I tasks.
Recently, researchers started realizing the potential of UI2I translation to solve image detection and segmentation problems. By considering objects with noise as one domain and objects without noise as another domain, the noisy data is translated into the clean data’s domain without the need for any labels. Xu et al. [32] proposed BlobDetGAN, a CycleGAN-based model to solve blob detection in two steps: image denoising and object segmentation. BlobDetGAN is entirely label-free as it leverages both the blobs’ geometric properties and noise distribution to denoise the real noisy image to the clean image with only target blobs. However, there are two disadvantages to using CycleGAN in this application. The first is that CycleGAN consists of two generators and two discriminators and was designed to support two-way direction training (Domain A → Domain B and Domain B → Domain A). However, training four networks simultaneously makes it unnecessary and time-consuming since only one direction is required for the denoising application. The second disadvantage is that the cycle consistency constraint used in CycleGAN-based models is sometimes too restrictive to the relationship between the two domains [15]. Taesung et al. simplified the CycleGAN from two generators and two discriminators to only one generator and one discriminator by using contrastive learning to propose a Contrastive Unpaired Translation (CUT) model [15]. CUT offers an alternative straightforward way of maintaining correspondence in content but not appearance by maximizing the mutual information between the corresponding input and output images. Although CUT may have better image translation capability and help with reducing the time required for training, it will not work well directly when denoising images that consist of small objects such as blobs and background noise. The aforementioned statement is attributed to the fact that CUT utilizes the dissimilarities between image patches as a source of information. Notably, these patches are not specifically engineered to detect discrepancies at the level of individual pixels/voxels. Rather, the focus of the method is on extracting collective statistics from the comparison of patches. In our denoising problem, pixels/voxels are also as important as the collective statistics because small blobs have similar sizes, shapes, and data distributions with noise and a pixel/voxel change may cause big differences in biomarker extraction. Thus, the CUT model needs to be modified to work properly to detect small blobs.

3. Materials and Methods

As Figure 1 illustrates, our proposed BlobCUT consists of three steps to identify the blobs (e.g., glomeruli) from the kidney CFE-MRI images: (1) Synthesize 3D blobs by using the 3D elliptical Gaussian function and randomly generate 3D blobs image with their blob mask as training input and shape constraints; (2) train a 3D CUT model with convexity consistency constraint and distribution constraint to translate real 3D blob images (e.g., real kidney CFE-MRI) to clean 3D image and produce a probability mask and a Hessian convexity mask; (3) A voxel union constraint operation is applied on the probability mask and the Hessian convexity mask to derive the final identification mask of blobs.

3.1. 3D Blob Synthesis through 3D Gaussian Function

Our proposed BlobCUT includes blob synthesis as the first step of blob detection. Blobs can range in size, shape, orientation and location in images. Some work has been done to use Gaussian function to construct small blobs in 2D space [20,33]. However, in 3D space, the shape and orientation of blobs are more complex. BlobDetGAN [32,34] firstly simulates 3D blobs by using the 3D elliptical Gaussian function from Equation (1), and it is proven to be effective and efficient in the literature.
F ( x , y , z ) = A · e ( a x 2 + b y 2 + c z 2 + d x y + e y z + f x z )
a = sin 2 θ cos 2 φ σ x 2 + sin 2 θ sin 2 φ σ y 2 + cos 2 θ σ z 2
b = cos 2 θ cos 2 φ σ x 2 + cos 2 θ sin 2 φ σ y 2 + sin 2 θ σ z 2
c = sin 2 φ σ x 2 + cos 2 φ σ y 2
d = sin 2 θ cos 2 φ σ x 2 + sin 2 θ sin 2 φ σ y 2 sin 2 θ σ z 2
e = cos θ sin 2 φ σ x 2 + cos θ sin 2 φ σ y 2
f = sin θ sin 2 φ σ x 2 + sin θ sin 2 φ σ y 2
where x, y and z are the coordinates of voxels near the blob center which is fixed at the origin point. The shape and orientation of 3D blobs are controlled by coefficients a, b, c, d, e and f via θ , φ , σ x , σ y and σ z . θ and φ are two angles in the spherical coordinate system, and σ defines the length along each direction. A is a normalization factor. More details are illustrated in Appendix A.
In our study, we employ the aforementioned formulas to generate 3D blobs in the same way of [32,34] and designate their masks as ground truth. During the synthesis of these images, the parameters, location, and quantity of blobs within each image are chosen randomly. Our primary focus is on segmenting glomeruli from kidneys, and therefore, BlobCUT is used to convert real 3D blob images (glomeruli images) into approximate synthesized blob images. The main purpose of synthesizing these 3D blob images is to approximate the actual distribution of the target blobs (glomeruli) while retaining their information, and only eliminating contextual details during the image translation process.

3.2. 3D Blob Images Detecting through 3D GAN with Contrastive Learning

We wish to translate images from the input domain–real noisy 3D glomeruli which we denote as the source domain S, to appear like an image from the output domain–synthetic clean 3D blob which we denote as the target domain T. We are given a dataset of unpaired instances. Our method only requires learning the mapping in one direction and avoids using inverse auxiliary generators and discriminators which are used in CycleGAN-based model. This can largely simplify the training procedure and reduce training time.
With training samples of real noisy blob images { I n } S and clean synthetic blob images { I c } T , we denote their respective data distributions as I n p d a t a ( I n ) and I c p d a t a ( I c ) .
The generator G consists of two parts, an encoder network G e n c and a decoder network G d e c , which are applied sequentially to produce output image y = G ( z ) = G d e c ( G e n c ( x ) ) . To facilitate effective model training, multiple loss functions were incorporated. The first is the Adversarial loss. Adversarial loss function [26] is a common loss function used in GAN training to incentivize the generated output to exhibit visual congruity with images originating from the designated target domain, as follows:
L G A N ( G , D ) = E I c p d a t a ( I c ) [ log ( D ( I c ) ) ] + E I n p d a t a ( I n ) [ 1 log ( D ( G ( I n ) ) ) ]
where G and D are the functions of the generator and discriminator of the basic GAN model of CUT, X is the input image, which is typically a real 3D glomeruli image, and Y is the output image, which is denoised 3D glomeruli image in our situation.
Another loss function that was incorporated is a loss function based on the Contrastive learning loss. The overall target of unpaired image-to-image translation methods is to associate the input and output data, and the loss function is the criterion to measure the correlations of images from different domains. However, traditional unpaired image-to-image translation methods suffer from the same issue-the need for a pre-specified, hand-designed loss function to measure predictive performance [35,36]. Recently, a family of methods based on maximizing mutual information has emerged to bypass the above issue [35,36,37,38,39,40,41]. These approaches employ noise contrastive estimation (NCE) [42] for the purpose of learning an embedding space that brings related signals into proximity while distinguishing them from other samples within the dataset. It is noteworthy that analogous concepts have historical roots in seminal work on metric learning, exemplified by Siamese networks [43]. Associated signals can be an image with itself [38,41,44,45,46], an image with its downstream representation [47], neighboring patches within an image [35,39,48], or multiple views of the input image [49], and most successfully, an image with a set of transformed versions of itself [37,40]. Thus, by maximizing the mutual information, we can address the above issue.
Contrastive Unpaired Translation (CUT) model first uses patchwise contrastive loss [35] under a noise contrastive estimation (NCE) framework [35] to maximize the mutual information between input and output at the patch level. It can enable one-sided image translation, which is more efficient than CycleGAN-based methods, with comparable good quality. The core concept underlying contrastive learning is the establishment of associations between two signals, denoted as the ’query’ and its corresponding ’positive’ example, as opposed to the remaining data points within the dataset, which are commonly referred to as ’negatives’. Thus, the cross-entropy loss can be calculated, representing the probability of the positive example being selected over the negatives. In our context, query refers to a small patch of clean 3D glomeruli images (output). Positives and negatives are the same and different position patches from real noisy 3D glomeruli images (input). For example, given a patch showing a lot of glomeruli from the output image, one should be able to more strongly associate it with the corresponding cortical area from the input image, more so than the other patches of the medullar areas or other noise areas. Even at the pixel/voxel level, the intensity of a glomerulus (close to zero) can be more strongly associated with the intensity of the glomerulus than with the background noise. Thus, we employ a multilayer, patch-based learning objective. We aim to match corresponding input-output patches at a specific location area (cortex, medullar, and noisy area). We can leverage the other area patches within the input as negatives. The multilayer patchwise contrastive loss (denoted as PatchNCE loss), as illustrated in Figure 2, is selected to be used:
L P a t c h N C E ( G , F , X ) = E x X l = 1 L s = 1 S t l ( z l s , z l S / s )
where x is a patch from an image X, and F is the feature map generated from a two-layer fully connected network after the generator. We pass patches of the image through a small two-layer MLP network F same as [37], generating features z l s = F l ( G e n c ( x ) ) and the other features in other locations as z l S / s . l is the layer index l 1 , 2 , , L and s 1 , 2 , , S l where S l is the number of locations in images. You can find more details in [15]. In our works, we pass both source domain images I n and target domain images I c into the network to guarantee the translation quality. Thus we have two patch NCE loss L P a t c h N C E ( G , F , I n ) and L P a t c h N C E ( G , F , I c ) .
However, based on the above settings, any small blobs’ geometric transformation in the translation have little effect on their distribution and cannot be identified by discriminators. The CUT utilizes the dissimilarities between image patches as a source of information. However, it is important to note that these patches are not specifically engineered to detect discrepancies at the level of individual pixels/voxels. Rather, the focus of the method is on extracting collective statistics from the comparison of patches. In our denoising problem, pixels/voxels are also as important as the collective statistics because small blobs have similar sizes, shapes, and data distributions with noise and a pixel/voxel change may cause big differences in biomarker extraction. Normally, directly applying the CUT model to kidney MRI images will cause geometric distortion and object blur, which essentially manifest as intensity changes at the pixel/voxel level. Thus, the translated 3D blobs in G d e c with the geometric distortion will have an inconsistent geometric property with the glomeruli in R e a l A . Therefore, in addition to the above losses, we propose a Convexity consistency constraint loss that is based on Hessian analysis in BlobCUT to preserve blob voxels’ convexity invariance to satisfy the convexity property of glomeruli.
L c o n v e x ( G , H I , M ) = E I n p d a t a ( I n ) [ | | ( H I ( J G ( I n ) ) H I ( J I n ) ) M ( I n ) | | 1 ]
where J denotes the unit matrix with all ones, and H I ( f ) is the Hessian indicator to measure the convexity of voxels in f, and it is defined by the corresponding Hessian matrix H ( · ) . The details can be found at Equations (11) and (12)
H ( I f ( i , j , k ) ) = 2 I f ( i , j , k ) i 2 2 I f ( i , j , k ) i j 2 I f ( i , j , k ) i k 2 I f ( i , j , k ) i j 2 I f ( i , j , k ) j 2 2 I f ( i , j , k ) j k 2 I f ( i , j , k ) i k 2 I f ( i , j , k ) j k 2 I f ( i , j , k ) k 2
H I ( i , j , k ) ( I f ) = 1 H ( I f ( i , j , k ) ) < 0 0 H ( I f ( i , j , k ) ) 0
This binary indicator matrix (referred subsequently to as Hessian convexity mask) identifies the blob voxels as 1 and the rest as 0. To preserve the geometry of 3D blobs in the translated images G ( I n ) , G ( I n ) should have fewer convex voxels than I n . Since G ( I n ) is a normalized image with darker blobs as compared to the background, we define the convexity consistent constraint as:
v = 1 N H I v ( J G ( I n ) ) v = 1 N H I v ( J I n )
where J denotes the unit matrix with all ones, N denotes the total number of voxels in I n , and v denotes index of voxel in I n . Equation (13) means the translated images should always have less or equal total value of Hessian indicator (i.e., less or equal convex objects) than that of input images.
Finally, It is not enough for the generative model to have the convexity consistency constraint to denoise the real glomeruli, because the original CUT was not designed to capture pixel/voxel differences. Thus, we enforce a Probability distribution constraint which involves a Kullback-Leibler divergence (KLD) loss to minimize the distance between two intensity distributions. We compare the distributions of the output denoised 3D gloms with our synthetic 3D blobs to make the BlobCUT have good generation quality. After obtaining the translated image ( G ( I n ) ), the number of glom can be determined by using connected component analysis, assuming there is no noise in the image. Subsequently, a random image with an equivalent number of gloms is generated using the 3D elliptical Gaussian function. These two images are then fed back into the encoder G e n c of the generator G to obtain the intensity distribution representations of the images. The Kullback-Leibler Divergence (KLD) loss is calculated for the two intensity distributions. By minimizing the KLD loss, the translated image can be ensured to have a similar distribution to the synthetic image. This approach effectively addresses the issue of small object blurring that is present in the original CUT model-generated image by utilizing the intensity distribution information.
L K L D ( G e n c , S ) = E x [ log p ( G e n c ( G ( I n ) ) ) log p ( G ( S ) ) ]
where S ( x ) is the distribution of synthetic clean blobs. G e n c is the encoder part of the generator.
Final objective. Our final objective is as follows. The generated image should exhibit realism, with a simultaneous requirement for the establishment of correspondence between patches in the input and output images. The final objective loss function becomes as follows:
L t o t a l = L G A N ( G , D ) + λ I n L P a t c h N C E ( G , F , I n ) + λ I c L P a t c h N C E ( G , F , I c ) + λ c o n v L c o n v e x ( G , H I , M ) + λ K L D L K L D ( G e n c , S )
where λ I n , λ I c , λ c o n v , and λ K L D are the parameters to tune our proposed method to get good denoising performance. Once trained, BlobCUT has the ability to perform image translation for the purpose of blob image denoising and obtaining a glom/blob probability mask.

3.3. 3D Blob Identification through Voxel Union Constraint Operation

The BlobCUT approach presented in Section 3.2 provides the Hessian convexity mask H I ( I n ) and the blob probability mask M ( G ( I n ) ) . Prior works [14] have shown that Hessian analysis tends to over-segment the blobs and the U-Net probability map tends to under-segment the blobs. To resolve the issues associated with Hessian analysis and U-Net in [14], we similarly use a voxel union constraint operation to derive the final blobs. Towards this, the final blob identification mask M f i n a l ( I n ) is derived by applying the joint operation on the two masks as follows:
M f i n a l ( I n ) = H I ( I n ) M ( G ( I n ) )
where the joint operator ⊙ is the Hadamard product. Figure 3 shows the step-by-step methodology of blob identification through BlobCUT and joint constraint operation. Figure 3a–c shows the original noisy blobs image, the clean blob image of blobs without noise, and the centroids of the blobs’ ground truth in respective order. While Figure 3d–e shows the Hessian convexity mask and the blob mask that were generated using BlobCUT. Finally, Figure 3f shows the final blob identification mask generated by the joint operation. The blue circle in parts (a), (b), and (c) shows a total of four blobs, three of which are overlapping while the fourth is separated. In the Hessian convexity mask shown in part (d), they are correctly identified as four distinct blobs. However, In the blob mask shown in part (e), all four blobs are incorrectly segmented as one blob highlighting the under-segmentation issue discussed previously. The yellow-colored ellipse shows a segment of the noise that needs to be cleaned from the image. It is obvious from part (e) that the blob mask is able to distinguish this segment as pure noise and successfully remove it while the Hessian mask in part (d) incorrectly identifies the noise as blobs, causing some of the noisy regions to be detected as blobs. Which highlights the over-segmentation issue in the Hessian mask. Therefore, by taking the intersection of the Hessian convexity mask and the blob mask, their respective issues of under-segmentation and over-segmentation are alleviated. The effectiveness of this operation has been validated by multiple significant experiments in [32]. The final blob mask derived by the joint constraint operation is shown in Figure 3f. Clearly, it is able to selectively segment the blobs while avoiding over-segmentation of noisy regions.
We define the true 3D blob candidate as a 26-connected voxel, so the final identified blobs set are represented as:
S b l o b = { ( i , j , k ) ( i , j , k ) I d , I b ( i , j , k ) = 1 }

4. Experiments and Results

4.1. Networks and Hardware

BlobCUT is directly trained on 3D images. We take 3D images with 64 × 64 × 32 voxels as input, then resize them to 128 × 128 × 64 voxels and normalize to [ 0 , 1 ] . The generator G adopts an encoder-decoder structure with residual blocks, similar to [50]. The Generator consists of a 7 × 7 × 7 3D Convolution-InstanceNormalization-ReLU layer. Then followed by 2 Downsampling layers, 6 Residual blocks, 2 Upsampling layers, and Tanh loss function. Each Downsampling layer is a 3 × 3 × 3 3D Convolution-InstanceNormalization-ReLU layer with stride size 2. Each Residual block consists of two 3 × 3 × 3 3D Convolution-InstanceNormalization-ReLU layer. Each Upsampling layer is a 3 × 3 × 3 3D fractional-strided-Convolution-InstanceNormalization-ReLU layer with a stride size of 2. A 3D replication padding with a padding size of 1 is used in all Convolution layers. The discriminator D adopts a 70 × 70 PatchGAN [51] for classifying the real images and translated images. It consists of three 4 × 4 × 4 3D Convolution-InstanceNormalization-LeakyReLU layers with a stride size of 2, one 4 × 4 × 4 3D Convolution-InstanceNormalization-LeakyReLU layer with a stride size of 1 and one 4 × 4 × 4 3D Convolution layer with a stride size of 1. The LeakyReLU in each layer has a slope of 0.2. Finally, a Sigmoid activation function is applied in the output layer.
Our experiments are conducted on the platform with 1 Intel CORE i7-10700K CPU, manufactured by Intel Corporation based in Santa Clara, California, USA, and 1 NVIDIA RTX 3090 GPU with 24 GB of memory, sourced from NVIDIA Corporation, also headquartered in Santa Clara, California, USA.

4.2. Dataset Description

We conducted two experiments to validate the performance of BlobCUT. The first experiment is applying the BlobCUT to synthetic noisy 3D blob images to see the significant improvement compared to the other methods. The second experiment is applying the BlobCUT to real mouse kidney CFE-MR images to check the segmentation performance on a real dataset. BlobCUT needs training datasets from source domain S and target domain T as input. Target domain T in BlobCUT should be fixed as synthetic clean 3D blobs, and all of them were generated by the function in Section 3.
In the context of the target domain, we randomly generated a dataset consisting of 1000 3D blob images, each with dimensions of 64 × 64 × 32 voxels, for use in both experimental settings. For each 3D blob image, blobs were spread in random locations. The number of blobs for each image ranged from 500 to 800. The parameters of the 3D elliptical Gaussian function for each synthetic blob are as follows: ϕ and θ are set to x σ 1 σ 2 are all set to [ 0.5 , 1.5 ] . We record the blob mask of each image.
For source domain S, in the first experiments, we synthesized another 1000 3D blobs images using the same 3D blobs synthesis function in Section 3, and record their label mask as ground truth. Then random noise was added to these images to make them become noisy 3D blob images. Noise was generated by the Gaussian function with μ = 0 and δ defined by:
σ n o i s e 2 = σ i m a g e 2 10 S N R 10
where the signal-to-noise ratio S N R was randomly set from 0.01 db to 1 db. The 1000 synthetic clean 3D blobs images and 1000 noisy 3D blobs images are treated as the training dataset for the first experiment. Figure 4 demonstrates the source domain and target domain images used by BlobCUT during training in Experiment I.
In the second experiment, we studied 62 mice kidney CFE-MR images as source domain S. Each mouse kidney MR image has voxel dimensions of 256 × 256 × 256 . The dataset comprises two distinct comparison groups: chronic kidney disease (CKD) with a sample size of 26 individuals in contrast to 18 controls, and acute kidney injury (AKI) with 10 individuals compared to 8 controls. The animal experiments were approved by the Institutional Animal Care and Use Committee (IACUC) under protocol #3929 on 4 July 2020 at the University of Virginia, in accordance with the National Institutes of Health Guide for the Care and Use of Laboratory Animals. They were imaged by CFE-MRI as described in [52]. For validation, we split the 62 mice kidneys into training and validation sets. The training dataset includes 44 mice kidneys, and the validation dataset consists of 18 mice kidneys. We randomly sampled 1000 3D patch images ( 64 × 64 × 32 voxels) from 44 mice kidneys. These 3D patch images (Figure 5) have varied locations in mice kidneys and were non-overlapping. The sampling process is performed in the cortex region because the medulla region of mouse kidney does not have glomeruli. The medulla and cortex regions of the mice kidneys were meticulously annotated by an expert in the field. The training dataset for the second experiment comprises 1000 synthetic 3D blob images and 1,000 3D patch images extracted from mice kidneys. Figure 4 demonstrates the source domain and target domain images used by BlobCUT during training in Experiment II.

4.3. Experiment I: 3D Synthetic Image Data

Because medical kidney images with experts’ labeling are rare and expensive, we use synthetic data that have label masks (ground truth) instead. Thus, with the label masks, we can quantify the segmentation performance of different methods. Our proposed method, BlobCUT, underwent a comparative analysis against six other methods: UHDoG [13], BTCAS [14], BlobDetGAN [18], UVCGAN [16], EGSDE [17], and CUT [15]. The selection of UHDoG and BTCAS was motivated by their pretraining on a public cellular dataset, allowing us to illustrate the advantages of eliminating the reliance on additional domain information. While BlobDetGAN is a state-of-the-art method in glom segmentation, its extended training time presented a challenge. The inclusion of UVCGAN, EGSDE, and CUT (acknowledged as cutting-edge techniques in unpaired image-to-image translation) serves the purpose of highlighting the intrinsic challenges associated with small blob detection. The objective is to demonstrate that, even among the most advanced methods in unpaired image-to-image translation (UI2I), achieving satisfactory performance in addressing the small blob detection problem is challenging without specialized design considerations. This rationale informs the choice of these specific methods for comparison against our proposed approach.
We evaluated the performance of these six methods by the following seven metrics: Detection Error Rate (DER), Precision, Recall, F-score, Dice coefficient, Intersection over Union (IoU) and Blobness. For detection, DER measures the difference ratio between number of detected blobs and ground truth. Precision quantifies the proportion of retrieved candidates that are validated by the ground truth, while recall quantifies the proportion of ground-truth data that is successfully retrieved. The F-score provides a comprehensive evaluation of the combined performance of precision and recall. In the context of segmentation, the Dice coefficient assesses the similarity between the segmented blob mask and the ground truth, and the Intersection over Union (IoU) measures the degree of overlap between the segmented blob mask and the ground truth. For synthesis, Blobness evaluates the likelihood of an object having a blob-like shape.
Since the blob locations (the coordinates of the blob centers) were already generated when synthesizing the 3D blob images, the ground-truth number of blobs for all 3D images was already recorded. DER can be calculated by Equation (19).
D E R = | N G T N D e t | N G T
where N G T represents the number of ground-truth blobs and N D e t represents the number of detected blobs. A candidate was classified as a true positive if the centroid of its magnitude fell within a detection pair ( i , j ) , where the nearest ground truth center j had not been previously associated, and the Euclidean distance D i j between ground truth center j and blob candidate i was less than or equal to the specified threshold d. To avoid double-counting, the number of true positives T P was computed using Equation (20). Precision, recall, and F-score were computed as per Equations (21)–(23).
T P = min { # { ( p i , p j ) : min i = 1 m D p i p j d δ } , # { ( p i , p j ) : min j = 1 n D p i p j d δ } }
P r e c i s i o n = T P n
R e c a l l = T P m
F s c o r e = 2 × P r e c i s i o n × R e c a l l P e c i s i o n + R e c a l l
where m is the number of true glomeruli and n is the number of blob candidates; d represents a threshold with a positive value in the range of ( 0 , + ) . When d is set to a smaller value, fewer blob candidates are considered, as it requires a closer proximity between the centroid of a blob candidate and the corresponding ground truth. Conversely, when d is set too large, more blob candidates are taken into account. In this context, given the potential presence of local intensity maxima within small, irregularly shaped blobs, we configure d to be equal to the average diameter of the blobs. The Dice coefficient and IoU were computed by comparing the segmented blob mask with the ground truth mask using Equations (24) and (25).
D i c e ( B M , B G ) = 2 | B M B G | | B M | + | B G |
I o U ( B M , B G ) = B M B G B M B G
where B M is the binary mask for segmentation result and B G is the binary mask for the ground truth. Based on identified blobs set S b l o b , blobness for each blob candidate b i is calculated by Equation (26):
B l o b n e s s b i S b l o b = 3 × | d e t ( H ( J f ) ) | 2 3 p m ( H ( J f ) )
where f denotes the normalized 3D blobs image, J represents the unit matrix with all ones, H symbolizes the Hessian matrix, and p m indicates the principal minors of the Hessian matrix. We make the assumption that the blobs in f correspond to dark blobs in Equation (26) for consistency with other equations. It’s important to note that in [22], the calculation of blobness is based on the DoG transformed blobs image, whereas here, we use the normalized 3D blobs image without the DoG transformation. Consequently, the transformation of these dark blobs into bright blobs is achieved by J f .
Note that since we already know blob mask of each synthetic image, we calculated the average blobness of all images as ground truth (in this experiment, the average blobness for all images is 0.519). UHDoG, BTCAS, UVCGAN and EGSDe firstly detect the gloms in 2D-space and then stack the results as 3D outputs. We limited our training time comparison to BlobCUT, BlobDetGAN, and 3D CUT, as these networks are specifically tailored for 3D segmentation, providing a more relevant and comparable benchmark.
Table 1 shows the comparative performance results of BlobCUT against SOTA methods on the synthetic data. In terms of evaluation metrics, BlobCUT attains best values in Dice, IoU, and Blobness. It demonstrates comparable performance to BlobDetGAN in DER, recall, and F-score, with a marginally lower precision. The diminished performance of UHDoG, BTCAS, UVCGAN, and EGSDE in Dice, IoU, and Blobness, relative to BlobCUT and BlobDetGAN, is unsurprising, given their reliance on two-dimensional spatial designed methods. In the assessment of F-score, the three generic methods in unpaired image-to-image translation (UVCGAN, EGSDE, and CUT) exhibit lower values than the blob detectors (UHDoG, BTCAS, BlobDetGAN, and BlobCUT). This underscores the inherent challenge of small blob detection, where generic approaches, lacking meticulous design modifications beyond alterations to the training set, struggle to attain commendable performance. Not only do the empirical findings presented in Table 1 indicate the quantitative superiority of BlobCUT against SOTA methods, but they also show that BlobCUT proves to be a faster and more efficient method for training models, completing the training process in 1393 s per epoch, a notable improvement over BlobDetGAN’s 2171 s per epoch.
Furthermore, as Figure 6 shows, it is noteworthy that BlobCUT attains its peak performance at 23 epochs, whereas BlobDetGAN reaches its optimal performance at 26 epochs. Accounting for the temporal investment per epoch, BlobCUT outperforms the state-of-the-art method BlobDetGAN, achieving superior results with only 56.6% of its training time. This emphasizes the efficacy of BlobCUT in precisely segmenting small blobs, concurrently achieving significant improvements in training efficiency.

4.4. Experiment II: 3D MR Images of Mouse Kidney

We conducted experiments involving CF-labeled glomeruli within a dataset of 3D magnetic resonance (MR) images to quantify the total count, denoted as N g l o m , of glomeruli in both healthy and diseased mouse kidneys. The segmentation of glomeruli was performed using four blob detectors: UH-DoG, BTCAS, BlobDetGAN, and the proposed BlobCUT. Additionally, we included two advanced unpaired image-to-image (UI2I) methods, UVCGAN and EGSDE, for comparative analysis. The results of HDoG with VBGMM from [19], confirmed by physicians are used as the ground truth in this paper. The parameter settings of DoG were: window size N = 7 , γ = 2 , Δ s = 0.001 . To denoise the 3D blob images by using trained U-Net, we firstly resized each slice to 512 × 512 by bilinear interpolation and each slice was fed into U-Net. The final probability map of whole kidney is reconstructed by combining all 2D patches and used for UH-DoG. The threshold for the U-Net probability map in UH-DoG was 0.5. Since mouse kidney has voxel dimensions of 256 × 256 × 256 , to validate through BlobDetGAN and BlobCUT, we divided each mouse kidney as 128 3D patches ( 64 × 64 × 32 ). The final identification mask of the whole kidney is also reconstructed by combining all 3D patches. To train the BlobDetGAN, λ c y c l e and λ c o n v e x were set to 10, λ i d e n t i t y was set to 0.5. They were both trained from scratch with a learning rate 0.0002. The training typically took about 20 epochs to converge so we did not set up the decay policy for learning rate. We used the Adam optimizer with a batch size set to 1.
We perform quality control by visually checking the identified glomeruli, visible as black spots in the images. For illustration, Figure 7 shows the comparison of glomerular segmentation results on mouse kidneys of CKD group and CKD control group using UH-DoG, BTCAS, BlobDetGAN, UVCGAN, EGSDE and BlobCUT, respectively. From the zoom-in regions from Figure 7, BlobCUT and BlobDetGAN exhibit relatively superior performance, manifested in their ability to accurately identify gloms without excessive omission of correct gloms. Moreover, they demonstrate outstanding resistance to image defects, such as dark edges (#429-111, #466-096). In contrast, UHDoG and BTCAS show comparatively inferior performance. Due to its reliance on a single threshold, UHDoG’s performance is less stable, as observed in its excellent performance in #466-096 but less impressive performance in #429-111. It is prone to noise interference, resulting in an over-segmentation problem. Conversely, BTCAS exhibits more stability in performance and is less susceptible to noise interference compared to UHDoG. However, in comparison to BlobCUT, it has a lower recognition rate, attributed to the limitations imposed by utilizing a pre-trained U-Net. As for the remaining two generic models, UVCGAN and EGSDE, designed for UI2I tasks, manifest similar issues of under-segmentation and sensitivity to noise. This is attributed to their lack of dedicated geometric and intensity constraints for glom detection, which causes shape distortion and unseparated overlapping gloms during inference.
N g l o m is reported in Table 2, where the UH-DoG, BTCAS, BlobDetGAN, UVCGAN, EGSDE and the proposed BlobCUT blob detectors are compared to HDoG with VBGMM from [22]. The differences between the results are also listed in Table 2. We observe that the performance of UHDoG is not very stable, showing differences of 1% and 6% within the same experimental group, primarily attributed to its lack of robustness when employing a single threshold on the probability map generated by U-Net in conjunction with the Hessian convexity map. In contrast, BTCAS, due to its multi-threshold design, exhibits greater stability compared to UHDoG but does show significant deviations in certain kidneys (ID 427, 462, 476). As mentioned earlier, UVCGAN and EGSDE share a common issue of under-segmentation, as evident in the table. Regarding BlobCUT and BlobDetGAN, both perform well on the dataset of mouse kidney magnetic resonance images. Similar to the previous experiment, BlobCUT demonstrates yet again its computational superiority over BlobDetGAN, as it achieves this performance using approximately 60% of the training time required by BlobDetGAN.

5. Discussion: Denoising and Applications

5.1. Denoising

Imaging biomarkers derived from blob images hold substantial potential to inform clinical decision-making. Illustratively, the glomerulus-based imaging biomarkers ( N g l o m ) examined in Experiment II stand as pivotal indicators in clinical trials, mitigating the financial and procedural burdens associated with early detection research in kidney diseases. Despite the clinical significance of these biomarkers, their historical acquisition has predominantly relied on invasive stereological methodologies, often conducted post-mortem. Notably, beyond the methodologies elucidated in this paper, a majority of AI approaches are founded upon histological section images, enhanced resolution but with higher costs.
The advanced development of Magnetic Resonance Imaging (MRI) technology facilitates the measurement of these biomarkers in a non-invasive imaging approach. However, glomeruli are relatively small compared to the imaging resolution, and they exhibit a visual appearance similar to noise. Therefore, denoising is indispensable for this kind of medical imaging detection and segmentation task.
Research [13] indicates that Hessian analysis lacks robustness against noise, making noise easily misidentified as false positives. To address this issue, UH-DoG and BTCAS utilize probability maps from a pretrained U-Net to denoise blob images. However, this may result in suboptimal denoising performance, as objects in pre trained public datasets still have different distributions from noise. To illustrate this point, Figure 8 compares the denoised results of U-Net, 3D CUT, and BlobCUT. From Figure 8b, it is evident that the blue circles in Figure 8a represent noise, but they appear as blobs in Figure 8c and are not visible in Figure 8d,e. This suggests that 3D CUT and BlobCUT outperform U-Net in denoising performance. However, noise may indirectly impact the translation of blobs, altering their shapes. Comparing the yellow circles in Figure 8b,d, it is clear that there are differences in the shapes of the blobs. The blobs within the yellow circle in Figure 8e exhibit a much similar shape to those in Figure 8b. This demonstrates that BlobCUT is capable of preserving the geometric properties of blobs during translation. However, without constraint on translation, CUT may experience geometric distortion. In conclusion, BlobCUT excels in denoising performance and effectively shields blobs from the influence of noise during the denoising process.

5.2. Applications

Aside from its application in detecting glomeruli in kidney MRI images, with its denoising capabilities and proficiency in detecting small blobs, BlobCUT can extend its utility to various projects in the medical imaging domain, including tasks such as nuclei/cell detection [53,54] and microparticle picking in electron microscopy imaging [55,56]. Taking nuclei detection as an illustrative example: 1. BlobCUT facilitates the identification and delineation of cell nuclei, thereby supporting the diagnosis and grading of diseases. In cancer diagnosis, for instance, the characterization of nuclei offers insights into tumor stage and aggressiveness. 2. BlobCUT enables quantitative analysis of diverse cellular features, including size, shape, and density. These objective measures contribute to a standardized assessment of tissue characteristics. 3. BlobCUT aids in comprehending the spatial distribution of cell nuclei, assisting in the development of targeted therapies and determining the extent of surgical interventions.
Moreover, It is noteworthy that the potential of BlobCUT is not limited to medical applications. BlobCUT demonstrates considerable potential across various real-world applications such as microbiological detection [57,58] and crowd counting [59,60]. One noteworthy application is in sonar image analysis, specifically for discerning objects or structures on the seafloor. Sonar images, with their diverse seabed structures, present challenges for accurate detection [61]. By integrating BlobCUT, a segmentation technique that distinguishes object features from the background, sonar images can be effectively segmented. This integration facilitates the identification of prominent objects, such as sunken ships or archaeological sites, enhancing feature detection and ensuring robustness in different seabed environments. Beyond maritime applications, BlobCUT finds relevance in Industrial Inspection and Quality Control. In industrial settings, where imagery analysis is pivotal for quality control and defect detection [62,63,64], BlobCUT can be employed to identify and segment-specific defects or irregularities in manufactured products, thereby enhancing the precision and efficiency of quality assessment processes. Additionally, BlobCUT extends its utility to Material Science and Microstructure Analysis. In the microscopic examination of materials, which often reveals intricate microstructures and particulate systems [65,66,67], BlobCUT can be valuable for segmenting and characterizing these microstructural elements. This capability facilitates advanced material analysis and supports research in fields such as metallurgy, nanotechnology, and material science.

6. Conclusions and Future Work

In this research, we propose BlobCUT for small blob detection and segmentation. This work provides three main contributions to the literature. First, we propose a novel 3D small blob detector BlobCUT which combines contrastive learning and generative adversarial network; Second, a blob distribution consistency constraint based on Kullback–Leibler divergence to maintain the same blobs’ distribution before and after denoising is constructed and implemented. Third, a convexity consistency constraint to preserve the blobs’ geometry property is implemented to improve the performance. In the validation experiments, BlobCUT significantly outperforms the widely used methods HDoG, UH-DoG, BTCAS and the original CUT model, yielding more accurate imaging biomarkers. Furthermore, BlobCUT exhibits comparable performance to the state-of-the-art method, requiring only around 56.6% of the training time.
While BlobCUT exhibits noteworthy efficacy in detecting small blobs, there remains scope for enhancement. The algorithm relies on a geometric assumption (e.g., for glomeruli detection is the elliptical shape) regarding blob shape, a constraint that may prove overly stringent in certain scenarios. Notably, BlobCUT exhibited a tendency to over-detect gloms in diseased kidneys of female animals, as these kidney MRIs often featured irregularly shaped gloms compared to their healthy counterparts. Furthermore, the algorithm’s performance faces skepticism from physicians due to its reliance on synthetic training data. Despite achieving high metrics in isolation, the lack of human guidance renders the results less meaningful. To address these limitations, we think it is necessary to leverage the expertise of human annotators and to optimize resource utilization using active learning based approaches. Our future endeavors include integrating human-in-the-loop active learning, a strategy designed to refine the training process. By doing so, we aim to cultivate a model that garners greater credibility among physicians and enhances the segmentation of small blobs in medical images.

Author Contributions

T.L., Y.X., T.W. and F.A.-H. wrote the main manuscript. T.L. and Y.X. developed the computational methods and conducted the experiments. J.R.C. and K.M.B. acquired and provided the imaging data. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by funds from the National Institute of Health award under grant number R01DK110622, R01DK111861. This work used the Bruker ClinScan 7T MRI in the Molecular Imaging Core which was purchased with support from NIH grant 1S10RR019911-01 and is supported by the University of Virginia School of Medicine.

Institutional Review Board Statement

Three human kidneys were obtain from the International Institute for the Advancement of Medicine (Edison, NJ) and were deemed non-human research by the Arizona State University IRB [68]. The animal experiments were approved by the Institutional Animal Care and Use Committee (IACUC) under protocol #3929 on 4 July 2020 at the University of Virginia, in accordance with the National Institutes of Health Guide for the Care and Use of Laboratory Animals.

Informed Consent Statement

Informed consent was waived as the human kidneys were deemed nonhuman research.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to confidentiality.

Conflicts of Interest

Author Jennifer R. Charlton is co-owner of Sindri Technologies, LLC; consultant for XN Biotechnologies; consultant for Medtronics; President Elect of the Board of the Neonatal Kidney Collaborative; investor in Zorro-Flow. Author Kevin M. Bennett is co-owner of XN Biotechnologies, LLC, Sindri Technologies LLC, and Nephrodiagnostics, LLC. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Appendix A

Proposition A1. 
Let F ( x , y , z ) be the 3D elliptical Gaussian function with the following general form:
F ( x , y , z ) = A · e ( ( a ( x x 0 ) 2 + b ( y y 0 ) 2 + c ( z z 0 ) 2 + d ( x x 0 ) ( y y 0 ) + e ( y y 0 ) ( z z 0 ) + f ( x x 0 ) ( z z 0 ) ) )
where A is a normalization factor, and x 0 , y 0 and z 0 are the coordinates of the center of the Gaussian function F ( x , y , z ) .
Assume the coefficients a, b, c, d, e and f control the shape and orientation of F ( x , y , z ) via θ, φ, σ x , σ y and σ z . We want to show that:
a = sin 2 θ cos 2 φ σ x 2 + sin 2 θ sin 2 φ σ y 2 + cos 2 θ σ z 2
b = cos 2 θ cos 2 φ σ x 2 + cos 2 θ sin 2 φ σ y 2 + sin 2 θ σ z 2
c = sin 2 φ σ x 2 + cos 2 φ σ y 2
d = sin 2 θ cos 2 φ σ x 2 + sin 2 θ sin 2 φ σ y 2 sin 2 θ σ z 2
e = cos θ sin 2 φ σ x 2 + cos θ sin 2 φ σ y 2
f = sin θ sin 2 φ σ x 2 + sin θ sin 2 φ σ y 2
Figure A1. Spherical coordinate systems.
Figure A1. Spherical coordinate systems.
Bioengineering 10 01372 g0a1
Proof of Proposition A1. 
Assume the probability density for two normally distributed variables is given by:
F ( x ) = A · e ( u a ) T B ( u a )
where A is a normalization factor, u = ( u 1 , u 2 , u 3 ) T and a = ( a 1 , a 2 , a 3 ) T . B is the inverse of the covariance matrix
C = σ 1 2 c o v ( u 1 , u 2 ) c o v ( u 1 , u 3 ) c o v ( u 1 , u 2 ) σ 2 2 c o v ( u 2 , u 3 ) c o v ( u 1 , u 3 ) c o v ( u 2 , u 3 ) σ 3 2
where σ 1 , σ 2 , σ 3 are the standard diviation of u 1 , u 2 , u 3 . With the correlation coefficient
ρ 12 = c o v ( u 1 , u 2 ) σ 1 σ 2
ρ 23 = c o v ( u 2 , u 3 ) σ 2 σ 3
ρ 13 = c o v ( u 1 , u 3 ) σ 1 σ 3
B is given by
B = 1 E σ 1 2 ( 1 ρ 23 2 ) σ 1 1 σ 2 1 ( ρ 13 ρ 23 ρ 12 ) σ 1 1 σ 3 1 ( ρ 12 ρ 23 ρ 13 ) σ 1 1 σ 2 1 ( ρ 13 ρ 23 ρ 12 ) σ 2 2 ( 1 ρ 13 2 ) σ 2 1 σ 3 1 ( ρ 12 ρ 13 ρ 23 ) σ 1 1 σ 3 1 ( ρ 12 ρ 23 ρ 13 ) σ 2 1 σ 3 1 ( ρ 12 ρ 13 ρ 23 ) σ 3 2 ( 1 ρ 12 2 )
where E = 1 ρ 12 2 ρ 13 2 ρ 23 2 + 2 ρ 12 ρ 13 ρ 23 , so
( u v ) T B ( u v ) = C
Let u 1 = u 1 a 1 , u 2 = u 2 a 2 , u 3 = u 3 a 3 , let C be a constant C = 1 , the Equation (A14) can be rewritten as:
( u v ) T B ( u v ) = 1 E [ u 1 2 ( 1 ρ 23 2 ) σ 1 2 + 2 u 1 u 2 ( ρ 13 ρ 23 ρ 12 ) σ 1 σ 2 + u 2 2 ( 1 ρ 13 2 ) σ 2 2 + 2 u 1 u 3 ( ρ 12 ρ 23 ρ 13 ) σ 1 σ 3 + u 3 2 ( 1 ρ 12 2 ) σ 3 2 + 2 u 2 u 3 ( ρ 12 ρ 13 ρ 23 ) σ 2 σ 3 ] = 1
In this problem, since we consider u 1 , u 2 , u 3 as three orthogonal directions in the 3D coordinate systems, to simplify the model, we can consider ρ 12 = ρ 23 = ρ 13 = 0 , then
( u v ) T B ( u v ) = u 1 2 σ 1 2 + u 2 2 σ 2 2 + u 3 2 σ 3 2 = 1
To consider the orientations of blobs, we need to transfer Equation (A16) into Spherical coordinate systems, so units can be transferred as:
x y z = sin θ cos φ sin θ sin φ cos θ cos θ cos φ cos θ sin φ sin θ sin φ cos φ 0 u 1 u 2 u 3
Then
u 1 u 2 u 3 = sin θ cos φ cos θ cos φ sin φ sin θ sin φ cos θ sin φ cos φ cos θ sin θ 0 x y z
Then
( u v ) T B ( u v ) = ( sin θ cos φ x + cos θ cos φ y sin φ z ) 2 σ 1 2 + ( sin θ sin φ x + cos θ sin φ y + cos φ z ) 2 σ 2 2 + ( cos θ x sin θ y ) 2 σ 3 2
Since σ 1 2 = σ x 2 , σ 2 2 = σ y 2 , σ 3 2 = σ z 2 , let x = x x 0 , y = y y 0 , z = z z 0
( u v ) T B ( u v ) = ( x x 0 ) 2 ( sin 2 θ cos 2 φ σ x 2 + sin 2 θ sin 2 φ σ y 2 + cos 2 θ σ z 2 ) + ( y y 0 ) 2 ( cos 2 θ cos 2 φ σ x 2 + cos 2 θ sin 2 φ σ y 2 + sin 2 θ σ z 2 ) + ( z z 0 ) 2 ( sin 2 φ σ x 2 + cos 2 φ σ y 2 ) + ( x x 0 ) ( y y 0 ) ( sin 2 θ cos 2 φ σ x 2 + sin 2 θ sin 2 φ σ y 2 sin 2 θ σ z 2 ) + ( y y 0 ) ( z z 0 ) ( cos θ sin 2 φ σ x 2 + cos θ sin 2 φ σ y 2 ) + ( x x 0 ) ( z z 0 ) ( sin θ sin 2 φ σ x 2 + sin θ sin 2 φ σ y 2 )
Let F ( x , y , z ) be the 3D elliptical Gaussian function with the following general form:
F ( x , y , z ) = A · e ( ( a ( x x 0 ) 2 + b ( y y 0 ) 2 + c ( z z 0 ) 2 + d ( x x 0 ) ( y y 0 ) + e ( y y 0 ) ( z z 0 ) + f ( x x 0 ) ( z z 0 ) ) )
where A is a normalization factor. Then,
a = sin 2 θ cos 2 φ σ x 2 + sin 2 θ sin 2 φ σ y 2 + cos 2 θ σ z 2
b = cos 2 θ cos 2 φ σ x 2 + cos 2 θ sin 2 φ σ y 2 + sin 2 θ σ z 2
c = sin 2 φ σ x 2 + cos 2 φ σ y 2
d = sin 2 θ cos 2 φ σ x 2 + sin 2 θ sin 2 φ σ y 2 sin 2 θ σ z 2
e = cos θ sin 2 φ σ x 2 + cos θ sin 2 φ σ y 2
f = sin θ sin 2 φ σ x 2 + sin θ sin 2 φ σ y 2

References

  1. Dolz, J.; Gopinath, K.; Yuan, J.; Lombaert, H.; Desrosiers, C.; Ayed, I.B. HyperDense-Net: A hyper-densely connected CNN for multi-modal image segmentation. IEEE Trans. Med. Imaging 2018, 38, 1116–1126. [Google Scholar] [CrossRef]
  2. Chang, S.; Chen, X.; Duan, J.; Mou, X. A CNN-based hybrid ring artifact reduction algorithm for CT images. IEEE Trans. Radiat. Plasma Med. Sci. 2020, 5, 253–260. [Google Scholar] [CrossRef]
  3. Gupta, H.; Jin, K.H.; Nguyen, H.Q.; McCann, M.T.; Unser, M. CNN-based projected gradient descent for consistent CT image reconstruction. IEEE Trans. Med. Imaging 2018, 37, 1440–1453. [Google Scholar] [CrossRef] [PubMed]
  4. Li, W.; Li, J.; Sarma, K.V.; Ho, K.C.; Shen, S.; Knudsen, B.S.; Gertych, A.; Arnold, C.W. Path R-CNN for prostate cancer diagnosis and gleason grading of histological images. IEEE Trans. Med. Imaging 2018, 38, 945–954. [Google Scholar] [CrossRef] [PubMed]
  5. Altarazi, S.; Allaf, R.; Alhindawi, F. Machine learning models for predicting and classifying the tensile strength of polymeric films fabricated via different production processes. Materials 2019, 12, 1475. [Google Scholar] [CrossRef] [PubMed]
  6. Alhindawi, F.; Altarazi, S. Predicting the tensile strength of extrusion-blown high density polyethylene film using machine learning algorithms. In Proceedings of the 2018 IEEE International Conference on Industrial Engineering and Engineering Management (IEEM), Bangkok, Thailand, 16–19 December 2018; pp. 715–719. [Google Scholar]
  7. Al-Hindawi, F.; Soori, T.; Hu, H.; Siddiquee, M.M.R.; Yoon, H.; Wu, T.; Sun, Y. A framework for generalizing critical heat flux detection models using unsupervised image-to-image translation. Expert Syst. Appl. 2023, 227, 120265. [Google Scholar] [CrossRef]
  8. Al-Hindawi, F.; Siddiquee, M.M.R.; Wu, T.; Hu, H.; Sun, Y. Domain-knowledge Inspired Pseudo Supervision (DIPS) for unsupervised image-to-image translation models to support cross-domain classification. Eng. Appl. Artif. Intell. 2024, 127, 107255. [Google Scholar] [CrossRef]
  9. Rassoulinejad-Mousavi, S.M.; Al-Hindawi, F.; Soori, T.; Rokoni, A.; Yoon, H.; Hu, H.; Wu, T.; Sun, Y. Deep learning strategies for critical heat flux detection in pool boiling. Appl. Therm. Eng. 2021, 190, 116849. [Google Scholar] [CrossRef]
  10. Moridian, P.; Ghassemi, N.; Jafari, M.; Salloum-Asfar, S.; Sadeghi, D.; Khodatars, M.; Shoeibi, A.; Khosravi, A.; Ling, S.H.; Subasi, A.; et al. Automatic autism spectrum disorder detection using artificial intelligence methods with MRI neuroimaging: A review. Front. Mol. Neurosci. 2022, 15, 999605. [Google Scholar] [CrossRef]
  11. Shoeibi, A.; Ghassemi, N.; Khodatars, M.; Moridian, P.; Khosravi, A.; Zare, A.; Gorriz, J.M.; Chale-Chale, A.H.; Khadem, A.; Rajendra Acharya, U. Automatic diagnosis of schizophrenia and attention deficit hyperactivity disorder in rs-fMRI modality using convolutional autoencoder model and interval type-2 fuzzy regression. In Cognitive Neurodynamics; Springer: Berlin/Heidelberg, Germany, 2022; pp. 1–23. [Google Scholar]
  12. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  13. Xu, Y.; Wu, T.; Gao, F.; Charlton, J.R.; Bennett, K.M. Improved small blob detection in 3D images using jointly constrained deep learning and Hessian analysis. Sci. Rep. 2020, 10, 326. [Google Scholar] [CrossRef]
  14. Xu, Y.; Wu, T.; Charlton, J.R.; Gao, F.; Bennett, K.M. Small blob detector using bi-threshold constrained adaptive scales. IEEE Trans. Biomed. Eng. 2020, 68, 2654–2665. [Google Scholar] [CrossRef] [PubMed]
  15. Park, T.; Efros, A.A.; Zhang, R.; Zhu, J.Y. Contrastive learning for unpaired image-to-image translation. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 319–345. [Google Scholar]
  16. Torbunov, D.; Huang, Y.; Yu, H.; Huang, J.; Yoo, S.; Lin, M.; Viren, B.; Ren, Y. Uvcgan: Unet vision transformer cycle-consistent gan for unpaired image-to-image translation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 702–712. [Google Scholar]
  17. Zhao, M.; Bao, F.; Li, C.; Zhu, J. Egsde: Unpaired image-to-image translation via energy-guided stochastic differential equations. Adv. Neural Inf. Process. Syst. 2022, 35, 3609–3623. [Google Scholar]
  18. Xu, Y.; Wu, T.; Gao, F. Deep Learning based Blob Detection Systems and Methods. US Patent 17/698,750, 6 October 2022. [Google Scholar]
  19. Beeman, S.C.; Zhang, M.; Gubhaju, L.; Wu, T.; Bertram, J.F.; Frakes, D.H.; Cherry, B.R.; Bennett, K.M. Measuring glomerular number and size in perfused kidneys using MRI. Am. J. Physiol.-Ren. Physiol. 2011, 300, F1454–F1457. [Google Scholar] [CrossRef] [PubMed]
  20. Kong, H.; Akakin, H.C.; Sarma, S.E. A generalized Laplacian of Gaussian filter for blob detection and its applications. IEEE Trans. Cybern. 2013, 43, 1719–1733. [Google Scholar] [CrossRef]
  21. Zhang, M.; Wu, T.; Bennett, K.M. Small blob identification in medical images using regional features from optimum scale. IEEE Trans. Biomed. Eng. 2014, 62, 1051–1062. [Google Scholar] [CrossRef] [PubMed]
  22. Zhang, M.; Wu, T.; Beeman, S.C.; Cullen-McEwen, L.; Bertram, J.F.; Charlton, J.R.; Baldelomar, E.; Bennett, K.M. Efficient small blob detection based on local convexity, intensity and shape information. IEEE Trans. Med. Imaging 2015, 35, 1127–1137. [Google Scholar] [CrossRef] [PubMed]
  23. Xu, Y.; Gao, F.; Wu, T.; Bennett, K.M.; Charlton, J.R.; Sarkar, S. U-net with optimal thresholding for small blob detection in medical images. In Proceedings of the 2019 IEEE 15th International Conference on Automation Science and Engineering (CASE), Vancouver, BC, Canada, 22–26 August 2019; pp. 1761–1767. [Google Scholar]
  24. Tan, J.; Gao, Y.; Liang, Z.; Cao, W.; Pomeroy, M.J.; Huo, Y.; Li, L.; Barish, M.A.; Abbasi, A.F.; Pickhardt, P.J. 3D-GLCM CNN: A 3-dimensional gray-level Co-occurrence matrix-based CNN model for polyp classification via CT colonography. IEEE Trans. Med. Imaging 2019, 39, 2013–2024. [Google Scholar] [CrossRef]
  25. Zreik, M.; Van Hamersvelt, R.W.; Wolterink, J.M.; Leiner, T.; Viergever, M.A.; Išgum, I. A recurrent CNN for automatic detection and classification of coronary artery plaque and stenosis in coronary CT angiography. IEEE Trans. Med. Imaging 2018, 38, 1588–1598. [Google Scholar] [CrossRef]
  26. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  27. Karras, T.; Laine, S.; Aila, T. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4401–4410. [Google Scholar]
  28. Nazeri, K.; Ng, E.; Ebrahimi, M. Image colorization using generative adversarial networks. In Proceedings of the Articulated Motion and Deformable Objects: 10th International Conference, AMDO 2018, Palma de Mallorca, Spain, 12–13 July 2018; pp. 85–94. [Google Scholar]
  29. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  30. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  31. Wang, L.; Li, R.; Zhang, C.; Fang, S.; Duan, C.; Meng, X.; Atkinson, P.M. UNetFormer: A UNet-like transformer for efficient semantic segmentation of remote sensing urban scene imagery. ISPRS J. Photogramm. Remote Sens. 2022, 190, 196–214. [Google Scholar] [CrossRef]
  32. Xu, Y. Novel Computational Algorithms for Imaging Biomarker Identification. Ph.D. Thesis, Arizona State University, Tempe, AZ, USA, 2022. [Google Scholar]
  33. Wang, G.; Lopez-Molina, C.; De Baets, B. Blob reconstruction using unilateral second order Gaussian kernels with application to high-ISO long-exposure image denoising. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4817–4825. [Google Scholar]
  34. Xu, Y.; Wu, T.; Charlton, J.R.; Bennett, K.M. GAN Training Acceleration Using Fréchet Descriptor-Based Coreset. Appl. Sci. 2022, 12, 7599. [Google Scholar] [CrossRef]
  35. Oord, A.V.D.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
  36. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
  37. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning. PMLR—2020, Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  38. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9729–9738. [Google Scholar]
  39. Henaff, O. Data-efficient image recognition with contrastive predictive coding. In Proceedings of the International conference on machine learning. PMLR—2020, Virtual, 13–19 July 2020; pp. 4182–4192. [Google Scholar]
  40. Misra, I.; Maaten, L.V.D. Self-supervised learning of pretext-invariant representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6707–6717. [Google Scholar]
  41. Wu, Z.; Xiong, Y.; Yu, S.X.; Lin, D. Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3733–3742. [Google Scholar]
  42. Gutmann, M.; Hyvärinen, A. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics—JMLR Workshop and Conference Proceedings, Chia Laguna Resort, Italy, 13–15 May 2010; pp. 297–304. [Google Scholar]
  43. Chopra, S.; Hadsell, R.; LeCun, Y. Learning a similarity metric discriminatively, with application to face verification. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 539–546. [Google Scholar]
  44. Malisiewicz, T.; Gupta, A.; Efros, A.A. Ensemble of exemplar-svms for object detection and beyond. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 16–17 November 2011; pp. 89–96. [Google Scholar]
  45. Shrivastava, A.; Malisiewicz, T.; Gupta, A.; Efros, A.A. Data-driven visual similarity for cross-domain image matching. ACM Trans. Graph. 2011, 30, 154. [Google Scholar] [CrossRef]
  46. Alexey, D.; Fischer, P.; Tobias, J.; Springenberg, M.R.; Brox, T. Discriminative unsupervised feature learning with exemplar convolutional neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1734–1747. [Google Scholar]
  47. Hjelm, R.D.; Fedorov, A.; Lavoie-Marchildon, S.; Grewal, K.; Bachman, P.; Trischler, A.; Bengio, Y. Learning deep representations by mutual information estimation and maximization. arXiv 2018, arXiv:1808.06670. [Google Scholar]
  48. Isola, P.; Zoran, D.; Krishnan, D.; Adelson, E.H. Learning visual groups from co-occurrences in space and time. arXiv 2015, arXiv:1511.06811. [Google Scholar]
  49. Tian, Y.; Krishnan, D.; Isola, P. Contrastive multiview coding. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 776–794. [Google Scholar]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  51. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  52. Charlton, J.R.; Xu, Y.; Wu, T.; deRonde, K.A.; Hughes, J.L.; Dutta, S.; Oxley, G.T.; Cwiek, A.; Cathro, H.P.; Charlton, N.P.; et al. Magnetic resonance imaging accurately tracks kidney pathology and heterogeneity in the transition from acute kidney injury to chronic kidney disease. Kidney Int. 2021, 99, 173–185. [Google Scholar] [CrossRef]
  53. Hollandi, R.; Moshkov, N.; Paavolainen, L.; Tasnadi, E.; Piccinini, F.; Horvath, P. Nucleus segmentation: Towards automated solutions. Trends Cell Biol. 2022, 32, 295–310. [Google Scholar] [CrossRef]
  54. Basu, A.; Senapati, P.; Deb, M.; Rai, R.; Dhal, K.G. A survey on recent trends in deep learning for nucleus segmentation from histopathology images. In Evolving Systems; Springer: Berlin/Heidelberg, Germany, 2023; pp. 1–46. [Google Scholar]
  55. Gyawali, R.; Dhakal, A.; Wang, L.; Cheng, J. Accurate cryo-EM protein particle picking by integrating the foundational AI image segmentation model and specialized U-Net. bioRxiv 2023. [Google Scholar] [CrossRef]
  56. Zhang, X.; Zhao, T.; Chen, J.; Shen, Y.; Li, X. EPicker is an exemplar-based continual learning approach for knowledge accumulation in cryoEM particle picking. Nat. Commun. 2022, 13, 2468. [Google Scholar] [CrossRef] [PubMed]
  57. Pawłowski, J.; Majchrowska, S.; Golan, T. Generation of microbial colonies dataset with deep learning style transfer. Sci. Rep. 2022, 12, 5212. [Google Scholar] [CrossRef] [PubMed]
  58. Xu, D.; Liu, B.; Wang, J.; Zhang, Z. Bibliometric analysis of artificial intelligence for biotechnology and applied microbiology: Exploring research hotspots and frontiers. Front. Bioeng. Biotechnol. 2022, 10, 998298. [Google Scholar] [CrossRef] [PubMed]
  59. Fan, Z.; Zhang, H.; Zhang, Z.; Lu, G.; Zhang, Y.; Wang, Y. A survey of crowd counting and density estimation based on convolutional neural network. Neurocomputing 2022, 472, 224–251. [Google Scholar] [CrossRef]
  60. Khan, M.A.; Menouar, H.; Hamila, R. Revisiting crowd counting: State-of-the-art, trends, and future perspectives. Image Vis. Comput. 2022, 129, 104597. [Google Scholar] [CrossRef]
  61. Tueller, P.; Kastner, R.; Diamant, R. Target detection using features for sonar images. IET Radar, Sonar Navig. 2020, 14, 1940–1949. [Google Scholar] [CrossRef]
  62. Pierleoni, P.; Belli, A.; Palma, L.; Palmucci, M.; Sabbatini, L. A machine vision system for manual assembly line monitoring. In Proceedings of the 2020 International Conference on Intelligent Engineering and Management (ICIEM), London, UK, 17–19 June 2020; pp. 33–38. [Google Scholar]
  63. De Vitis, G.A.; Foglia, P.; Prete, C.A. Algorithms for the detection of blob defects in high speed glass tube production lines. In Proceedings of the 2019 IEEE 8th International Workshop on Advances in Sensors and Interfaces (IWASI), Otranto, Italy, 13–14 June 2019; pp. 97–102. [Google Scholar]
  64. De Vitis, G.A.; Di Tecco, A.; Foglia, P.; Prete, C.A. Fast Blob and Air Line Defects Detection for High Speed Glass Tube Production Lines. J. Imaging 2021, 7, 223. [Google Scholar] [CrossRef] [PubMed]
  65. DeCost, B.L.; Holm, E.A. A computer vision approach for automated analysis and classification of microstructural image data. Comput. Mater. Sci. 2015, 110, 126–133. [Google Scholar] [CrossRef]
  66. Agbozo, R.; Jin, W. Quantitative metallographic analysis of GCr15 microstructure using mask R-CNN. J. Korean Soc. Precis. Eng. 2020, 37, 361–369. [Google Scholar] [CrossRef]
  67. Ge, M.; Su, F.; Zhao, Z.; Su, D. Deep learning analysis on microscopic imaging in materials science. Mater. Today Nano 2020, 11, 100087. [Google Scholar] [CrossRef]
  68. Beeman, S.C.; Cullen-McEwen, L.A.; Puelles, V.G.; Zhang, M.; Wu, T.; Baldelomar, E.J.; Dowling, J.; Charlton, J.R.; Forbes, M.S.; Ng, A.; et al. MRI-based glomerular morphology and pathology in whole human kidneys. Am. J. Physiol.-Ren. Physiol. 2014, 306, F1381–F1390. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Steps to identify blobs by using BlobCUT.
Figure 1. Steps to identify blobs by using BlobCUT.
Bioengineering 10 01372 g001
Figure 2. The training process of our proposed BlobCUT.
Figure 2. The training process of our proposed BlobCUT.
Bioengineering 10 01372 g002
Figure 3. Illustration of blob identification through joint constraint operation. (a) Original noisy blobs image. (b) Clean blob image of blobs without noise. (c) Ground truth of blob centers. (d) Hessian convexity mask. (e) Blob mask from networks. (f) Final blob identification mask.
Figure 3. Illustration of blob identification through joint constraint operation. (a) Original noisy blobs image. (b) Clean blob image of blobs without noise. (c) Ground truth of blob centers. (d) Hessian convexity mask. (e) Blob mask from networks. (f) Final blob identification mask.
Bioengineering 10 01372 g003
Figure 4. Illustration of the training datasets used by different experiments. For Exp. I, synthetic noisy blob images were used as the source domain images; for Exp. II, real kidney MR images were used as the source domain images. For both experiments, synthetic clean blob images were used as target domain images to encourage the model to have denoising capability.
Figure 4. Illustration of the training datasets used by different experiments. For Exp. I, synthetic noisy blob images were used as the source domain images; for Exp. II, real kidney MR images were used as the source domain images. For both experiments, synthetic clean blob images were used as target domain images to encourage the model to have denoising capability.
Bioengineering 10 01372 g004
Figure 5. Illustration of training input images of BlobCUT (a) Synthesized 3D blobs image from domain clean 3D blobs. (b) Blob mask of (a). (c) Synthesized 3D noisy blobs image from domain noisy 3D blobs. (d) Synthesized 3D mouse kidney image patch from domain noisy 3D blobs. (e) Real 3D mice kidney image patch from domain T.
Figure 5. Illustration of training input images of BlobCUT (a) Synthesized 3D blobs image from domain clean 3D blobs. (b) Blob mask of (a). (c) Synthesized 3D noisy blobs image from domain noisy 3D blobs. (d) Synthesized 3D mouse kidney image patch from domain noisy 3D blobs. (e) Real 3D mice kidney image patch from domain T.
Bioengineering 10 01372 g005
Figure 6. Illustration of the learning curve. (a) Loss curve comparison between training and validating of BlobCUT. (b) Testing F-score comparison between BlobCUT and BlobDetGAN.
Figure 6. Illustration of the learning curve. (a) Loss curve comparison between training and validating of BlobCUT. (b) Testing F-score comparison between BlobCUT and BlobDetGAN.
Bioengineering 10 01372 g006
Figure 7. Comparison of glomerular segmentation results from 3D MR images of mouse kidneys using BlobCUT, BlobDetGAN, UH-DOG, BTCAS, UVCGAN and EGSDE. Identified glomeruli are marked in red. Three slices are illustrated: kidney #429 slice 111, #466 slice 96 and #469 slice 81.
Figure 7. Comparison of glomerular segmentation results from 3D MR images of mouse kidneys using BlobCUT, BlobDetGAN, UH-DOG, BTCAS, UVCGAN and EGSDE. Identified glomeruli are marked in red. Three slices are illustrated: kidney #429 slice 111, #466 slice 96 and #469 slice 81.
Bioengineering 10 01372 g007
Figure 8. Denoising results of noisy synthetic blobs image using U-Net, 3D CUT, BlobCUT and compared with ground truth. (a) Original noisy blobs image. (b) Ground truth. (c) Denoised result of U-Net. (d) Denoised result of 3D CUT. (e) Denoised result of BlobCUT.
Figure 8. Denoising results of noisy synthetic blobs image using U-Net, 3D CUT, BlobCUT and compared with ground truth. (a) Original noisy blobs image. (b) Ground truth. (c) Denoised result of U-Net. (d) Denoised result of 3D CUT. (e) Denoised result of BlobCUT.
Bioengineering 10 01372 g008
Table 1. Comparison (AVG ± STD) of UHDoG, BTCAS, UVCGAN, EGSDE, BlobDetGAN, CUT, and BlobCUT on 3D synthetic images (* significance p < 0.05 ).
Table 1. Comparison (AVG ± STD) of UHDoG, BTCAS, UVCGAN, EGSDE, BlobDetGAN, CUT, and BlobCUT on 3D synthetic images (* significance p < 0.05 ).
MetricsBlobCUTUHDoGBTCASUVCGANEGSDEBlobDetGAN3D CUT
DER0.085 ± 0.035 0.166 ± 0.040 0.093 ± 0.061 0.117 ± 0.075 0.128 ± 0.058 0.081 ± 0.039 0.126 ± 0.052
(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001) ( 0.096 ) (* < 0.001)
Precision0.972 ± 0.0013 0.940 ± 0.040 0.802 ± 0.086 0.867 ± 0.026 0.901 ± 0.018 0.980 ± 0.013 0.954 ± 0.015
(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)
Recall0.896 ± 0.031 0.786 ± 0.029 0.988 ± 0.004 0.825 ± 0.016 0.812 ± 0.028 0.904 ± 0.034 0.736 ± 0.052
(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001) ( 0.786 ) (* < 0.001)
F-score0.932 ± 0.018 0.856 ± 0.026 0.885 ± 0.072 0.845 ± 0.021 0.854 ± 0.026 0.940 ± 0.039 0.830 ± 0.021
(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001) ( 0.062 ) (* < 0.001)
Dice0.816 ± 0.061 0.296 ± 0.005 0.348 ± 0.012 0.475 ± 0.011 0.458 ± 0.009 0.795 ± 0.005 0.587 ± 0.020
(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)
IoU0.694 ± 0.074 0.174 ± 0.003 0.210 ± 0.009 0.462 ± 0.013 0.429 ± 0.009 0.666 ± 0.004 0.548 ± 0.023
(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)
Blobness0.542 ± 0.291 0.550 ± 0.301 0.562 ± 0.300 0.560 ± 0.321 0.563 ± 0.315 0.548 ± 0.213 0.551 ± 0.299
(GT:0.519)(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)(* < 0.001)
Avg. Time1393\\\\21761246
(s/epoch)
Table 2. Mouse kidney glomerular segmentation ( N g l o m ) from CFE-MRI using UH-DoG, BTCAS, UVCGAN, EGSDE, BlobDetGAN and the proposed BlobCUT compared to HDoG with VBGMM method (difference % compared with results of HDoG with VBGMM). Results with bold font are the best.
Table 2. Mouse kidney glomerular segmentation ( N g l o m ) from CFE-MRI using UH-DoG, BTCAS, UVCGAN, EGSDE, BlobDetGAN and the proposed BlobCUT compared to HDoG with VBGMM method (difference % compared with results of HDoG with VBGMM). Results with bold font are the best.
Mouse KidneyHDoG with VBGMM (Ground Truth)BlobCUT (Diff. %)BlobDet-GAN (Diff. %)UH-DoG (Diff. %)BTCAS (Diff. %)UVCGAN (Diff. %)EGSDE
(Diff. %)
CKDID 42976567621 (0.46%)7633 (0.30%)7346 (4.05%)7719 (0.82%)7358 (3.89%)7295 (4.72%)
ID 46686658843 (2.05%)8912 (2.85%)8138 (6.08%)8228 (5.04%)8365 (3.46%)8317 (4.01%)
ID 46785498812 (3.08%)8802 (2.96%)8663 (1.33%)8595 (0.54%)8368 (2.11%)8128 (4.92%)
AVG82908425 (1.63%)8449 (1.92%)8049 (2.91%)8181 (2.13%)8030 (3.13%)7913 (4.54%)
Control for CKDID 42712,72412,573 (1.18%)12,683 (0.32%)12,701 (0.18%)12,008 (5.63%)12,486 (1.87%)12,423 (2.37%)
ID 46910,82910,897 (0.63%)10,921 (0.85%)11,347 (4.78%)11,048 (2.02%)10,604 (2.08%)10,458 (3.43%)
ID 47010,70410,579 (1.17%)10,774 (0.65%)11,309 (5.65%)10,969 (2.48%)10,281 (3.95%)10,299 (3.78%)
ID 47111,93412,488 (4.56%)12,692 (6.27%)12,279 (2.81%)12,058 (0.96%)11,685 (2.08%)11,718 (1.81%)
ID 47212,56912,590 (0.16%)12,786 (1.73%)12,526 (0.34%)13,418 (4.75%)11,952 (4.90%)12,152 (3.31%)
ID 47312,24512,058 (1.53%)12,058 (1.53%)11,853 (3.20%)12,318 (0.60%)12,025 (1.79%)11,825 (3.43%)
AVG11,83611,864 (0.24%)11,986 (1.27%)12,003 (1.41%)11,970 (3.07%)11,505 (2.77%)11,479 (2.99%)
AKIID 43311,04611256 (1.90%)11618 (5.18%)11,033 (0.12%)10,752 (2.66%)10,826 (1.99%)10,751 (2.67%)
ID 46211,29211,420 (1.13%)11,445 (1.35%)10,779 (4.54%)10,646 (5.75%)10,892 (3.54%)11,082 (1.86%)
ID 46311,54211533 (0.07%)11,544 (0.02%)10,873 (5.80%)11,820 (2.41%)11,058 (4.19%)11,282 (2.25%)
ID 46411,90611,704 (1.70%)11,562 (2.89%)11,340 (4.75%)11,422 (3.33%)11,526 (3.19%)11,290 (5.17%)
AVG11,44711,478 (0.27%)11,542 (0.83%)11,006 (3.85%)11,015 (3.78%)11,075 (3.24%)11,101 (3.01%)
Control for AKIID 46510,33610,482 (1.41%)10,214 (1.18%)10,115 (2.14%)10,393 (0.55%)10,205 (1.26%)10,059 (2.68%)
ID 47410,87410,928 (0.50%)10,955 (0.74%)11,157 (2.60%)11,034 (1.47%)10,989 (1.05%)10,428 (4.10%)
ID 47510,29210,196 (0.91%)10,222 (0.68%)10,132 (1.55%)9985 (2.98%)9982 (3.01%)9907 (3.74%)
ID 47610,95410,764 (1.71%)11452 (4.55%)10,892 (0.57%)11,567 (5.60%)10,735 (2.00%)10,429 (4.79%)
ID 47710,88511,047 (1.49%)10,929 (0.40%)11,335 (4.13%)11,143 (2.37%)10,789 (0.88%)10,216 (6.15%)
AVG10,66810,683 (0.14%)10,754 (0.81%)10,726 (0.54%)10,824 (2.59%)10,540 (1.20%)10,208 (4.32%)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, T.; Xu, Y.; Wu, T.; Charlton, J.R.; Bennett, K.M.; Al-Hindawi, F. BlobCUT: A Contrastive Learning Method to Support Small Blob Detection in Medical Imaging. Bioengineering 2023, 10, 1372. https://doi.org/10.3390/bioengineering10121372

AMA Style

Li T, Xu Y, Wu T, Charlton JR, Bennett KM, Al-Hindawi F. BlobCUT: A Contrastive Learning Method to Support Small Blob Detection in Medical Imaging. Bioengineering. 2023; 10(12):1372. https://doi.org/10.3390/bioengineering10121372

Chicago/Turabian Style

Li, Teng, Yanzhe Xu, Teresa Wu, Jennifer R. Charlton, Kevin M. Bennett, and Firas Al-Hindawi. 2023. "BlobCUT: A Contrastive Learning Method to Support Small Blob Detection in Medical Imaging" Bioengineering 10, no. 12: 1372. https://doi.org/10.3390/bioengineering10121372

APA Style

Li, T., Xu, Y., Wu, T., Charlton, J. R., Bennett, K. M., & Al-Hindawi, F. (2023). BlobCUT: A Contrastive Learning Method to Support Small Blob Detection in Medical Imaging. Bioengineering, 10(12), 1372. https://doi.org/10.3390/bioengineering10121372

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop