Image Processing and Analysis for Preclinical and Clinical Applications

and AUC. The results showed good agreement with patient outcomes, and since no relevant correlations were found between CSH and AUC and the

Preclinical and clinical imaging aims to characterize and measure biological processes and diseases in animals [1] and humans [2]. In recent years, there has been growing interest in the quantitative analysis of clinical images using techniques such as positron emission tomography (PET) [3], computerized tomography (CT) [4], and magnetic resonance imaging (MRI) [5], mainly applied to texture analysis and radiomics. Various image processing and analysis algorithms based on pattern recognition, artificial intelligence, and computer graphics methods have been proposed to extract features from biomedical images. These quantitative approaches are expected to have a positive clinical impact on quantitatively analyzing images, to reveal biological processes and diseases, and to predict response to treatment.
This Special Issue presents a collection of high-quality studies covering state-of-the-art and innovative approaches focusing on image processing and analysis across a variety of imaging modalities as well as the expected clinical applicability of these innovative approaches for personalized patient-tailored medicine.
The topics/keywords covered by this Special Issue includes the following: In response to the call for papers, nineteen papers were submitted to this Special Issue, of which fourteen were accepted for publication. These papers address several research challenges related to image processing and analysis in both preclinical and clinical applications.
Among the published research papers, five of them focus on segmentation and detection applications, including prostate gland segmentation [6,7], retroperitoneal sarcoma segmentation [8], basal cell carcinoma detection [9], and fracture detection in patients with maxillofacial trauma [10].
In one of these papers, the authors estimated prostate volume using ultrasound imaging, which offers many advantages such as portability, low cost, lack of ionizing radiations, and suitability for real-time operation [6]. Since experts usually consider automatic end-toend volume-estimation procedures as non-transparent and uninterpretable systems, the authors proposed a system that directly estimated the diameter parameters of the standard ellipsoid formula to produce the prostate volume in a dataset of 305 patients. The proposed system detects four diameter endpoints from the transverse images and two diameter endpoints from the sagittal images, as defined by the classical procedure. These endpoints are estimated using a new image-patch voting method to address characteristic problems of ultrasound images. Furthermore, the dataset included 75 MRI images of the initial 305 patients. The results showed optimal performance, confirming that this system can be used in clinical practice.
Another prostate gland segmentation method based on T2-weighted MRI was proposed by Comelli et al. [7]. The authors presented the efficient neural network (ENet) to tackle the fully automated, real-time, and 3D delineation process of the prostate. ENet is mainly applied in self-driving cars to compensate for limited hardware availability while still achieving accurate segmentation. The authors applied this network to a limited set of 85 manual prostate segmentations using the k-fold validation strategy and the Tversky loss function [11] and compared the results with UNet and ERFNet (efficient residual factorized convNet). The results showed that ENet and UNet were more accurate than ERFNet, with ENet much faster than UNet. Specifically, ENet obtains a dice similarity coefficient of 90.89% and a segmentation time of about 6 s using central processing unit (CPU) hardware to simulate real clinical conditions where the graphics processing unit (GPU) is not always available.
In a similar study, Salvaggio et al. [8] used ENet and ERFNet for the automatic segmentation of retroperitoneal sarcoma (RPS) in 94 CT examinations. The volume estimation of RPS is often difficult due to its huge dimensions and irregular shape; thus, it often requires manual segmentation, which is time-consuming and operator-dependent. For this reason, the authors assessed the existence of significant differences between manual segmentation performed by two radiologists and automatic segmentation based on ENet and ERFNet using analysis of variance (ANOVA). A set of performance indicators for the shape comparison were calculated, namely sensitivity, positive predictive value, dice similarity coefficient, volume overlap error, and volumetric differences. There were no significant differences found between the RPS volumes obtained using manual segmentation and deep learning methods. Furthermore, all performance indicators were optimal for both ENet and ERFNet. Finally, ENet took around 15 s for segmentation versus 13 s for ERFNet by using GPU. In the case of CPU, ENet took around 2 min versus 1 min for ERFNet. The manual approach required approximately one hour per segmentation. In conclusion, fully automatic deep learning networks were reliable methods for RPS volume assessment.
Vélez et al. [9] proposed a tool for the detection of basal cell carcinoma (BCC) to provide a prioritization in the tele-dermatology consultation. BCC is the most frequent skin cancer, and its increasing incidence is producing a high overload in dermatology services. The authors analyzed if pre-segmentation of the lesion improved the classification of the lesion. After that, they analyzed three deep neural networks to distinguish between BCC and nevus, or other skin lesions. The best segmentation results were obtained with SegNet with accuracies of 98% and 95% for distinguishing BCC from nevus and other skin lesions, respectively. This method outperformed the winner of the challenge International Skin Imaging Collaboration (ISIC) 2019. Furthermore, the authors concluded that when deep neural networks are used to classify, a pre-segmentation of the lesion does not improve the classification results.
Finally, a novel maxillofacial fracture detection system (MFDS), based on convolutional neural networks and transfer learning, was proposed by Amodeo et al. [10] to detect traumatic fractures in patients. A convolutional neural network pre-trained on non-medical images was re-trained and fine-tuned using 148 CT images to produce a model for the classification of future CTs as fracture or not fracture. The validation and test datasets were characterized by 30 patients: both datasets contained 5 patients without fractures and 25 with fractures. The results showed an accuracy of 80% in classifying the maxillofacial fractures. Consequently, the proposed model can be used as a care support, reducing the risk of human error, preventing patient harm by minimizing diagnostic delays, and reducing the incongruous burden of hospitalization.
Among the other research papers, three of them focus on radiomics applications, including restaging in metastatic colorectal cancer [12], evaluating the robustness of PET radiomics features after MRI co-registration [13], and predicting pathologic complete response after neoadjuvant chemoradiation therapy for rectal cancer [14].
Alongi et al. [12] investigated the application of [18F]FDG PET/CT image-based textural features analysis to early predict disease progression and survival outcome in 52 metastatic colorectal cancer (MCC) patients after first adjuvant therapy. For this purpose, radiomics features from PET and low-dose CT images were extracted. The hybrid descriptive-inferential method [15] was used for feature selection while the discriminant analysis [16] was used for the predictive model implementation. The prediction performance was evaluated for per-lesion analysis, per-patient analysis, and per liver lesions analysis. All results showed that the proposed radiomics model was feasible and potentially useful in the predictive evaluation of disease progression in MCC.
Stefano et al. [13] studied the variability in PET radiomics features under the impact of co-registration with MRI using the difference percentage coefficient and the Spearman's correlation coefficient for three groups of images: (i) original PET, (ii) PET after co-registration with T1-weighted MRI, and (iii) PET after co-registration with FLAIR MRI. For this purpose, 77 patients with brain cancers undergoing [11C]-Methionine PET were considered. Successively, PET images were co-registered with MRI sequences and 107 features were extracted for each mentioned group of images. The variability analysis revealed that shape features, first-order features, and two subgroups of higher-order features possessed a good robustness, unlike the remaining groups of features, which showed large differences in the difference percentage coefficient. Furthermore, using Spearman's correlation coefficient, approximately 40% of the selected features differed from the three mentioned groups of images. This is an important consideration for users conducting radiomics studies with image co-registration constraints to avoid errors in cancer diagnosis, prognosis, and clinical outcome prediction.
Lee et al. [14] evaluated the MRI assessment after neoadjuvant chemoradiotherapy (nCRT) in 912 patients with rectal cancer for staging and treatment planning purposes. They proposed a pathologic complete response (pCR) prediction method based on a novel multi-parametric MRI embedding technique. Specifically, multiple MRI sequences were encapsulated into multi-sequence fusion images (MSFI). Subsequently, radiomics features were extracted and used to predict pCR through a random forest classifier. The results demonstrated that the use of all given MRI sequences is the most effective method regardless of the dimension reduction method. Furthermore, it outperformed four competing baselines in terms of the area under the receiver operating characteristic curve (AUC) and F1-score.
Among the other research papers, four of them focus on biomedical image quantification, including the early monitoring response to therapy in patients with brain lesions [17], the quantification of cancer cell mass evolution in zebrafish [18], the clinical comparison of the glomerular filtration rate calculated from different renal depths and formulae [19], and the assessment of the left atrial flow stasis in patients undergoing pulmonary vein isolation for paroxysmal atrial fibrillation [20].
Stefano et al. [17] evaluated new PET prognostic indices for the early assessment of response to Gamma Knife (GK) treatment. GK is an alternative to traditional brain surgery and whole-brain radiation therapy for the treatment of tumors inaccessible through conventional treatments [21]. Semi-quantitative PET parameters currently used in the clinical setting can be affected by statistical fluctuation errors and/or cannot provide information on tumor extent and heterogeneity. To overcome these limitations, the cumulative standardized uptake value histogram (CSH) and AUC were considered as additional information on the response to GK treatment. Specifically, the absolute level of [11C]-Methionine (MET) uptake was measured and its heterogeneity distribution within PET lesions was evaluated by calculating the CSH and AUC. The results showed good agreement with patient outcomes, and since no relevant correlations were found between CSH and AUC and the indices usually used in PET imaging, these innovative parameters could be a useful tool for assessing patient responses to therapy.
In [18], the authors considered zebrafish as it is a model organism for the study of human cancer and, compared with the murine model, it has several properties that are ideal for personalized therapies. The transparency of the zebrafish embryos and the development of the pigment-deficient "casper" zebrafish line give the capacity to directly observe cancer formation and progression in the living animal. Nevertheless, the automatic quantification of cellular proliferation in vivo is still critical. For this reason, the authors proposed a new tool, namely ZFTool, to automatically quantify the cancer cellular evolution. ZFTool is capable of establishing a base threshold that eliminates the embryo autofluorescence, to automatically measure the area and intensity of green-fluorescent protein marked cells, and to define a proliferation index. As result, the proliferation index computed on different targets demonstrated the efficiency of ZFTool in providing a good automatic quantification of cancer mass evolution in zebrafish, eliminating the influence of its autofluorescence.
In the study proposed by Hsu et al. [19], the authors aimed to compare the differences in renal depths in a camera-based method using Technetium-99 m diethylenetriaminepentaacetic acid (Tc-99 m DTPA). This method is commonly used to calculate the glomerular filtration rate (GFR) as it can easily calculate split renal function. Renal depth is the main factor affecting the measurement of GFR accuracy. For this reason, the difference in renal depths between three formulae (Tonnesen's, Itoh K's, and Taylor's) and a CT scan were compared and used to calculate the GFRs using four methods. For this purpose, 51 patients underwent a laboratory test within one month and a CT scan within two months. The results showed that the renal depths measured using the three formulae were smaller than those measured using the CT scan, and the right renal depth was always larger than the left.
In [20], the authors aimed to demonstrate that left atrial (LA) stasis, derived from 4D-flow, is a useful biomarker of LA recovery in patients with atrial fibrillation (AF). AF is associated with systemic thrombo-embolism and stroke events, which do not appear significantly reduced following successful pulmonary vein (PV) ablation. The authors' hypothesis was that LA recovery was associated with a reduction in LA stasis. For this purpose, 148 subjects with paroxysmal AF and 24 controls were recruited and underwent a cardiac MRI, inclusive of 4D-flow. LA was isolated within the 4D-flow dataset to constrain stasis maps. The results showed that the mean LA stasis in the control was lower than that in the pre-ablation cohort and that the mean LA stasis was reduced in the post-ablation cohort compared with in the pre-ablation cohort. The study demonstrated that 4D flowderived LA stasis mapping was clinically relevant and revealed stasis changes in the LA body pre-and post-pulmonary vein ablation.
Finally, the last two published studies concern an image registration technique based on local feature of retinal vessels [22], and the hardware optimizations of the X-ray preprocessing using the field programmable gate array (FPGA) [23].
In the first of these two studies [22], an innovative method, namely CURVE, is presented to accurately extract feature points on retinal vessels and throughout the fundus image. The CURVE performance was tested on different datasets and compared with six state-of-the-art feature extraction methods. The results showed that the feature extraction accuracy of CURVE significantly outperformed the existing feature extraction methods. Then, CURVE was paired with a scale-invariant feature transform (SIFT) descriptor to test its registration capability on the fundus image registration (FIRE) dataset. CURVE-SIFT successfully registered 44% of the image pairs while existing feature-based techniques registered less than 27% of the image pairs. The last study [23] proposed the optimization of the X-ray pre-processing in CT imaging to compute total attenuation projections by avoiding the intermediate step of converting detector data to intensity images. Furthermore, a configurable hardware architecture for data acquisition systems on FPGAs was proposed to fulfill the real-time requirements and with the aim of achieving "on-the-fly" pre-processing of 2D projections. Finally, this architecture was configured for exploring and analyzing different arithmetic representations, such as floating-point and fixed-point data formats. In this way, the best representation and data format that minimized execution time and hardware costs was found without affecting image quality. By comparing the proposed solution with the state-of-the-art pre-processing algorithm, the latency decreased by 4.125× and the resource utilization decreased by ∼6.5×. By using fixed-point representation in the different data precisions, the latency and the resource utilization were further decreased.
In conclusion, this Special Issue covers recent trends in biomedical imaging applications, such as quantification, detection, radiomics, registration, and optimization, constituting a good sample of the current state-of-the-art results in this field.