Next Article in Journal
Immunohistological Examination of HEATR1 and SLC27A2 Expression in ccRCC Samples to Evaluate Their Potential as Prognostic Markers—A Preliminary Study
Previous Article in Journal
Discordance Between Radiological and Pathological Responses to Pembrolizumab in Mismatch Repair-Deficient Metastatic Colorectal Cancer: Implications for Precision Oncology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Artificial Intelligence and Rectal Cancer: Beyond Images

by
Tommaso Novellino
1,*,
Carlotta Masciocchi
2,
Andrada Mihaela Tudor
1,
Calogero Casà
1,3,
Giuditta Chiloiro
2,
Angela Romano
1,2,
Andrea Damiani
2,
Giovanni Arcuri
2,
Maria Antonietta Gambacorta
1,2 and
Vincenzo Valentini
3
1
Department of Medicine and Surgery, Università Cattolica del Sacro Cuore, 00168 Rome, Italy
2
Centro di Medicina dell’Invecchiamento, Fondazione Policlinico Universitario Agostino Gemelli—IRCCS, 00168 Rome, Italy
3
Ospedale Isola Tiberina—Gemelli Isola, 00186 Rome, Italy
*
Author to whom correspondence should be addressed.
Cancers 2025, 17(13), 2235; https://doi.org/10.3390/cancers17132235
Submission received: 21 May 2025 / Revised: 25 June 2025 / Accepted: 27 June 2025 / Published: 3 July 2025

Simple Summary

The cancer burden, particularly in rectal cases, can be alleviated through the use of artificial intelligence models, provided they are properly designed, implemented, and validated. Artificial intelligence encompasses machine learning, which in turn includes deep learning. Artificially intelligent models can be developed based on various types of data, including images, numerical values, and texts. We believe there is currently considerable hype around image-based models, and that more intensive exploration of other data types—such as electronic health records and omics—could greatly enhance both research and clinical practice. By analyzing the literature, we confirm this idea and offer some recommendations that we ultimately consider beneficial for patients, especially by promoting multimodal approaches beyond simply imaging.

Abstract

Introduction: The variability of cancers and medical big data can be addressed using artificial intelligence techniques. Artificial intelligence models can accept different input types, including images as well as other formats such as numerical data, predefined categories, and free text. Non-image sources are as important as images in clinical practice and the literature; nevertheless, the secondary literature tends to focus exclusively on image-based inputs. This article reviews such models, using non-image components as a use case in the context of rectal cancer. Methods: A literature search was conducted using PubMed and Scopus, without temporal limits and in English; for the secondary literature, appropriate filters were employed. Results and Discussion: We classified artificial intelligence models into three categories: image (image-based input), non-image (non-image input), and combined (hybrid input) models. Non-image models performed significantly well, supporting our hypothesis that disproportionate attention has been given to image-based models. Combined models frequently outperform their unimodal counterparts, in agreement with the literature. However, multicenter and externally validated studies assessing both non-image and combined models remain under-represented. Conclusions: To the best of our knowledge, no previous reviews have focused on non-image inputs, either alone or in combination with images. Non-image components require substantial attention in both research and clinical practice. The importance of multimodality—extending beyond images—is particularly relevant in the context of rectal cancer and potentially other pathologies.

1. Introduction

Rectal cancer is a leading cause of morbidity and mortality worldwide. It ranks as the cancer type with the eighth highest incidence (≈700,000 new cases per year) and has significant mortality (≈300,000 deaths per year) [1].
Cancer remains a fundamental burden on human society, despite significant technological and clinical progress. Globally, it is the second leading cause of death [2], and its prevalence is increasing [1]. One of the main reasons for this persistent burden is its heterogeneity. This variability spans multiple levels: the disease level (100+ cancer types), the inter-patient level (each patient is unique, as per their history and current condition), and the intra-patient level (cancer evolves dynamically, particularly at the genomic level). This daunting complexity hinders both its biological understanding and clinical management (prevention, diagnosis, therapy, and prognosis).
Fortunately, a promising paradigmatic shift is underway in medicine, particularly in oncology, as evidence-based medicine is gradually being complemented by personalized medicine [3]. Evidence-based medicine, which is more traditional and well-established through clinical trials, has been fundamental in the development of medical knowledge and has shaped the standard of care over the past three decades [3,4]. Nonetheless, this methodology generates evidence based on an ‘average patient’—a ‘one size fits all’ approach—making its application to actual, unique patients problematic [3,5]. Personalized medicine, with its more modern, precise, and individualized perspective, offers a valuable complement to evidence-based medicine, particularly in oncology. Certain enablers of personalized medicine—such as big data (particularly real-world data) and artificial intelligence (particularly machine learning)—hold great promise [3].
Big data refer to digital information generated in enormous amounts (≈1012 terabyte/year, with a growing trend). To qualify as big data, such data must not only have high ‘volume’ (quantity) but also high ‘variety’ (multiple variables and sources), ‘velocity’ (acquisition speed), and ‘veracity’ (quality-related parameter). Medical big data have significant potential to improve both research and clinical practice [6], but they must be processed using proper mathematical and computational methodologies. In particular, artificial intelligence-based models—when rigorously designed, developed, and validated [7]—can successfully manage these data [8].
Artificial intelligence imitates human intelligence via machine computation [9]. An artificial intelligence model represents a mathematical relationship between one or more inputs (independent variables) and one or more outputs (dependent, target variables). Inputs can be categorized as structured or unstructured. Structured data—including numbers and predefined categories—are more standardized and generally of higher quality. Unstructured data—including images and uncategorized (freely typed) text—usually need to be mined (e.g., radiomics for images or natural language processing for texts). In particular, radiomics is an approach that aims to extract structured, quantitative information (often consisting of several hundreds of predictors) from unstructured medical images [10].
Machine learning is a notable and intensively researched subset of artificial intelligence. It mimics human learning via computer algorithms [11], enabling models to ‘learn’ features and patterns directly from input data without being explicitly programmed. Thus, a model’s structure and performance are strongly influenced by the data used for input (training) and validation (evaluation). Generally, increasing the quantity of these data (i.e., using larger sample sizes) provides more reliable results. As in statistics, this demonstrates the advantage of big data over ‘traditional’ datasets. However, big data must also meet sufficient quality standards to produce reliable and realistic results. Therefore, data collection must be performed in a curated and standardized manner [10,12].
A noteworthy subset of machine learning is deep learning, which employs artificial neural networks with more than one hidden layer. These networks are inspired by the human brain: they comprise relatively simple units (‘artificial neurons’, modeled after biological neurons) connected in a complex fashion (Figure 1).
Because of their architectural connectivity, deep learning models can accomplish complex tasks, such as classification and generation. In medicine, they are mainly applied to images.
Medical images, such as imaging, histology, and endoscopy, are fundamental for both diagnostic and therapeutic purposes [13]. However, images are not the only type of data that support medical decision making: structured data, as well as unstructured data other than images, are also essential [6]. Real-world data [14,15]—defined as all data generated outside clinical trials [3,16]—are an important source for evidence generation. In particular, electronic health records (or electronic medical records)—including medical reports and nurse diaries—can be a rich source of patient information [17]. Hence, medical big data are intrinsically multimodal; i.e., they encompass heterogeneous types and come from various sources.
Multimodality is also a key feature of oncological care, particularly for rectal cancer [3,5]. Multimodal integration can benefit the whole medical workflow in both clinical practice and research, including aspects such as disease understanding, prevention, diagnosis, therapy, prognosis, and follow-up. The integration of different roles—including but not limited to physicians, nurses, biomedical technicians, and data scientists—is of paramount importance for advancing knowledge and improving healthcare. Medicine is a multi-source, synthetic, and holistic process that cannot be reduced to images only.
Nonetheless, to the best of our knowledge, all reviews on artificial intelligence applied to rectal cancer have focused on models whose inputs are based solely on images (herein termed ‘image models’). This seems to contradict not only the abovementioned medical holism, but also the literature. Indeed, models with only non-image input, which we call ‘non-image models,’ can show high performance—for example, in pan-cancer applications based entirely on real-world data/electronic health records [18], or in rectal cancer specifically, as presented in Section 3. Moreover, there are also ‘multimodal’ models which accept multiple input-types; i.e., both image and non-image sources. Such models (herein named ‘combined models’; also known in the literature as ‘hybrid models’ or ‘holistic models’) can outperform their unimodal constituent models [6,10,19], further highlighting the importance of multimodality.
While three of the identified reviews [10,12,20] have briefly mentioned the existence of combined models, all reviews found [10,12,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36] focused solely on image models. As non-image inputs—whether integrated (‘combined models’) or standalone (‘non-image models’) with images—are as essential as image inputs, a review of the existing primary literature on them is needed.
This review focuses on original articles reporting artificial intelligence models (particularly machine learning algorithms) that include some non-image input and are applied to rectal cancer, without specific restrictions.

2. Methods

Articles on artificial intelligence applied to rectal cancer were searched on PubMed and Scopus, with no limitations on publication date. Only articles written in the English language were considered, using the following search string: “rectal cancer” AND (“artificial intelligence” OR “machine learning” OR “deep learning” OR “predictive model” OR “nomogram” OR “general linear model” OR “logistic regression” OR “Cox regression” OR “decision tree” OR “random forest” OR “support vector machine” OR “artificial neural network”). In PubMed, secondary-literature articles were filtered according to the following three literature types: reviews, systematic reviews, and meta-analyses.
There were no strict eligibility criteria—that is, no restrictions regarding study design or outcomes, patient age or genre, or cohort size.

3. Results and Discussion

Table 1 and Table 2 present the primary literature (N = 23 articles, for a total of 11,941 patients) on artificial intelligence applied to rectal cancer, addressing different outcomes, and including at least one non-image input (i.e., either non-image or combined models).
Table 1 contains general information, while Table 2 provides more in-depth analysis.
As shown in Table 1, most articles address the predictive response to neoadjuvant therapy (neoadjuvant chemoradiotherapy, part of the standard treatment for locally advanced rectal cancer). Additional purposes include predicting metastases and other rectal cancer features. Specifically, response prediction is analyzed in 11 articles (47.8%); prediction of clinical outcomes (overall survival, local recurrence, and distant metastases) and risk factors (KRAS mutation, tumor deposit, peri-neural invasion, extramural venous invasion) is the focus of 4 articles (17.4%) in each subset; prediction of lymph node staging in 3 articles (13.0%); and histological examination prediction in 1 article (4.4%).
Sample sizes ranged from hundreds to thousands of patients per study.
Classification according to the TRIPOD (Transparent Reporting of a Multivariable Prediction model for Individual Prognosis or Diagnosis) statement [60] is reported. Generally, based on this classification, artificial intelligence-based prediction models can be regrouped in the following manner: 1a training only; 1b training plus validation via resampling (cross validation, bootstrapping, …); 2a training plus validation via random split (between training and validation data sets); 2b training plus validation via non-random split (e.g., based on the temporal sequence of patients); 3 training plus external validation (i.e., training and validation on different populations, centers, institutions, or even countries and continents); 4 external validation only (i.e., validation of an already published prediction model).
Most of the identified artificial intelligence models for rectal cancer belong to the combined model category (N = 20), compared to non-image models (N = 3). The variability among selected artificial intelligence models—(i.e., the reason so many models continue to be used rather than one optimal, ideal model)—is linked to the ‘no free lunch theorem’ [61]. This fundamental theorem holds (in its original form or in analogous formulations) for many artificial intelligence tools (supervised models, unsupervised models, feature selection techniques, etc.), and declares that no tool is absolutely the best across all scenarios, contexts, and datasets. This helps to explain the variety of methodologies found. The most common type of model found (69.6%), including both non-image and combined models, is regression (either logistic or Cox). Specifically, most models (52.2%) belong to the logistic regression type, likely due to its interpretability—model coefficients directly reflect predictor importance. When models are ranked by decreasing patient count, the first three models are all Cox regression type. Another notable technique is the support vector machine (8.7%).
Table 2. Details about studies listed in Table 1.
Table 2. Details about studies listed in Table 1.
ReferenceFirst AuthorYearModel Input(s)PerformanceNumber of CentersExternal
Validation(s)?
Note
Non-Images?Images?CM Better?
NIMs
[37]Peng J.2014Yes (demographic and clinicopathological) C-Index = 0.73–0.761No
[38]Sun Y.2017Yes (clinicopathological) C-Index = 0.711No
[39]Valentini V.2011Yes C-Index = 0.68–0.735YesSubsequently re-validated by Reference [62]
CMs
[40]Chen L.D.2020Yes (clinical)Yes (radiomics, ANN, US)YesAUC = 0.801No
[41]Cheng Y.2021YesYes (radiomics, MRI)YesAUC = 0.91–0.941No
[42]Cui Y.2019Yes (from EHRs)Yes (radiomics, MRI)YesAUC = 0.971No
[43]Dinapoli N.2018Yes (clinical)Yes (radiomics, MRI) AUC = 0.753YesSubsequently re-validated by Reference [63]
[44]Ding L.2020YesYes (DL, MRI) AUC = 0.89–0.921No
[45]Huang Y.Q.2016Yes (clinicopathological)Yes (radiomics, CT) C-Index = 0.781No
[46]Jin C.2021Yes (blood tumor markers)Yes (DL, MRI)YesAUC = 0.973Yes
[47]Kleppe A.2022Yes (markers)Yes (DL, histopathology)YesHR = 3.06–10.713Yes
[48]Li M.2021Yes (clinical)Yes (radiomics, CT)YesAUC = 0.801No
[49]Liu H.2022Yes (clinical)Yes (DL, MRI)YesAUC = 0.841No
[50]Liu S.2021Yes (clinical)Yes (radiomics, MRI)YesAUC = 0.861No
[51]Liu X.2021Yes (clinicopathological)Yes (radiomics, DL, MRI)YesC-Index = 0.783Yes
[52]Mao Y.2022Yes (clinicopathological)Yes (radiomics, CT)YesAUC = 0.871No
[53]Peterson K.J.2023Yes (clinical [from EHRs])Yes (radiomics, MRI)YesAUC = 0.731No
[54]van Stiphout R.G.P.M.2011Yes (clinical)Yes (PET-CT)YesAUC = 0.864Yes
[55]van Stiphout R.G.P.M.2014Yes (clinical)Yes (PET-CT) AUC = 0.702Yes
[56]Wan L.2019Yes (clinical)Yes (MRI)YesAUC = 0.841No
[57]Wei Q.2023Yes (clinical)Yes (radiomics, MRI)YesAUC = 0.872Yes
[58]Wei Q.2023Yes (clinical)Yes (radiomics, MRI)YesAUC = 0.851No
[59]Yi X.2019Yes (radiological-clinicopathological)Yes (radiomics, MRI) AUC = 0.90–0.931No
Legend. AUC = Area under the ROC curve; C-Index = Concordance index; CM = Combined model; CT = Computerized tomography; EHR = Electronic health record; HR = Hazard ratio; MRI = Magnetic resonance imaging; NIM = Non-Image model; PET = Positron emission tomography; ROC = Receiver operating characteristic; US = Ultrasound.
Table 2 provides additional information about the articles listed in Table 1; particularly details about the artificial intelligence models in terms of their input(s) and validation.
For those models that include an image component (combined models), the most common sources of images are imaging techniques—mainly magnetic resonance imaging (65.0%) and computerized tomography (25.0%). An intensively employed methodology for handling images is radiomics; as mentioned above, deep learning is also frequently applied to images.
Performance is expressed through either the area under the receiver operating characteristic curve, concordance index (a generalization of the previous quantity [39]), or hazard ratio. Both non-image and combined models show relatively good performance, with ranges of area under the curve and concordance index of 0.70–0.97 and 0.68–0.78, respectively. Interestingly, in combined models, integrated models (e.g., deep learning plus pathological staging markers [47]) very often (75.0%) outperform their individual components. Both these aspects suggest that multimodality is highly desirable, in agreement with the literature [64]. This applies more generally, even for images alone; i.e., without any non-image input, such as when multiple imaging modalities are combined together within the same model [65]. We believe that multimodality is important and often leads to superior performance because each input source type captures different aspects of the same phenomenon (rectal cancer). Integrating multiple sources into a single computational framework provides a deeper, more holistic, and more complete view of the reality in oncology and possibly other contexts. While challenges exist in integrating data of different types (e.g., heterogeneity), we believe that the implications for clinical practice and the translational potential can be significant. Areas for future research include increased simultaneous use of multiple image sources or, more generally, multiple data sources (multimodality), as well as data fusion.
Stratifying the primary articles by outcome type—short-term (e.g., pathological complete response) or long-term (e.g., 5-year survival)—yields the following: short-term results show a concordance index range of 0.71–0.78 and an area under the curve of 0.70–0.97; long-term results show a concordance index range of 0.68–0.76.
More details about combined models and their performance are shown in Table 3.
When we calculated the incremental gain (i.e., the difference between the combined model and each of its unimodal components) in terms of area under the curve, we obtained differences ranging from 0.02 to 0.18, as reported in references [46,57], respectively. While the minimum difference has little effect on results, the maximum (≈0.2 difference in area under the curve, ranging from 0.5 to 1) becomes significant and may be clinically meaningful. Hence, integrating different input modalities (i.e., employing combined models) can offer clinical advantages.
Some notes about how performance is reported: if the model is externally validated, the performance refers to that external validation; otherwise, it reflects internal validation. If multiple models are presented for the same outcome, the best-performing model is presented. If a study investigated more than one outcome, performance is reported as a range.
It is worth noting that too many studies (65.2%) remain monocentric (with a median of 1.0 center [interquartile range 1.5]). Multicenter projects tend to be superior, providing better input data, as per both quantity (number of patients, including pooled datasets [39]) and quality (greater diversity and heterogeneity of populations, which reduces overfitting and promotes generalizability), leading to improved model performance and clinical applicability. Although externally validated designs are increasingly common compared to a decade ago, their presence should continue to rise for the reasons previously stated. To summarize, more multicenter, externally validated, and prospective studies are needed [66].
To explain the observed variability, a more in-depth analysis of performance and validation was conducted. Variation in model performance across studies can generally be attributed to different study designs, conditions, and aims; more specifically, they may relate to the type of validation. Notably, externally validated studies, while less in number, involve more patients than internally validated ones: the former included 7239 patients (60.6%), while the latter involved 4702 (39.4%). On average, externally validated studies enroll three times more patients per study. As is well known, larger sample sizes tend to correlate with higher model quality. However, in terms of discriminative performance, there appears to be no substantial difference between internal and external validation, based on both area under the curve and concordance index (Table 4).
This can be explained in terms of validation type. Indeed, the external validation setting of a model allows for increased generalizability, as the model is exposed to a population significantly different from the one it was trained on. However, this typically comes at the cost of performance, which tends to be lower than in internal validation, where—by design—the validation dataset is intrinsically more similar to the training one. The latter is widely considered more limited and less robust compared to the external validation design, as it tends to result in higher overfitting (optimism) and lower generalizability, potentially reducing clinical relevance and applicability.
Regarding data preprocessing, most articles (82.6%) have reported at least one methodology.
The most frequently reported approach (60.9%) is feature selection, which seems reasonable given the well-known importance of the ‘dimensionality curse’. The concept refers to the need to empirically find an optimal balance between underfitting—where the proposed model is too simple to accurately describe the data and produce reliable predictions—and overfitting, where the model is overly complex, replicating the training data patterns but failing to properly generalize them to new, unseen data. This trade-off is reached by choosing an appropriate number of model predictors, which should correspond to the number of data items (e.g., quantities such as events per variable).
Other reported methods include data normalization and standardization (39.1%), imputation (13.0%), augmentation (4.3%), and binning (4.3%). Data normalization involves mathematically ‘compacting’ a variable’s data values to the [0, 1] interval, while data standardization transforms them into a standard Gaussian distribution (with mean 0 and standard deviation 1). Data imputation is a set of techniques for handling missing values, typically by estimating them based on various mathematical methods—the simplest being substitution of a missing value with the feature’s median or mean. Data augmentation refers to the synthetic generation of data, which can be particularly useful in cases of class imbalance. In binary outcome scenarios, this means that the proportion of data items with a certain outcome value (label) is significantly lower compared to that with the other, complementary label.
A detailed overview of preprocessing methodologies across the various studies is presented in Table 5.
From the previous table, it can be seen that the most recurring methods fall into two main categories. For feature selection, these include the least absolute shrinkage and selection operator and the inter-/intra-class correlation coefficient. For normalization/standardization, the most common approach is z-score transformation (which consists of subtracting the mean and then dividing all data points by the standard deviation, reaching a null mean and unitary standard deviation). Instead, data imputation and augmentation appear to be scarcely performed or reported.
As previously mentioned, the rationale for exploring and analyzing the primary articles presented above is that all (N = 19) secondary literature papers [10,12,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36] focused solely on image models. At most, a few briefly mentioned the existence of the combined models, but none reported on non-image models. Thus, a review of artificial intelligence models whose input has at least one non-image element is warranted. This provides the motivation for the present study. Further details on the evaluation of the review status are summarized in Table 6. Notably, almost half of the identified reviews (42.1%) address radiomics.
This review has several limitations. First, as a narrative review, it does not include all articles relevant to the topic; however, this applies only to the primary literature, as all identified secondary sources (i.e., reviews) are presented. Second, our classification (non-image, combined, and image models) is shown to be helpful but is limited to a qualitative evaluation; a quantitative approach (e.g., intergroup comparison) could provide additional information. Third, a specific appraisal of primary studies in terms of their risk of bias (e.g., PROBAST [67]) was not performed.
In order to address these limitations, as a prosecution of this review, we are currently working on a more strict and focused methodology—namely, a systematic review targeting a narrowly defined outcome, still within the context of rectal cancer disease. By including all identified studies on the target topic and subsequently performing a meta-analysis of their results, this approach may allow for a more precise evaluation of performance and comparison among the groups (i.e., non-image, combined, and image categories).
Regarding challenges for clinical translation, two aspects can be noted. First, data heterogeneity, which can be seen both as a challenge and an opportunity: while technical difficulties certainly exist in integrating multiple types of data (e.g., tabular data and images)—as in data fusion [68]—heterogeneous inputs may improve model performance and generalizability. Second, another critical challenge is missing data. As it is well known, medical datasets lack a significant number of data points across different features. Besides imputation—which, as previously discussed, is rarely performed or reported—another method to mitigate missingness is data integration. While a single data source is rarely complete, several, interconnected sources are more likely to generate an improved, holistic picture of medical reality.

4. Conclusions

First, this work introduced a new, qualitative classification of artificially intelligent models, based on the type of input (non-image, image, or combined models). Second, it highlighted that no existing review focuses on models applied to rectal cancer that utilize non-image inputs. Third, it demonstrated—consistent with oncological clinical practice and the existing literature—that non-image models, whether used alone or in combination with image inputs (combined models), can reliably characterize medical datasets and predict clinical outcomes, performing comparably to image-based models. Fourth, combined models usually exhibit superior performance compared to their individual components. Together, these aspects highlight the importance of the synergic, holistic use of multimodal data sources to more comprehensively describe reality (including complex conditions such as human cancer) and to find actionable features. Just as radiomics revolutionized image-based models, we envision that emerging techniques and data sources (e.g., the -omics fields [genomics, proteomics, fragmentomics, etc.] and natural language processing of electronic health records) will play an equally important role for non-image and combined models.
From a clinical perspective, rectal cancer is well suited as a pathological model to be employed with artificial intelligence methodologies: it necessitates multimodal therapies (radiotherapy, chemotherapy, surgery are all involved); offers opportunities to intensify therapies based on risk (e.g., whether to pursue adjuvant chemotherapy, demolitive surgery, or radiotherapy dose escalation); and has good healing possibility, creating a need to identify high-risk patients to be cured or those with low risk to avoid unnecessary toxicity. For these reasons, the existing literature has chosen to investigate key clinical questions through artificial intelligence techniques, such as outcome stratification, prediction of histological risk factors, lymph node staging, and treatment response prediction.
The limitations of this review include its narrative and qualitative nature, language restrictions, and the lack of a formal risk-of-bias assessment.
To conclude, the application of artificial intelligence to rectal cancer requires integration across different domains: multiple data types (image and non-image inputs), medical procedures (neoadjuvant chemoradiotherapy and surgery), and complementary knowledge branches (multidisciplinary clinical boards and research teams). More generally, this integrative approach may also benefit other diseases, both oncological and non-oncological. In today’s increasingly patient-centric landscape, it is ultimately beneficial to incorporate data sources and features beyond images.

Author Contributions

Conceptualization, T.N., C.M. and V.V.; methodology, T.N., C.M., A.M.T., C.C. and G.C.; validation, V.V., M.A.G., A.D., A.R. and G.A.; formal analysis, T.N. and A.M.T.; investigation, T.N.; resources, A.D. and G.A.; data curation, T.N.; writing—original draft preparation, T.N. and C.M.; writing—review and editing, T.N., V.V., M.A.G., G.C., A.M.T. and C.C.; visualization, T.N.; supervision, V.V., M.A.G. and A.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef]
  2. Vos, T.; Lim, S.S.; Abbafati, C.; Abbas, K.M.; Abbasi, M.; Abbasifard, M.; Abbasi-Kangevari, M.; Abbastabar, H.; Abd-Allah, F.; Abdelalim, A.; et al. Global Burden of 369 Diseases and Injuries in 204 Countries and Territories, 1990–2019: A Systematic Analysis for the Global Burden of Disease Study 2019. Lancet 2020, 396, 1204–1222. [Google Scholar] [CrossRef] [PubMed]
  3. Li, A.; Bergan, R.C. Clinical Trial Design: Past, Present, and Future in the Context of Big Data and Precision Medicine. Cancer 2020, 126, 4838–4846. [Google Scholar] [CrossRef] [PubMed]
  4. Damiani, A.; Onder, G.; Valentini, V. Large Databases (Big Data) and Evidence-Based Medicine. Eur. J. Intern. Med. 2018, 53, 1–2. [Google Scholar] [CrossRef]
  5. De Maria Marchiano, R.; Di Sante, G.; Piro, G.; Carbone, C.; Tortora, G.; Boldrini, L.; Pietragalla, A.; Daniele, G.; Tredicine, M.; Cesario, A.; et al. Translational Research in the Era of Precision Medicine: Where We Are and Where We Will Go. J. Pers. Med. 2021, 11, 216. [Google Scholar] [CrossRef] [PubMed]
  6. Acosta, J.N.; Falcone, G.J.; Rajpurkar, P.; Topol, E.J. Multimodal Biomedical AI. Nat. Med. 2022, 28, 1773–1784. [Google Scholar] [CrossRef]
  7. Lambin, P.; van Stiphout, R.G.P.M.; Starmans, M.H.W.; Rios-Velazquez, E.; Nalbantov, G.; Aerts, H.J.W.L.; Roelofs, E.; van Elmpt, W.; Boutros, P.C.; Granone, P.; et al. Predicting Outcomes in Radiation Oncology—Multifactorial Decision Support Systems. Nat. Rev. Clin. Oncol. 2012, 10, 27–40. [Google Scholar] [CrossRef]
  8. Elmore, J.G.; Lee, C.I. Data Quality, Data Sharing, and Moving Artificial Intelligence Forward. JAMA Netw. Open 2021, 4, e2119345. [Google Scholar] [CrossRef]
  9. Torkamani, A.; Andersen, K.G.; Steinhubl, S.R.; Topol, E.J. High-Definition Medicine. Cell 2017, 170, 828–843. [Google Scholar] [CrossRef]
  10. Cuocolo, R.; Caruso, M.; Perillo, T.; Ugga, L.; Petretta, M. Machine Learning in Oncology: A Clinical Appraisal. Cancer Lett. 2020, 481, 55–62. [Google Scholar] [CrossRef]
  11. Lambin, P.; Leijenaar, R.T.H.; Deist, T.M.; Peerlings, J.; de Jong, E.E.C.; van Timmeren, J.; Sanduleanu, S.; Larue, R.T.H.M.; Even, A.J.G.; Jochems, A.; et al. Radiomics: The Bridge between Medical Imaging and Personalized Medicine. Nat. Rev. Clin. Oncol. 2017, 14, 749–762. [Google Scholar] [CrossRef]
  12. Meldolesi, E.; van Soest, J.; Damiani, A.; Dekker, A.; Alitto, A.R.; Campitelli, M.; Dinapoli, N.; Gatta, R.; Gambacorta, M.A.; Lanzotti, V.; et al. Standardized Data Collection to Build Prediction Models in Oncology: A Prototype for Rectal Cancer. Future Oncol. 2016, 12, 119–136. [Google Scholar] [CrossRef] [PubMed]
  13. Cusumano, D.; Meijer, G.; Lenkowicz, J.; Chiloiro, G.; Boldrini, L.; Masciocchi, C.; Dinapoli, N.; Gatta, R.; Casà, C.; Damiani, A.; et al. A Field Strength Independent MR Radiomics Model to Predict Pathological Complete Response in Locally Advanced Rectal Cancer. Radiol. Med. 2021, 126, 421–429. [Google Scholar] [CrossRef] [PubMed]
  14. Valentini, V.; Alfieri, S.; Coco, C.; D’Ugo, D.; Crucitti, A.; Pacelli, F.; Persiani, R.; Sofo, L.; Picciocchi, A.; Doglietto, G.B.; et al. Four Steps in the Evolution of Rectal Cancer Managements through 40 Years of Clinical Practice: Pioneering, Standardization, Challenges and Personalization. Radiother. Oncol. 2024, 194, 110190. [Google Scholar] [CrossRef] [PubMed]
  15. Savino, M.; Chiloiro, G.; Masciocchi, C.; Capocchiano, N.D.; Lenkowicz, J.; Gottardelli, B.; Gambacorta, M.A.; Valentini, V.; Damiani, A. A Process Mining Approach for Clinical Guidelines Compliance: Real-World Application in Rectal Cancer. Front. Oncol. 2023, 13, 1090076. [Google Scholar] [CrossRef]
  16. Ramagopalan, S.V.; Simpson, A.; Sammon, C. Can Real-World Data Really Replace Randomised Clinical Trials? BMC Med. 2020, 18, 13. [Google Scholar] [CrossRef]
  17. Penberthy, L.T.; Rivera, D.R.; Lund, J.L.; Bruno, M.A.; Meyer, A. An Overview of Real-World Data Sources for Oncology and Considerations for Research. CA Cancer J. Clin. 2021, 72, 287–300. [Google Scholar] [CrossRef]
  18. Morin, O.; Vallières, M.; Braunstein, S.; Ginart, J.B.; Upadhaya, T.; Woodruff, H.C.; Zwanenburg, A.; Chatterjee, A.; Villanueva-Meyer, J.E.; Valdes, G.; et al. An Artificial Intelligence Framework Integrating Longitudinal Electronic Health Records with Real-World Data Enables Continuous Pan-Cancer Prognostication. Nat. Cancer 2021, 2, 709–722. [Google Scholar] [CrossRef]
  19. Bi, W.L.; Hosny, A.; Schabath, M.B.; Giger, M.L.; Birkbak, N.J.; Mehrtash, A.; Allison, T.; Arnaout, O.; Abbosh, C.; Dunn, I.F.; et al. Artificial Intelligence in Cancer Imaging: Clinical Challenges and Applications. CA Cancer J. Clin. 2019, 69, 127–157. [Google Scholar] [CrossRef]
  20. Jia, L.-L.; Zheng, Q.-Y.; Tian, J.-H.; He, D.-L.; Zhao, J.-X.; Zhao, L.-P.; Huang, G. Artificial Intelligence with Magnetic Resonance Imaging for Prediction of Pathological Complete Response to Neoadjuvant Chemoradiotherapy in Rectal Cancer: A Systematic Review and Meta-Analysis. Front. Oncol. 2022, 12, 1026216. [Google Scholar] [CrossRef]
  21. Bedrikovetski, S.; Dudi-Venkata, N.N.; Kroon, H.M.; Seow, W.; Vather, R.; Carneiro, G.; Moore, J.W.; Sammour, T. Artificial Intelligence for Pre-Operative Lymph Node Staging in Colorectal Cancer: A Systematic Review and Meta-Analysis. BMC Cancer 2021, 21, 1058. [Google Scholar] [CrossRef]
  22. Coppola, F.; Giannini, V.; Gabelloni, M.; Panic, J.; Defeudis, A.; Lo Monaco, S.; Cattabriga, A.; Cocozza, M.A.; Pastore, L.V.; Polici, M.; et al. Radiomics and Magnetic Resonance Imaging of Rectal Cancer: From Engineering to Clinical Practice. Diagnostics 2021, 11, 756. [Google Scholar] [CrossRef]
  23. Kalantar, R.; Lin, G.; Winfield, J.M.; Messiou, C.; Lalondrelle, S.; Blackledge, M.D.; Koh, D.-M. Automatic Segmentation of Pelvic Cancers Using Deep Learning: State-of-The-Art Approaches and Challenges. Diagnostics 2021, 11, 1964. [Google Scholar] [CrossRef]
  24. Kuntz, S.; Krieghoff-Henning, E.; Kather, J.N.; Jutzi, T.; Höhn, J.; Kiehl, L.; Hekler, A.; Alwers, E.; von Kalle, C.; Fröhling, S.; et al. Gastrointestinal Cancer Classification and Prognostication from Histology Using Deep Learning: Systematic Review. Eur. J. Cancer 2021, 155, 200–215. [Google Scholar] [CrossRef]
  25. Kwok, H.C.; Charbel, C.; Danilova, S.; Miranda, J.; Gangai, N.; Petkovska, I.; Chakraborty, J.; Horvat, N. Rectal MRI Radiomics Inter- and Intra-Reader Reliability: Should We Worry about That? Abdom. Radiol. 2022, 47, 2004–2013. [Google Scholar] [CrossRef] [PubMed]
  26. Miranda, J.; Tan, G.X.V.; Fernandes, M.C.; Yildirim, O.; Sims, J.A.; Araujo-Filho, J.d.A.B.; de M Machado, F.A.; Assuncao, A.N., Jr.; Nomura, C.H.; Horvat, N. Rectal MRI Radiomics for Predicting Pathological Complete Response: Where We Are. Clin. Imaging 2022, 82, 141–149. [Google Scholar] [CrossRef] [PubMed]
  27. Namikawa, K.; Hirasawa, T.; Yoshio, T.; Fujisaki, J.; Ozawa, T.; Ishihara, S.; Aoki, T.; Yamada, A.; Koike, K.; Suzuki, H.; et al. Utilizing Artificial Intelligence in Endoscopy: A Clinician’s Guide. Expert. Rev. Gastroenterol. Hepatol. 2020, 14, 689–706. [Google Scholar] [CrossRef] [PubMed]
  28. Pacal, I.; Karaboga, D.; Basturk, A.; Akay, B.; Nalbantoglu, U. A Comprehensive Review of Deep Learning in Colon Cancer. Comput. Biol. Med. 2020, 126, 104003. [Google Scholar] [CrossRef]
  29. Qin, Y.; Zhu, L.-H.; Zhao, W.; Wang, J.-J.; Wang, H. Review of Radiomics- and Dosiomics-Based Predicting Models for Rectal Cancer. Front. Oncol. 2022, 12, 913683. [Google Scholar] [CrossRef]
  30. Reginelli, A.; Nardone, V.; Giacobbe, G.; Belfiore, M.P.; Grassi, R.; Schettino, F.; Del Canto, M.; Grassi, R.; Cappabianca, S. Radiomics as a New Frontier of Imaging for Cancer Prognosis: A Narrative Review. Diagnostics 2021, 11, 1796. [Google Scholar] [CrossRef]
  31. Staal, F.C.R.; van der Reijd, D.J.; Taghavi, M.; Lambregts, D.M.J.; Beets-Tan, R.G.H.; Maas, M. Radiomics for the Prediction of Treatment Outcome and Survival in Patients with Colorectal Cancer: A Systematic Review. Clin. Color. Cancer 2021, 20, 52–71. [Google Scholar] [CrossRef] [PubMed]
  32. Stanzione, A.; Verde, F.; Romeo, V.; Boccadifuoco, F.; Mainenti, P.P.; Maurea, S. Radiomics and Machine Learning Applications in Rectal Cancer: Current Update and Future Perspectives. World J. Gastroenterol. 2021, 27, 5306–5321. [Google Scholar] [CrossRef] [PubMed]
  33. Tabari, A.; Chan, S.M.; Omar, O.M.F.; Iqbal, S.I.; Gee, M.S.; Daye, D. Role of Machine Learning in Precision Oncology: Applications in Gastrointestinal Cancers. Cancers 2022, 15, 63. [Google Scholar] [CrossRef]
  34. Wang, P.-P.; Deng, C.-L.; Wu, B. Magnetic Resonance Imaging-Based Artificial Intelligence Model in Rectal Cancer. World J. Gastroenterol. 2021, 27, 2122–2130. [Google Scholar] [CrossRef]
  35. Wong, C.; Fu, Y.; Li, M.; Mu, S.; Chu, X.; Fu, J.; Lin, C.; Zhang, H. MRI-Based Artificial Intelligence in Rectal Cancer. J. Magn. Reson. Imaging 2022, 57, 45–56. [Google Scholar] [CrossRef] [PubMed]
  36. Xu, Q.; Xu, Y.; Sun, H.; Jiang, T.; Xie, S.; Ooi, B.Y.; Ding, Y. MRI Evaluation of Complete Response of Locally Advanced Rectal Cancer after Neoadjuvant Therapy: Current Status and Future Trends. Cancer Manag. Res. 2021, 13, 4317–4328. [Google Scholar] [CrossRef]
  37. Peng, J.; Ding, Y.; Tu, S.; Shi, D.; Sun, L.; Li, X.; Wu, H.; Cai, S. Prognostic Nomograms for Predicting Survival and Distant Metastases in Locally Advanced Rectal Cancers. PLoS ONE 2014, 9, e106344. [Google Scholar] [CrossRef]
  38. Sun, Y.; Lin, H.; Lu, X.; Huang, Y.; Xu, Z.; Huang, S.; Wang, X.; Chi, P. A Nomogram to Predict Distant Metastasis after Neoadjuvant Chemoradiotherapy and Radical Surgery in Patients with Locally Advanced Rectal Cancer. J. Surg. Oncol. 2017, 115, 462–469. [Google Scholar] [CrossRef]
  39. Valentini, V.; van Stiphout, R.G.P.M.; Lammering, G.; Gambacorta, M.A.; Barba, M.C.; Bebenek, M.; Bonnetain, F.; Bosset, J.-F.; Bujko, K.; Cionini, L.; et al. Nomograms for Predicting Local Recurrence, Distant Metastases, and Overall Survival for Patients with Locally Advanced Rectal Cancer on the Basis of European Randomized Clinical Trials. J. Clin. Oncol. 2011, 29, 3163–3172. [Google Scholar] [CrossRef]
  40. Chen, L.-D.; Li, W.; Xian, M.-F.; Zheng, X.; Lin, Y.; Liu, B.-X.; Lin, M.-X.; Li, X.; Zheng, Y.-L.; Xie, X.-Y.; et al. Preoperative Prediction of Tumour Deposits in Rectal Cancer by an Artificial Neural Network-Based US Radiomics Model. Eur. Radiol. 2020, 30, 1969–1979. [Google Scholar] [CrossRef]
  41. Cheng, Y.; Luo, Y.; Hu, Y.; Zhang, Z.; Wang, X.; Yu, Q.; Liu, G.; Cui, E.; Yu, T.; Jiang, X. Multiparametric MRI-Based Radiomics Approaches on Predicting Response to Neoadjuvant Chemoradiotherapy (NCRT) in Patients with Rectal Cancer. Abdom. Radiol. 2021, 46, 5072–5085. [Google Scholar] [CrossRef] [PubMed]
  42. Cui, Y.; Yang, X.; Shi, Z.; Yang, Z.; Du, X.; Zhao, Z.; Cheng, X. Radiomics Analysis of Multiparametric MRI for Prediction of Pathological Complete Response to Neoadjuvant Chemoradiotherapy in Locally Advanced Rectal Cancer. Eur. Radiol. 2018, 29, 1211–1220. [Google Scholar] [CrossRef]
  43. Dinapoli, N.; Barbaro, B.; Gatta, R.; Chiloiro, G.; Casà, C.; Masciocchi, C.; Damiani, A.; Boldrini, L.; Gambacorta, M.A.; Dezio, M.; et al. Magnetic Resonance, Vendor-Independent, Intensity Histogram Analysis Predicting Pathologic Complete Response after Radiochemotherapy of Rectal Cancer. Int. J. Radiat. Oncol. Biol. Phys. 2018, 102, 765–774. [Google Scholar] [CrossRef] [PubMed]
  44. Ding, L.; Liu, G.; Zhang, X.; Liu, S.; Li, S.; Zhang, Z.; Guo, Y.; Lu, Y. A Deep Learning Nomogram Kit for Predicting Metastatic Lymph Nodes in Rectal Cancer. Cancer Med. 2020, 9, 8809–8820. [Google Scholar] [CrossRef]
  45. Huang, Y.; Liang, C.; He, L.; Tian, J.; Liang, C.; Chen, X.; Ma, Z.; Liu, Z. Development and Validation of a Radiomics Nomogram for Preoperative Prediction of Lymph Node Metastasis in Colorectal Cancer. J. Clin. Oncol. 2016, 34, 2157–2164. [Google Scholar] [CrossRef] [PubMed]
  46. Jin, C.; Yu, H.; Ke, J.; Ding, P.; Yi, Y.; Jiang, X.; Duan, X.; Tang, J.; Chang, D.T.; Wu, X.; et al. Predicting Treatment Response from Longitudinal Images Using Multi-Task Deep Learning. Nat. Commun. 2021, 12, 1851. [Google Scholar] [CrossRef] [PubMed]
  47. Kleppe, A.; Skrede, O.-J.; Raedt, S.D.; Hveem, T.S.; Askautrud, H.A.; Jacobsen, J.E.; Church, D.N.; Nesbakken, A.; Shepherd, N.A.; Novelli, M.; et al. A Clinical Decision Support System Optimising Adjuvant Chemotherapy for Colorectal Cancers by Integrating Deep Learning and Pathological Staging Markers: A Development and Validation Study. Lancet Oncol. 2022, 23, 1221–1232. [Google Scholar] [CrossRef]
  48. Li, M.; Jin, Y.-M.; Zhang, Y.-C.; Zhao, Y.-L.; Huang, C.-C.; Liu, S.-M.; Song, B. Radiomics for Predicting Perineural Invasion Status in Rectal Cancer. World J. Gastroenterol. 2021, 27, 5610–5621. [Google Scholar] [CrossRef]
  49. Liu, H.; Yin, H.; Li, J.; Dong, X.; Zheng, H.; Zhang, T.; Yin, Q.; Zhang, Z.; Lu, M.; Zhang, H.; et al. Deep Learning Model Based on MRI and Clinical Factors Facilitates Noninvasive Evaluation of KRAS Mutation in Rectal Cancer. J. Magn. Reson. Imaging 2022, 56, 1659–1668. [Google Scholar] [CrossRef]
  50. Liu, S.; Yu, X.; Yang, S.; Hu, P.; Hu, Y.; Chen, X.; Li, Y.; Zhang, Z.; Li, C.; Lu, Q. Machine Learning-Based Radiomics Nomogram for Detecting Extramural Venous Invasion in Rectal Cancer. Front. Oncol. 2021, 11, 610338. [Google Scholar] [CrossRef]
  51. Liu, X.; Zhang, D.; Liu, Z.; Li, Z.; Xie, P.; Sun, K.; Wei, W.; Dai, W.; Tang, Z.; Ding, Y.; et al. Deep Learning Radiomics-Based Prediction of Distant Metastasis in Patients with Locally Advanced Rectal Cancer after Neoadjuvant Chemoradiotherapy: A Multicentre Study. EBioMedicine 2021, 69, 103442. [Google Scholar] [CrossRef] [PubMed]
  52. Mao, Y.; Pei, Q.; Fu, Y.; Liu, H.; Chen, C.; Li, H.; Gong, G.; Yin, H.; Pang, P.; Lin, H.; et al. Pre-Treatment Computed Tomography Radiomics for Predicting the Response to Neoadjuvant Chemoradiation in Locally Advanced Rectal Cancer: A Retrospective Study. Front. Oncol. 2022, 12, 850774. [Google Scholar] [CrossRef]
  53. Peterson, K.J.; Simpson, M.T.; Drezdzon, M.K.; Szabo, A.; Ausman, R.A.; Nencka, A.S.; Knechtges, P.M.; Peterson, C.Y.; Ludwig, K.A.; Ridolfi, T.J. Predicting Neoadjuvant Treatment Response in Rectal Cancer Using Machine Learning: Evaluation of MRI-Based Radiomic and Clinical Models. J. Gastrointest. Surg. 2023, 27, 122–130. [Google Scholar] [CrossRef] [PubMed]
  54. van Stiphout, R.G.P.M.; Lammering, G.; Buijsen, J.; Janssen, M.H.M.; Gambacorta, M.A.; Slagmolen, P.; Lambrecht, M.; Rubello, D.; Gava, M.; Giordano, A.; et al. Development and External Validation of a Predictive Model for Pathological Complete Response of Rectal Cancer Patients Including Sequential PET-CT Imaging. Radiother. Oncol. 2011, 98, 126–133. [Google Scholar] [CrossRef]
  55. van Stiphout, R.G.P.M.; Valentini, V.; Buijsen, J.; Lammering, G.; Meldolesi, E.; van Soest, J.; Leccisotti, L.; Giordano, A.; Gambacorta, M.A.; Dekker, A.; et al. Nomogram Predicting Response after Chemoradiotherapy in Rectal Cancer Using Sequential PETCT Imaging: A Multicentric Prospective Study with External Validation. Radiother. Oncol. 2014, 113, 215–222. [Google Scholar] [CrossRef]
  56. Wan, L.; Zhang, C.; Zhao, Q.; Meng, Y.; Zou, S.; Yang, Y.; Liu, Y.; Jiang, J.; Ye, F.; Ouyang, H.; et al. Developing a Prediction Model Based on MRI for Pathological Complete Response after Neoadjuvant Chemoradiotherapy in Locally Advanced Rectal Cancer. Abdom. Radiol. 2019, 44, 2978–2987. [Google Scholar] [CrossRef] [PubMed]
  57. Wei, Q.; Chen, Z.; Tang, Y.; Chen, W.; Zhong, L.; Mao, L.; Hu, S.; Wu, Y.; Deng, K.; Yang, W.; et al. External Validation and Comparison of MR-Based Radiomics Models for Predicting Pathological Complete Response in Locally Advanced Rectal Cancer: A Two-Centre, Multi-Vendor Study. Eur. Radiol. 2022, 33, 1906–1917. [Google Scholar] [CrossRef]
  58. Wei, Q.; Yuan, W.; Jia, Z.; Chen, J.; Li, L.; Yan, Z.; Liao, Y.; Mao, L.; Hu, S.; Liu, X.; et al. Preoperative MR Radiomics Based on High-Resolution T2-Weighted Images and Amide Proton Transfer-Weighted Imaging for Predicting Lymph Node Metastasis in Rectal Adenocarcinoma. Abdom. Radiol. 2023, 48, 458–470. [Google Scholar] [CrossRef]
  59. Yi, X.; Pei, Q.; Zhang, Y.; Zhu, H.; Wang, Z.; Chen, C.; Li, Q.; Long, X.; Tan, F.; Zhou, Z.; et al. MRI-Based Radiomics Predicts Tumor Response to Neoadjuvant Chemoradiotherapy in Locally Advanced Rectal Cancer. Front. Oncol. 2019, 9, 552. [Google Scholar] [CrossRef]
  60. Roselló, S.; Frasson, M.; García-Granero, E.; Roda, D.; Jordá, E.; Navarro, S.; Campos, S.; Esclápez, P.; García-Botello, S.; Flor, B.; et al. Integrating Downstaging in the Risk Assessment of Patients with Locally Advanced Rectal Cancer Treated with Neoadjuvant Chemoradiotherapy: Validation of Valentini’s Nomograms and the Neoadjuvant Rectal Score. Clin. Color. Cancer 2018, 17, 104–112.e2. [Google Scholar] [CrossRef]
  61. Tang, B.; Lenkowicz, J.; Peng, Q.; Boldrini, L.; Hou, Q.; Dinapoli, N.; Valentini, V.; Diao, P.; Yin, G.; Orlandini, L.C. Local Tuning of Radiomics-Based Model for Predicting Pathological Response to Neoadjuvant Chemoradiotherapy in Locally Advanced Rectal Cancer. BMC Med. Imaging 2022, 22, 44. [Google Scholar] [CrossRef] [PubMed]
  62. Collins, G.S.; Reitsma, J.B.; Altman, D.G.; Moons, K.G.M. Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD): The TRIPOD Statement. BMJ 2015, 350, g7594. [Google Scholar] [CrossRef] [PubMed]
  63. Montesinos-López, O.A.; Montesinos-López, A.; Mosqueda-Gonzalez, B.A.; Montesinos-López, J.C.; Crossa, J. Accounting for Correlation between Traits in Genomic Prediction. Methods Mol. Biol. 2022, 2467, 285–327. [Google Scholar] [PubMed]
  64. Lipkova, J.; Chen, R.J.; Chen, B.; Lu, M.Y.; Barbieri, M.; Shao, D.; Vaidya, A.J.; Chen, C.; Zhuang, L.; Williamson, D.F.K.; et al. Artificial Intelligence for Multimodal Data Integration in Oncology. Cancer Cell 2022, 40, 1095–1110. [Google Scholar] [CrossRef]
  65. Feng, L.; Liu, Z.; Li, C.; Li, Z.; Lou, X.; Shao, L.; Wan, J.; Huang, Y.; Chen, H.; Pang, X.; et al. Development and Validation of a Radiopathomics Model to Predict Pathological Complete Response to Neoadjuvant Chemoradiotherapy in Locally Advanced Rectal Cancer: A Multicentre Observational Study. Lancet Digit. Health 2022, 4, e8–e17. [Google Scholar] [CrossRef]
  66. Kim, D.W.; Jang, H.Y.; Kim, K.W.; Shin, Y.; Park, S.H. Design Characteristics of Studies Reporting the Performance of Artificial Intelligence Algorithms for Diagnostic Analysis of Medical Images: Results from Recently Published Papers. Korean J. Radiol. 2019, 20, 405. [Google Scholar] [CrossRef]
  67. Wolff, R.F.; Moons, K.G.M.; Riley, R.D.; Whiting, P.F.; Westwood, M.; Collins, G.S.; Reitsma, J.B.; Kleijnen, J.; Mallett, S. PROBAST: A Tool to Assess the Risk of Bias and Applicability of Prediction Model Studies. Ann. Intern. Med. 2019, 170, 51–58. [Google Scholar] [CrossRef]
  68. Huang, S.C.; Pareek, A.; Seyyedi, S.; Banerjee, I.; Lungren, M.P. Fusion of medical imaging and electronic health records using deep learning: A systematic review and implementation guidelines. NPJ Digit. Med. 2020, 3, 136. [Google Scholar] [CrossRef]
Figure 1. Schematic representation of a typical artificial neural network, with multiple hidden layers (i.e., deep learning). Circles represent artificial neurons.
Figure 1. Schematic representation of a typical artificial neural network, with multiple hidden layers (i.e., deep learning). Circles represent artificial neurons.
Cancers 17 02235 g001
Table 1. Literature on non-image artificial intelligence techniques applied to rectal cancer.
Table 1. Literature on non-image artificial intelligence techniques applied to rectal cancer.
ReferenceFirst AuthorYearJournalAimNumber of PatientsTRIPODAI Model(s)
TotalTrainingValidation
NIMs
[37]Peng J.2014PLOS OnePredict OS, DMs (and LR)917833842bCox regression
[38]Sun Y.2017Journal of Surgical OncologyPredict DMs after nCRT522425972bCox regression
[39]Valentini V.2011Journal of Clinical OncologyPredict OS, LR, DMs279522425533Cox regression
CMs
[40]Chen L.D.2020European RadiologyPredict TDs12787402bANN
[41]Cheng Y.2021Abdominal RadiologyPredict response (particularly, pCR) to nCRT193128652aLogistic regression
[42]Cui Y.2019European RadiologyPredict response (particularly, pCR) to nCRT186131552aLogistic regression
[43]Dinapoli N.2018International Journal of Radiation OncologyPredict response (particularly, pCR) to nCRT221162593GLM
[44]Ding L.2020Cancer MedicinePredict preoperative LN metastases5453621832aLogistic regression, DL
[45]Huang Y.Q.2016Journal of Clinical OncologyPredict preoperative LN metastases5263262002bLogistic regression
[46]Jin C.2021Nature CommunicationsPredict response (particularly, pCR) to nCRT6223213013DL
[47]Kleppe A.2022Lancet OncologyOptimize adjuvant therapy207299710753Cox regression, DL
[48]Li M.2021World Journal of GastroenterologyPredict PNI303242612aLogistic regression
[49]Liu H.2022Journal of Magnetic Resonance ImagingEvaluate KRAS mutation376288882aLogistic regression, DL
[50]Liu S.2021Frontiers in OncologyDetect preoperative EMVI281198832bLogistic regression
[51]Liu X.2021Lancet EBioMedicinePredict DMs after nCRT235170653DL
[52]Mao Y.2022Frontiers in OncologyPredict response (particularly, pCR) to nCRT216151652aLogistic regression
[53]Peterson K.J.2023Journal of Gastrointestinal SurgeryPredict response (particularly, pCR) to nCRT131111202aLogistic regression
[54]van Stiphout R.G.P.M.2011Radiotherapy & OncologyPredict response (particularly, pCR) to nCRT953Various groupings3SVM
[55]van Stiphout R.G.P.M.2014Radiotherapy & OncologyPredict response (particularly, pCR) to nCRT190112783Logistic regression
[56]Wan L.2019Abdominal RadiologyPredict response (particularly, pCR) to nCRT12084362bLogistic regression
[57]Wei Q.2023European RadiologyPredict response (particularly, pCR) to nCRT151100513 (plus 2a)RF
[58]Wei Q.2023Abdominal RadiologyPredict preoperative LN metastases12580452aLogistic regression
[59]Yi X.2019Frontiers in OncologyPredict response (particularly, pCR) to nCRT134101332aSVM, RF, LASSO
Legend. CM = Combined model; DM = Distant metastasis; EMVI = Extramural venous invasion; GLM = General linear model; LASSO = Least absolute shrinkage and selection operator; LN = Lymph node; LR = Local recurrence; nCRT = Neoadjuvant chemoradiotherapy; NIM = Non-image model; OS = Overall survival; pCR = Pathological complete response; PNI = Peri-neural invasion; RF = Random forest; SVM = Support vector machine; TD = Tumor deposit; TRIPOD = Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis.
Table 3. Details about combined models listed in Table 1 and Table 2. (A) subtable contains the main information, while (B) subtable refers specifically to only one, more-complex reference.
Table 3. Details about combined models listed in Table 1 and Table 2. (A) subtable contains the main information, while (B) subtable refers specifically to only one, more-complex reference.
(A)ReferenceFirst AuthorYearPerformances
TypeModel Component(s)PerformanceCombined ModelPerformanceNotes
CMs
[40]Chen L.D.2020AUCUS–Radiomics0.74US–Radiomics + Clinical0.80
[41]Cheng Y.2021Clinical0.84–0.91MRI–Radiomics + Clinical0.91–0.94Range values: minimum, pCR; maximum, GR
[42]Cui Y.2019MRI–Radiomics0.94MRI–Radiomics + Clinical0.97
[43]Dinapoli N.2018 MRI–Radiomics + Clinical0.75
[44]Ding L.2020 MRI Nomograms (2 interrelated outcomes)0.89–0.92Range values: minimum, degree outcome; maximum, status outcome
[45]Huang Y.Q.2016C-Index CT–Radiomics + Clinical0.78
[46]Jin C.2021AUCMRI (2 external validations)0.92–0.95MRI + Blood Markers0.97
[47]Kleppe A.2022HRH&E–DL (Poor vs. Good)3.04 (Adjusted)–3.84 (Unadjusted)H&E-DL + Pathological Markers (High vs. Low, Intermediate vs. Low)10.71, 3.06
[48]Li M.2021AUCClinical0.67Clinical + CT–Radiomics0.80
[49]Liu H.2022Clinical|MRI0.67|0.77Clinical + MRI0.84
[50]Liu S.2021Radiomics + Clinical|Radiomics + mrEMVI0.70|0.77Radiomics + Clinical + mrEMVI0.83
[51]Liu X.2021C-IndexMRI–Radiomics0.75MRI–Radiomics + Clinical0.78
[52]Mao Y.2022AUCClinical | CT–Radiomics0.79|0.83Clinical + CT–Radiomics0.87
[53]Peterson K.J.2023Clinical Markers0.64–0.69Clinical Markers + MRI–Radiomics0.73
[54]van Stiphout RGPM2011Clinical0.69Clinical + PET–CT (Post-nCRT)0.86
[55]van Stiphout R.G.P.M.2014 Clinical + PET–CT0.70
[56]Wan L.2019mrTRG|rT2wSI-related|CATV-relatedTraining: 0.68, 0.77, 0.83MRI + Clinical0.84
[57]Wei Q.2023Clinical|MRI-Radiomics0.69|0.83Clinical + MRI-Radiomics0.87Performance of best model (RF)
[58]Wei Q.2023MRI–Radiomics0.85Clinical + MRI-Radiomics0.93Performance of best model (LR)
[59]Yi X.2019See (B) sub-tableThree outcomes
(B)ReferenceFirst AuthorYearOutcomeModel
Image1Image2 + ClinicalCombined
[59]Yi X.2019pCR0.820.760.88
2019GR0.790.770.90
2019DS0.800.850.89
Legend. AUC = Area under the curve; C-Index = Concordance index; CM = Combined model; DL = Deep learning; DS = Down-staging; EMVI = Extramural venous invasion; GR = Good response; HR = Hazard ratio; LR = Logistic regression; nCRT = Neoadjuvant chemoradiotherapy; pCR = Pathological complete response; RF = Random forest.
Table 4. Discrimination values (median and interquartile range) based on validation type.
Table 4. Discrimination values (median and interquartile range) based on validation type.
ValidationArea Under the CurveConcordance Index
Internal0.860.080.750.03
External0.860.120.740.04
Table 5. Preprocessing in the analyzed primary literature.
Table 5. Preprocessing in the analyzed primary literature.
ReferenceFirst AuthorYearFeature-Selection
Technique(s)
Data
NormalizationImputationAugmentation
NIMs
[37]Peng J.2014
[38]Sun Y.2017
[39]Valentini V.2011 Z-score transformationExpectation–maximization
CMs
[40]Chen L.D.2020Spearman correlation + LASSO
[41]Cheng Y.2021Intra-class correlation + Mann–Whitney U test + LASSO
[42]Cui Y.2019Univariate LR + LASSO + Pearson correlationZ-score transformation
[43]Dinapoli N.2018Mann-Whitney test
[44]Ding L.2020Univariate/Multivariate analyses (NOS)
[45]Huang Y.Q.2016LASSO
[46]Jin C.2021 Image data are normalized
[47]Kleppe A.2022
[48]Li M.2021Intra-class correlation + 6 other methods (analysis of variance, Pearson, mutual information, L1-based, tree-based, recursive) + univariate LRZ-score transformation
[49]Liu H.2022 MRI image normalization to [0, 255] Yes
[50]Liu S.2021Spearman correlation + mRMR + LASSOImage normalization
[51]Liu X.2021 Image Z-score transformation
[52]Mao Y.2022Inter-/Intra-observer correlation + LASSO + LR (uni-/multi-variate)
[53]Peterson K.J.2023Correlation (NOS) + LASSOZ-score transformation
[54]van Stiphout R.G.P.M.2011Spearman correlation + Wilcoxon rank-sum testZ-score transformationMean
[55]van Stiphout R.G.P.M.2014Spearman correlation + Wilcoxon rank-sum test Expectation–maximization
[56]Wan L.2019Univariate analysis (t and Mann–Whitney U or chi-square and Fisher exact tests) + LASSO + Intraclass correlation
[57]Wei Q.2023Inter-class correlation + LASSOImage intensity normalization to [0, 255]
[58]Wei Q.2023Intra-class correlation + Spearman correlation + multivariate LRImage intensity Z-score transformationMedian
[59]Yi X.2019Inter-/Intra-observer correlation + LASSO
Legend. CM = Combined model; LASSO = Least absolute shrinkage and selection operator; LR = Logistic regression; mRMR = Minimum redundancy maximum correlation; NIM = Non-image model.; NOS = Not otherwise specified.
Table 6. Review articles addressing artificial intelligence models of rectal cancer identified in our initial secondary literature evaluation.
Table 6. Review articles addressing artificial intelligence models of rectal cancer identified in our initial secondary literature evaluation.
ReferenceFirst authorYearJournalIMsCMsNIMsNotes
[11]Lambin P.2017Nature Reviews Clinical OncologyMain focusBriefly mentionedNot presentRadiomics
[12]Meldolesi E.2016Future Oncology
[20]Jia L.L.2022Frontiers in OncologySystematic review (with meta-analysis)
[21]Bedrikovetski S.2021BMC Cancer
[22]Coppola F.2021Diagnostics Radiomics
[23]Kalantar R.2021Diagnostics Deep learning
[24]Kuntz S.2021European Journal of Cancer Deep learning; Systematic review
[25]Kwok H.C.2022Abdominal Radiology Radiomics
[26]Miranda J.2022Clinical Imaging
[27]Namikawa K.2020Expert Review of Gastroenterology & Hepatolog
[28]Pacal I.2020Computers in Biology and Medicine Deep learning
[29]Qin Y.2022Frontiers in Oncology Radiomics
[30]Reginelli A.2021Diagnostics
[31]Staal F.C.R.2021Clinical Colorectal Cancer Radiomics; Systematic review
[32]Stanzione A.2021World Journal of Gastroenterology Radiomics
[33]Tabari A.2022Cancers
[34]Wang P.P.2021World Journal of Gastroenterology
[35]Wong C.2023Journal of Magnetic Resonance Imaging
[36]Xu Q.2021Cancer Management and Research
Legend. CM = Combined model. IM = Image model. NIM = Non-image model.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Novellino, T.; Masciocchi, C.; Tudor, A.M.; Casà, C.; Chiloiro, G.; Romano, A.; Damiani, A.; Arcuri, G.; Gambacorta, M.A.; Valentini, V. Artificial Intelligence and Rectal Cancer: Beyond Images. Cancers 2025, 17, 2235. https://doi.org/10.3390/cancers17132235

AMA Style

Novellino T, Masciocchi C, Tudor AM, Casà C, Chiloiro G, Romano A, Damiani A, Arcuri G, Gambacorta MA, Valentini V. Artificial Intelligence and Rectal Cancer: Beyond Images. Cancers. 2025; 17(13):2235. https://doi.org/10.3390/cancers17132235

Chicago/Turabian Style

Novellino, Tommaso, Carlotta Masciocchi, Andrada Mihaela Tudor, Calogero Casà, Giuditta Chiloiro, Angela Romano, Andrea Damiani, Giovanni Arcuri, Maria Antonietta Gambacorta, and Vincenzo Valentini. 2025. "Artificial Intelligence and Rectal Cancer: Beyond Images" Cancers 17, no. 13: 2235. https://doi.org/10.3390/cancers17132235

APA Style

Novellino, T., Masciocchi, C., Tudor, A. M., Casà, C., Chiloiro, G., Romano, A., Damiani, A., Arcuri, G., Gambacorta, M. A., & Valentini, V. (2025). Artificial Intelligence and Rectal Cancer: Beyond Images. Cancers, 17(13), 2235. https://doi.org/10.3390/cancers17132235

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop