Considering the nonspecific nature of gastrointestinal complaints and the broad differentials of gastrointestinal symptomatology, imaging plays a vital role in the formulation of diagnoses. As a result, artificial intelligence (AI) tools have emerged to assist radiologists in the interpretation of gastrointestinal imaging and to mitigate diagnostic errors. Among the main subtypes of AI applied in this field is deep learning (DL), a subfield of machine learning (ML) that uses artificial neural networks to analyze data and has proven to be superior to traditional ML methods in radiologic imaging analysis. In this review, we discuss DL applications in gastrointestinal imaging across different modalities, including x-ray imaging, ultrasonography, computed tomography, magnetic resonance tomography, and positron emission tomography. Moreover, we outline the challenges and ethical considerations facing the growing role of AI in clinical practice.

Gastrointestinal (GI) diseases are a wide range of disorders that affect the digestive tract, including the esophagus, stomach, small bowel, and colon, and the accessory organs including the liver, gallbladder, and pancreas.[1] Patients with these disorders can present with nonspecific symptoms such as abdominal pain, nausea, vomiting, diarrhea, and constipation, which can lead to diagnostic challenges. Radiologic imaging techniques such as ultrasonography, x-ray imaging, computed tomography (CT), magnetic resonance imaging (MRI), and positron emission tomography (PET) play an essential role in diagnosing and managing various GI disorders.[2] Artificial intelligence (AI) was first defined in 1956 and significant advancements in healthcare have been made possible with the use of AI since then.[3] AI is a computer system that can perform human tasks by making decisions learned from previous experiences.[3] AI systems have many applications in healthcare, including assisting in patient diagnosis, predicting patients’ prognosis, transcribing notes, and organizing patients’ data.[4] Machine learning (ML) and deep learning (DL) are two AI methods widely used in healthcare settings.[5] ML emphasizes the use of machines to perform intelligent tasks by learning from their previous mistakes, while DL is a subfield of ML that uses artificial neural network architectures for processing and extracting data.[6] A conventional neural network (CNN) is a type of DL that uses fully connected layers with standard weights for image classification and analysis.[5] The use of AI in diagnostic radiologic imaging has evolved significantly over the current decade.[7] ML has demonstrated great performance in classifying and analyzing radiologic images.[7] DL has also been used in image analysis by using neural networks and has shown greater performance than traditional ML models.[7] AI-based radiologic imaging can improve diagnostic efficiency, reduce interpretation errors, and optimize the imaging workflow.[8] The use of AI in radiologic imaging of GI conditions has been reported widely in the literature, such as organ segmentation, automated lesion detection, and computer-aided diagnosis.[9]

AI has significantly advanced GI imaging and played a crucial role in developing predictive models, leading to improvements in the detection, diagnosis, and prognosis of many diseases.[10,11] The radiologic techniques that have benefited from AI include ultrasonography, x-ray imaging, CT, MRI, and PET-CT. Furthermore, the automated analysis of GI images commonly involves ML, natural language processing (NLP), and CNN.[9] The rise of AI in healthcare, especially in diagnostic radiology, has opened unparalleled possibilities for improving the standard and effectiveness of patient care. However, this rapid advancement brings numerous challenges, including the need for sufficient data quality and quantity, incorporation into clinical routines, software security, and numerous ethical and legal issues.[10,12] Although there is extensive discussion in the literature regarding the ethical aspects and challenges associated with the application of AI, a universally accepted and comprehensive framework for the ethical development and implementation of AI in healthcare is yet to be established.[13] To fully leverage the potential of AI, it is essential to invest in building the necessary infrastructure that fosters data sharing, adheres to uniform information governance policies, and encourages innovation while ensuring the confidentiality of patient information. Additionally, it is crucial to establish the groundwork for educating and preparing the workforce for an AI-driven healthcare system in the future.[14] In this review, we aim to explore the current role of DL in radiologic imaging of GI diseases.

We searched through Google Scholar and PubMed databases and included articles that discussed DL application in radiologic imaging of different GI conditions. Our focus was on studies that fit within the scope of our review purpose from 2018 onward, and we cited older studies only if they had an impact on the role of DL in GI imaging. Our search terms in the field of DL in radiologic imaging included gastrointestinal, esophageal, gastric, pancreatic, hepatobiliary, small bowel, and colorectal diseases.

Throughout this review we will be discussing articles that report a Dice similarity coefficient (DSC) or simply, Dice score, which is a measure of segmentation performance and spatial overlap between two regions by considering both false positives and false negatives. The DSC ranges from 0 to 1, where 0 means no overlap, and 1 is perfect agreement between the proposed method and ground truth.[15] We will also be discussing radiomics, which is the process of quantifying medical images into high-dimensional data related to predictive targets and extracting valuable information to enhance clinical decision-making.[16] In addition, the receiver operating characteristic curve is a graphical representation used to evaluate the performance of a classification model or diagnostic test. A commonly reported statistic is the area under the curve (AUC), which quantifies the overall ability of the model to discriminate between positive and negative values. AUC values range from 0.5 to 1.0. An AUC of 0.5 means the test performs no better than random chance in distinguishing between benign and malignant tumors, for example, whereas an AUC of 1.0 signifies perfect discrimination.[17]Another commonly reported statistic, the C-index, or concordance index, is primarily used in survival analysis and binary classification to evaluate the discriminatory power of a predictive model. It measures how well the model can distinguish between different outcomes or risk levels.[18]

GI organ segmentation on radiologic images is a critical but time-consuming task for radiologists to obtain manually. Consequentially, multiple semi-automatic and fully automatic DL methods to segment different GI organs on CT and MRI images have emerged over the past couple of years.

In a study reported by Gibson et al,[19] the authors proposed a DL-based segmentation model on abdominal CT images that can segment the pancreas, esophagus, stomach, liver, spleen, gallbladder, left kidney, and duodenum more accurately than previously reported DL methods. In another study, Xue et al[20] developed a cascade multitask 3D conventional network to address the pancreatic segmentation challenges on CT images that outperformed the state-of-the-art methods. In a study from China, researchers developed a 2.5D CNN to automatically segment the pancreas on 3D CT images, with a DSC of 86.21 ± 4.37%, sensitivity of 87.49 ± 6.38%, and specificity of 85.11 ± 6.49%.[21] In another study reported by Zheng et al,[22] the authors proposed a 2D DL-based iterative pancreas segmentation workflow method to describe the uncertain regions on pancreatic MRI images. This method improved pancreatic segmentation by addressing the pancreatic segmentation challenges, especially the critical areas for pancreatic cancer diagnosis, which include the pancreatic heads, tails, boundaries and disconnected topology. Furthermore, Panda et al[23] developed a highly accurate and precise 3D CNN model for fully automated segmentation of the pancreas on a large CT scan dataset. A limitation of this study was that the proposed segmentation method was for the morphologically normal pancreas rather than the diseased pancreas.

In 2020, Kim et al[24] proposed a CNN model for abdominal multiorgan segmentation, including the liver, stomach, duodenum, and kidneys, from 3D-patch–based CT images. The average segmentation time without automation was 22.6 minutes, compared to 7.1 minutes with automation using the CNN model. The proposed CNN also showed comparable accuracy to radiographers, especially for the liver and kidneys. However, the CNN model showed a relatively weak performance for the stomach and duodenum owing to their low-intensity gradients. One of the limitations of the study was using only the time reduction parameter to show the efficiency of CNN auto-segmentation.[24]

Furthermore, DL models were also developed to automatically segment different GI organs on MRI images. Nainamalai et al[25] assessed the applicability of two automated DL algorithms for liver parenchyma segmentation of contrast-enhanced T1-weighted MR images to generate patient-specific models that aid clinicians in planning surgeries. They achieved a Dice score of 0.9696 and concluded that around one-third of the segmentations using the DL algorithms can produce results without any manual correction necessary.

Fu et al[26] performed a retrospective study on 3D MRI images and proposed a CNN model to automatically segment the liver, kidneys, stomach, duodenum, and bowel, which showed good accuracy. In another study, Chen et al[27] developed a CNN model with a high accuracy for fully automated multiorgan segmentation on abdominal MRI images, including the liver, spleen, pancreas, left and right kidneys, stomach, duodenum, small intestine, spinal cord, and vertebral bodies. One limitation of the study is the small testing dataset with only 20 cases.

The integration of DL models in imaging modalities has shown significant promise in enhancing diagnostic accuracy for various GI conditions and cancers. Studies have demonstrated that DL models can often outperform traditional methods in certain aspects.

A study from Japan assessed the performance of a CNN-based DL identification system for esophageal cancer from CT images, which was able to detect the cancer with higher accuracy than radiologists.[28] The CNN-based system demonstrated a diagnostic accuracy of 84.2%, a specificity of 90%, and a sensitivity of 71.7%. Another study by Lin et al[29] compared the diagnostic efficacy of a DL-based model built on noncontrast CT imaging with the performance of radiologists in detecting esophageal cancer. They retrospectively analyzed and trained the DL model with a group of patients with esophageal cancer and with pathologically confirmed noncontrast CT imaging and another group of healthy individuals with endoscopy-proven esophageal cancer–free status. The model achieved an accuracy level of 88.2%, which was higher than that of junior radiologists but lower than the rate of senior radiologists.[29] In this study, a complex case featuring a 77-year-old man with T1 stage esophageal cancer was overlooked by the radiologists but was successfully identified by the DL model. However, a significant limitation of this study was that most patients with early-stage esophageal cancer who were enrolled in the study were not identified by either the DL model or the radiologists. Further studies are needed to determine whether implementing this model will improve patient care.

DL has been applied in different imaging modalities to aid in the detection of hepatocellular carcinoma (HCC). Xu et al[30] developed an AI pipeline with an integrated CNN segmentation model using a large-scale ultrasound dataset, which was able to detect liver masses with an AUC of 0.990 (95% CI, 0.986–0.992), a sensitivity of 95.1%, and a specificity of 96.1%. However, owing to the limited number of pathology results, this model was tested only on three subtype classifications for benign and malignant masses. Besides ultrasonographic imaging of the liver, Ghoniem[31] proposed a DL algorithm for the detection of liver masses on CT images. This method achieved a Dice score of 0.968 and 0.97 across two different datasets, outperforming previous segmentation methods. Considering the different primary and metastatic liver masses, the differentiation of liver tumors is a fertile ground for the application of AI.[32] Hu et al[33] developed a contrast-enhanced ultrasound (CEUS)–based model that achieved a higher accuracy in the differentiation of liver masses than resident physicians and was found to be on par with the performance of hepatic CEUS experts. DL has also been used in CT imaging to differentiate benign and malignant liver masses; Midya et al[34] developed a deep neural network to differentiate liver tumors, using single portal venous phase contrast-enhanced CT images. This model identified four types of liver masses: intrahepatic cholangiocarcinoma, HCC, colorectal liver metastasis, and benign masses, with an accuracy of 96.27%.[34] Despite the promising results, this study is limited by the small sample size of benign tumors, and selection bias for patients who had undergone surgical intervention for liver masses. Another recent study has evaluated the use of DL to differentiate the primary source of liver metastases. The DL radiomics-based model in this study outperformed ML models in identifying the source of liver metastasis, with an AUC of 0.907 and 0.809 for digestive and nondigestive tract malignancies, respectively.[35] One more potential benefit of DL is exposing the patients to less radiation. In one study, three-phase CT images yielded a similar diagnostic accuracy to a four-phase protocol in distinguishing HCC from other focal liver lesions.[36]

In addition to the detection and differentiation of malignancies, DL has a promising future in grading the stages of liver fibrosis. Lee et al[37] developed a deep convolutional neural network–based model for the classification of liver fibrosis, using ultrasonography. The model achieved an accuracy of 83.5% and 76.4% on the internal and external datasets, respectively.

In terms of the application of DL in biliary tract imaging, Obaid et al[38] used four different CNN models for the detection of different gallbladder pathologies (including cholecystitis, gallstones, perforation, gallbladder wall thickening), using ultrasound imaging. Of these four models, MobileNet achieved the highest accuracy of 98.35%. DL has also been used in differentiating intrahepatic cholangiocarcinoma from HCC; Ponnoprat et al[39] have achieved an accuracy of 88% in differentiating intrahepatic cholangiocarcinomas and HCC on multiphase abdominal CT, using a machine-based model that integrates CNN.

Another application of DL is the automated detection of pancreatic cancers on CT images. A retrospective study from Taiwan, which included two local datasets and one dataset from the United States, used a CNN model to distinguish cancerous from noncancerous pancreatic tissues on CT images.[40] The model showed acceptable accuracy and higher sensitivity (98.3%) than radiologists (92.9%). Additionally, CNN missed 3 (1.7%) of 176 pancreatic cancers, while radiologists missed 12 (7%) of 168 pancreatic cancers, of which 11 (92%) were correctly classified by using CNN.[40] However, the CNN-based model achieving higher sensitivity than radiologists should be interpreted cautiously, as radiologists had additional clinical information when interpreting CT images, whereas the CNN model only had access to the CT images. Additionally, the study used a small dataset of Asian participants from one institution, so it may not cover the full range of pancreatic cancer CT-imaging manifestations. Zhu et al[41] also used a DL model to analyze CT images from 303 patients with pancreatic cancer and 136 controls and achieved 94.1% sensitivity and 98.5% specificity in detecting pancreatic cancer. One study from the Netherlands developed a DL model to automatically detect pancreatic ductal adenocarcinoma (PDAC) on CT images.[42] The proposed model achieved a maximum AUC of 0.914 in the external test set and 0.876 for the subgroup of tumors smaller than 2 cm, suggesting that it can assist radiologists in diagnosing PDAC. However, the model was trained with a relatively low number of patients and only included tumors in the pancreatic head. In a recent study in 2023 that included 6239 patients across 10 centers, the DL model showed an AUC of 0.986–0.996 for lesion detection on noncontrast CT images, outperforming the mean radiologist performance by 34.1% in sensitivity and 6.3% in specificity for PDAC identification.[43] Moreover, for lesion detection in a real-world multiscenario validation consisting of 20,530 patients, the DL model showed a sensitivity of 92.9% and specificity of 99.9%. The DL model was trained by using multicenter data but mainly from the East Asian population and hospitals. Therefore, further validation in external real-world centers, international cohorts, and prospective studies is necessary.

Another application of DL in radiologic imaging is detecting small-bowel obstruction (SBO). Kim et al[44] developed a DL model to identify SBO on plain abdominal radiographs with high accuracy, achieving an AUC of 0.961, sensitivity of 91%, and specificity of 93%. However, the clinical relevance of these findings is limited because abdominal radiographs are less sensitive than CT scans for diagnosing SBO. A single abdominal x-ray image should never be the sole method to diagnose SBO and must be correlated with the clinical context. Furthermore, Cheng et al[45,46] used CNN to analyze abdominal radiographs for diagnosing SBO. In their first study, 2210 abdominal radiographs were evaluated by using a CNN diagnostic model and showed a sensitivity of 83.8% and specificity of 68.1%.[45] In their second study, abdominal radiographs were increased to 7768, which resulted in the increase of diagnostic sensitivity and specificity to 99.4% and 99.9%, respectively.[46] These two studies suggested that expanding the number of radiographs has significantly improved the accuracy of SBO detection by CNNs.[45,46] Another recent retrospective cohort study from South Korea developed a 3D CNN model to detect high-risk patients with acute SBO, based on abdominal CT images.[47] The study analyzed 578 cases from 250 normal subjects, 209 high-grade SBOs, and 119 low-grade SBOs, based on over 38,000 CT images. The 3D CNN model detected high-risk acute SBO with high efficiency and accuracy. However, the study was done in a single tertiary center with a potential selection bias. Additionally, the study findings are relatively weak in terms of usability and generalizability because the model lacks external validation.

DL has also been studied in the detection of colorectal polyps and cancers. One study from Germany that used DL was able to differentiate between benign (regular mucosa or hyperplastic) and premalignant (adenoma) colorectal polyps detected on CT colonography in correspondence to histopathology, with an AUC of 0.83.[48] A limitation of this study was the small sample size, and the study was only applicable to polyps detected clearly on CT colonography, resulting in a potential selection bias. Optical coherence tomography (OCT) has been used to diagnose retinal diseases, but recent research suggests it could detect colorectal cancer (CRC). A study from Washington University in Saint Louis reported the first study to differentiate normal colorectal tissues from neoplastic ones by using DL-based pattern recognition OCT.[49] The proposed CNN model achieved 100% sensitivity and 99.7% specificity with an AUC of 0.998, suggesting a promising potential for real-time diagnosis of CRCs. Although this study showed promising results, the DL model was tested on a limited number of tumors that had received radiation and chemotherapy treatments. Table 1 summarizes multiple studies on DL application in the detection of different GI lesions on radiologic images.

Table 1

Multiple studies on deep learning application in radiologic imaging for detecting different gastrointestinal lesions

Multiple studies on deep learning application in radiologic imaging for detecting different gastrointestinal lesions
Multiple studies on deep learning application in radiologic imaging for detecting different gastrointestinal lesions

AI is being increasingly used in radiologic imaging to predict the response of various GI conditions to therapy. DL models can analyze large volumes of radiologic imaging data to identify patterns and features that may not be discernible to the human eye. These DL models can predict how patients with various GI conditions, such as esophageal, gastric, pancreatic, and colorectal cancers, respond to different therapies.

Advancements in pattern recognition, medical image analysis, and AI have laid the groundwork for the rapid evolution of radiomics, particularly in applications to esophageal carcinoma.[16] Several studies have incorporated PET and CT radiomics features with deep convolutional neural networks to predict treatment response in patients with esophageal carcinoma.[17,18] Several limitations of these current studies need to be highlighted. A significant challenge in this research area is the creation of a comprehensive public database with ample annotated medical imaging data to train numerous parameters in neural networks. Such a database would greatly enhance the ability to provide clinically relevant features, leading to improved model performance. Additionally, the limited number of tumor volumes available for these studies could result in overfitting, which might impair the algorithm’s ability to generalize to new, unseen test examples. Another group of researchers developed a pretreatment CT-based DL model for predicting the response to chemoradiation therapy in esophageal carcinoma.[23] This model was able to accurately and consistently stratify patients by their objective response to treatment, demonstrating its potential to guide clinical decisions and bridge the gap between precision medicine and biological imaging data.[23] A major challenge for this study was that the inner workings of the DL algorithm remain largely opaque, functioning as a “black box.” The processes of feature extraction and selection are automated and not transparent. This presents a broader challenge for physicians, who may find it difficult to intuitively and clearly interpret the results produced by the model. Consequently, understanding the reasons behind these results and determining whether specific features have a positive or negative impact can be quite challenging.

Researchers from China developed a DL model, Nomo-LDLM-2F, which uses multilesion and time series CT images to predict the benefits of anti-HER2 targeted therapy in stage IV gastric cancer.[50] The study found that patients with a low Nomo-LDLM-2F score derived significant survival benefits from anti-HER2 targeted therapy as compared to those with a high score (all p < 0.05). Thus, the Nomo-LDLM-2F score, based on these imaging techniques, shows promise for effectively predicting overall survival in patients with HER2-positive stage IV gastric cancer who are undergoing this targeted therapy.[50] Another study aimed to identify the diagnostic accuracy of an AI model and compare its performance against clinical methods and imaging methods (IMs) through a systematic review of head-to-head comparative studies.[51] Results from nine studies involving 3313 patients showed that the AI model had a pooled sensitivity of 0.75 and specificity of 0.77. The AI model was found to have higher sensitivity but lower specificity than the IM in most subgroups. The study concluded that AI is a viable tool for predicting the response of patients with gastric cancer to neoadjuvant chemotherapy. Particularly, CT-based DL models were effective in extracting tumor features and predicting responses.[51] However, in this review, there were three times as many male participants as female participants, with all articles originating from Asia. This uneven representation may have led to a model that is better suited to a specific population cohort, potentially introducing selection bias. Additionally, differences in baseline characteristics across the articles—such as cutoff values, cohort size and type, and gold standards—required extensive subgroup analysis to assess their impact on the stability of the conclusions. Moreover, none of the included articles reported the cutoff values, and most did not explain how these values were determined, which undermines the reliability of the diagnostic conclusions.

DL models have also been studied to predict the response of pancreatic tumors to neoadjuvant therapy. Watson et al[52] reported the first study to examine the ability of a DL model to predict pancreatic adenocarcinoma response to neoadjuvant therapy, based on pretreatment and posttreatment CT images. The pure CNN model achieved an AUC of 0.738, which increased to 0.785 when they incorporated a hybrid model that accounted for a 10% decrease in carbohydrate antigen (CA) 19–9 levels. Interestingly, the pure CA 19-9 model yielded an AUC of 0.564 for response prediction, suggesting that the pure CA 19-9 model alone essentially performs no better than random guesswork. However, this single-institution study included only patients who underwent resection, resulting in a potential selection bias. Recently in 2023, Shao et al[53] reported the first study to evaluate the feasibility of DL models in predicting the efficacy of neoadjuvant chemotherapy in pancreatic cancer, based on contrast-enhanced ultrasound videos. The DL models showed high performance with AUCs of 0.892 and 0.908. They were also able to accurately classify over 40% of the original videos that were misclassified by the naked eye.[53]

Another application of DL models is detecting CRC response to therapy. Zhang et al[54] developed a DL model to predict rectal cancer response to neoadjuvant chemoradiotherapy from diffusion keratosis and T2-weighted MRI images. The study found that radiologists had a higher error rate than the DL model for predicting pathologic complete response (26.9% and 24.8% for raters 1 and 2, respectively, versus 2.2% for the DL model). In addition, this study found that radiologists had lower error rates (13% and 14% for raters 1 and 2, respectively) with the assistance of the DL model. The main limitation of the study is that the DL model was based on the date of a single center so there is a need for multicenter investigation with a large dataset to generalize and validate the DL model. In another study, Shi et al[55] used radiomics and CNN based on pretreatment and mid-radiation follow-up MRI images to predict rectal cancer pathologic response to neoadjuvant radiation therapy. The study included 51 patients: 45 with pretreatment MRI, 41 with mid-radiation therapy MRI, and 35 with both sets. Using radiomics to predict pathologic complete response (pCR) versus non-pCR, the AUC was 0.8 for pretreatment, 0.82 for mid-radiation, and 0.86 for both MRI sets. Using CNN to predict pCR versus non-pCR, the AUC was 0.59 for pretreatment, 0.74 for mid-radiation, and 0.83 for both MRI sets. The CNN model’s prediction accuracy was lower than for radiomics owing to the small case number insufficient for training.[55] In another study that analyzed 1028 patients with metastatic CRC, Lu et al[56] developed a DL model that characterized tumor morphologic and size changes on CT images in response to anti–vascular endothelial growth factor therapies. Table 2 summarizes multiple studies on DL applications in the prediction of GI lesions’ response to therapy using radiologic imaging.

Table 2

Multiple studies on deep learning application in radiologic imaging for predicting gastrointestinal lesions’ response to therapy

Multiple studies on deep learning application in radiologic imaging for predicting gastrointestinal lesions’ response to therapy
Multiple studies on deep learning application in radiologic imaging for predicting gastrointestinal lesions’ response to therapy

DL models have shown great promise in predicting the prognosis and survival of various GI cancers, including esophageal, gastric, pancreatic, and colorectal. In addition, DL models have been studied to predict the severity of GI conditions such as Crohn disease and acute pancreatitis.

A study from Taiwan trained a 3D CNN model with PET scans to predict esophageal cancer prognosis with acceptable accuracy (AUC = 0.738).[57] One limitation in this study is the inherent opacity of DL networks. The processes of defining, extracting, and selecting features are automated and occur implicitly, making the imaging characteristics they assess quite obscure. This lack of transparency contrasts sharply with the clearly defined radiomic features developed by experts. For instance, morphometric measures like core muscle size from cross-sectional imaging have been linked to sarcopenia, which independently predicts clinical outcomes in various GI cancers. In contrast, CNN classification might rely on features related to tumors or nontumors. Another multicenter retrospective study involved 411 patients with pathologically confirmed esophageal squamous cell carcinoma (ESCC) from two hospitals.[58] Researchers developed radiomics signatures for each feature set and combined these with clinical risk factors to create multiple radiomics models. The results showed that the integrated radiomics model outperformed traditional clinical factors and existing methods, achieving C-statistics of 0.875, 0.874, and 0.840 in the development, internal validation, and external validation cohorts, respectively. Nomogram and decision curve analysis confirmed the model’s clinical utility, indicating it can effectively predict lymph node metastasis in patients with ESCC and assist in treatment planning.[58] This study was limited because authors used 2D features extracted from the maximum tumor rather than 3D features. Additionally, previous studies have demonstrated that genetic events, such as ZNF750 mutations, are linked to metastasis in patients with ESCC. In the future, when genetic data become available, incorporating these gene markers could enhance the predictive value of the model.

Survival prediction for esophageal cancer is crucial for personalized treatment planning, but traditional methods relying on handcrafted features from medical images are limited by incomplete medical knowledge, resulting in poor predictions. To overcome this, a novel DL-based framework has been proposed, featuring a 3D coordinate attention convolutional autoencoder (CACA) and an uncertainty-based jointly optimizing Cox model (UOCM). The CACA captures latent representations and encodes 3D spatial characteristics with precise positional information, whereas the UOCM jointly optimizes the autoencoder and survival prediction tasks to model interactions between patient features and clinical outcomes, predicting reliable hazard ratios. Extensive experiments on a dataset of 285 patients with esophageal cancer demonstrated that this method achieved a C-index of 0.72, outperforming existing state-of-the-art methods.[59]

Because current staging methods fall short in accurately predicting the risk of relapse in patients with gastric cancer, Jiang et al[60] devised a prognostic tool based on DL by using preoperative CT images to predict disease-free and overall survival and predict the potential benefits of adjuvant chemotherapy. The introduced AI model demonstrated enhanced prognostic precision, offering a valuable means to identify patients who are most likely to benefit from adjuvant chemotherapy. This study had several limitations. First, as a retrospective study, it is subject to inherent biases and unknown confounders, despite including many patients from two centers and adjusting for common clinicopathologic factors. Second, the use of adjuvant chemotherapy was not randomized; patients might have declined chemotherapy owing to personal preferences, concerns about side effects, or financial reasons. To address potential selection bias, a propensity score matching strategy was used. Finally, the model was developed and validated by using data from East Asian patients, and its generalizability to Western populations has yet to be established. Huang et al[61] similarly developed a CNN model for the diagnosis of occult peritoneal metastasis in patients with advanced gastric cancer, using preoperative CT imaging. This DL model exhibited a diagnostic accuracy surpassing that of a traditional clinical model constructed through logistic regression for comparison.[61] In this study, however, retrospective datasets were used to develop the CNN model, and a relatively small number of clinical factors were examined. Other factors, such as serologic tumor markers, which are not initially available from CT scans, may have contributed to any gaps in the data.

Another domain where DL was used is assessing the prognosis of HCC. For example, CEUS-based DL radiomics models have been developed to predict the prognosis of HCC that had been treated with radiofrequency ablation (RFA) or surgical resection (SR).[62] When applied to predict 2-year progression-free survival, these models achieved AUCs of 0.820 and 0.815 in training and validation cohorts for RFA, as well as 0.863 and 0.828 for SR.[62]

In 2023, Vezakis et al[63] developed a 3D CNN model to predict the prognosis of PDAC. The researchers trained the model using an external dataset to generate automated pancreas and tumor segmentations. They extracted radiomics features from CT images of 40 patients with PDAC. These features, along with the TNM system staging parameters and the patient’s age, were combined to predict survival. The results were promising, with a mean C-index of 0.731 for survival modeling and a mean accuracy of 0.76 in predicting 2-year survival. A potential limitation of this study is that the fully automated model may result in suboptimal radiomics feature extraction due to segmentation inaccuracies. Another limitation is the small dataset from a single institution. In another study, Chang et al[64] also proposed a 3D CNN to predict the presence of lymph node metastasis and the postoperative positive margin status, based on 881 preoperative CT scans of patients with PDAC. The study showed an accuracy of 90% for per-patient analysis and 75% for per-scan analysis for lymph node metastasis prediction, whereas for postoperative margin prediction, the study showed an accuracy of 81% for per-patient analysis and 76% for per-scan analysis. The major limitation of the study is that it included a small sample size of 110 patients and most were Caucasian, which makes the study less representative of the general PDAC population. DL models have also been studied for predicting the severity of acute pancreatitis. One recent study from China developed a combined clinical and imaging DL model that was able to accurately predict the severity of mild (AUC = 0.820) and severe (AUC = 0.920) acute pancreatitis by using nonenhanced CT images in a cohort of 783 patients.[65]

Meng et al[66] developed a DL model to classify bowel fibrosis in patients with Crohn disease, using CT enterography images. The study included 312 bowel segments of 235 patients with Crohn disease and compared the performance of the DL model to that of a radiomics model and two radiologists. The DL model accurately distinguished none to mild from moderate to severe bowel fibrosis, with an AUC of 0.811 in the total test cohort. In addition, the DL model performance was superior to radiologists (AUCs of 0.579 and 0.646) and not inferior to the radiomics model (AUC of 0.813) with much shorter total processing time than the radiomics model. However, most of the enrolled patients had moderate-severe bowl fibrosis, raising the potential of selection bias.

Accurate identification of lymph nodes on MRI is crucial for proper CRC staging, but it can be time-consuming. Several DL models have been studied for staging CRCs. Two studies from China reported the use of a faster region-based CNN (Faster R-CNN) model for identifying metastatic lymph nodes of rectal cancers in pelvic MRI images.[67,68] Their CNN model provided lymph node staging similar to that of radiologists, but the model’s average diagnostic time was 20 seconds compared to 600 seconds by radiologists. These results indicate that the Faster R-CNN model enables accurate and rapid diagnosis of lymph node metastasis in rectal cancer. Another study from China developed an automated lymph node detection and segmentation DL model based on multiparametric MRI images.[69] The study demonstrated that this model can accurately identify and segment lymph nodes from MRI images with higher performance than that of junior radiologists. In addition, it showed that the model’s average lymph node detection and segmentation time was 1.3 seconds compared to 200 seconds by radiologists.

Accurate prediction of circumferential resection margins (CRMs) is also essential for predicting the prognosis of CRC, as CRM invasion has a low survival rate. In a single-center retrospective study, Wang et al[70] reported the use of the Faster R-CNN model to identify rectal cancer–positive CRMs on high-resolution MRI images. The study showed that the Faster R-CNN model can automatically recognize an image in 0.2 seconds, with high accuracy and feasibility. However, as it is a single-center study with insufficient data, the DL model accuracy could not be validly assessed, including its ability to identify lymph nodes or extramural vascular invasion. Peritoneal carcinomatosis is considered a terminal stage in CRC with poor outcomes, so early diagnosis is essential to start treatment. In another study, a CT-based DL algorithm was developed and showed great potential for predicting synchronous peritoneal carcinomatosis in patients with CRC, with high accuracy, sensitivity, and specificity.[71] The limitations of the study included the small sample size and potential bias as all the data were collected from a single center in East Asia. Table 3 summarizes multiple studies on DL application in the prediction of GI lesions’ prognosis and survival using radiologic images.

Table 3

Multiple studies on deep learning application in radiologic imaging for predicting the prognosis of different gastrointestinal lesions

Multiple studies on deep learning application in radiologic imaging for predicting the prognosis of different gastrointestinal lesions
Multiple studies on deep learning application in radiologic imaging for predicting the prognosis of different gastrointestinal lesions

AI models have been studied to determine the Kirsten rat sarcoma viral oncogene (KRAS) mutation, DNA mismatch repair (MMR), and microsatellite instability (MSI) status in patients with CRC without the need for invasive biopsies. He et al[72] conducted a study to determine if a CT-based DL model could predict KRAS mutation status in colorectal cancer. The DL model had the potential to noninvasively estimate the presence of KRAS mutations, with an AUC of 0.9. In another recent study, Cao et al73] developed a DL model to predict the MMR status in patients with CRC, based on pretreatment CT images, which showed promising results with an AUC of 0.986 in the internal validation cohort and an AUC of 0.915 in the external validation cohort. Zhang et al[74] proposed MRI-based DL models to predict MSI in patients with rectal cancer. The imaging-based DL models were compared to the performance of a clinical model that recognizes MSI status using only clinical factors. The pure imaging-based DL model and combined DL model correctly classified 75% and 85.4% of MSI status with AUC values of 0.820 and 0.868, whereas the clinical model classified 37.5% of MSI status with an AUC value of 0.573. Table 4 summarizes two studies on DL applications in the prediction of CRC mutation status, using radiologic images.

Table 4

Two studies on deep learning application in radiologic imaging for predicting colorectal cancer mutation status

Two studies on deep learning application in radiologic imaging for predicting colorectal cancer mutation status
Two studies on deep learning application in radiologic imaging for predicting colorectal cancer mutation status

DL has enhanced the diagnostic applications of radiologic imaging in various GI conditions. This includes organ segmentation, lesion detection, prognosis, and response to therapy. For lesion detection, DL has been studied to detect esophageal, gastric, hepatic, pancreatic, and colorectal cancers. Additionally, DL models demonstrated potential in differentiating benign from malignant liver masses and regular mucosa or hyperplastic polyps from adenomas. DL models also help predict various GI lesions’ responses to different types of treatment, including surgery, radiotherapy, chemotherapy, and immunotherapy. DL has also shown promising results in predicting the prognosis, patient survival, and severity of various GI conditions such as pancreatic and colorectal cancers. It can also help predict mutation status in CRC, based on radiologic imaging. Some of the main recurring limitations to DL application in imaging was the need for large datasets for validation of results, potential selection bias, and the fact the training DL and ML models relies on annotated data to define the ground truth, making the imaging characteristics they assess quite obscure. Finally, the legal and ethical implications of using DL models in clinical decision-making is the problem. Radiologists must take responsibility for clinical decisions and cannot delegate this responsibility to AI developers. Despite some DL applications outperforming radiologists in certain tasks, the ultimate legal accountability for any diagnostic errors made by these applications rests with radiologists and clinicians. Further research is necessary to address potential challenges, errors, biases, and ethical considerations associated with the implementation of DL in radiologic imaging of GI diseases.

1.
Ogobuiro
I,
Gonzales
J,
Shumway
KR,
Tuma
F.
Physiology, gastrointestinal
. In: StatPearls.
StatPearls Publishing
;
2024
. Accessed Feb 6, 2024. www.ncbi.nlm.nih.gov/books/NBK537103/
2.
Sun
B,
Liu
J,
Li
S,
et al.
Imaging of gastrointestinal tract ailments
.
J Imaging
.
2023
;
9
(
6
):
115
.
3.
Thomas
LB,
Mastorides
SM,
Viswanadhan
NA,
et al.
Artificial intelligence: review of current and future applications in medicine
.
Fed Pract
.
2021
;
38
:
527
538
.
4.
Basu
K,
Sinha
R,
Ong
A,
Basu
T.
Artificial intelligence: how is it changing medical sciences and its future
?
Indian J Dermatol
.
2020
;
65
:
365
370
.
5.
Ghaffar Nia
N,
Kaplanoglu
E,
Nasab
A.
Evaluation of artificial intelligence techniques in disease diagnosis and prediction
.
Discov Artif Intell
.
2023
;
3
:
5
.
6.
Pettit
RW,
Fullem
R,
Cheng
C,
Amos
CI.
Artificial intelligence, machine learning, and deep learning for clinical outcome prediction
.
Emerg Top Life Sci
.
2021
;
5
:
729
745
.
7.
Moawad
AW,
Fuentes
DT,
ElBanan
MG,
et al.
Artificial Intelligence in diagnostic radiology: where do we stand, challenges, and opportunities
.
J Comput Assist Tomogr
.
2022
;
46
:
78
90
.
8.
Hosny
A,
Parmar
C,
Quackenbush
J,
et al.
Artificial intelligence in radiology
.
Nat Rev Cancer
.
2018
;
18
:
500
510
.
9.
Berbís
MA,
Aneiros-Fernández
J,
Mendoza Olivares
FJ,
et al.
Role of artificial intelligence in multidisciplinary imaging diagnosis of gastrointestinal diseases
.
World J Gastroenterol
.
2021
;
27
:
4395
4412
.
10.
Najjar
R.
Redefining radiology: a review of artificial intelligence integration in medical imaging
.
Diagnostics
.
2023
;
13
:
2760
.
11.
Kumar
R,
Khan
FU,
Sharma
A,
et al.
Recent applications of artificial intelligence in the detection of gastrointestinal, hepatic and pancreatic diseases
.
Curr Med Chem
.
2022
;
29
:
66
85
.
12.
Jiang
L,
Wu
Z,
Xu
X,
et al.
Opportunities and challenges of artificial intelligence in the medical field: current application, emerging problems, and problem-solving strategies
.
J Int Med Res
.
2021
;
49
:
03000605211000157
.
13.
Karimian
G,
Petelos
E,
Evers
SMAA.
The ethical issues of the application of artificial intelligence in healthcare: a systematic scoping review
.
AI Ethics
.
2022
;
2
:
539
551
.
14.
Ahmad
OF,
Stoyanov
D,
Lovat
LB.
Barriers and pitfalls for artificial intelligence in gastroenterology: ethical and regulatory issues
.
Tech Innov Gastrointest Endosc
.
2020
;
22
:
80
84
.
15.
Zou
KH,
Warfield
SK,
Bharatha
A,
et al.
Statistical validation of image segmentation quality based on a spatial overlap index
.
Acad Radiol
.
2004
;
11
:
178
189
.
16.
Koçak
B,
Durmaz
EŞ,
Ateş
E,
Kılıçkesmez
Ö.
Radiomics with artificial intelligence: a practical guide for beginners
.
Diagn Interv Radiol
.
2019
;
25
:
485
495
.
17.
Çorbacıoğlu
ŞK,
Aksel
G.
Receiver operating characteristic curve analysis in diagnostic accuracy studies: a guide to interpreting the area under the curve value
.
Turk J Emerg Med
.
2023
;
23
:
195
198
.
18.
Longato
E,
Vettoretti
M,
Di Camillo
B.
A practical perspective on the concordance index for the evaluation and selection of prognostic time-to-event models
.
J Biomed Inform
.
2020
;
108
:
103496
.
19.
Gibson
E,
Giganti
F,
Hu
Y,
et al.
Automatic multi-organ segmentation on abdominal CT with dense V-networks
.
IEEE Trans Med Imaging
.
2018
;
37
:
1822
1834
.
20.
Xue
J,
He
K,
Nie
D,
et al.
Cascaded MultiTask 3-D fully convolutional networks for pancreas segmentation
.
IEEE Trans Cybern
.
2021
;
51
:
2153
2165
.
21.
Laino
ME,
Ammirabile
A,
Lofino
L,
et al.
Artificial intelligence applied to pancreatic imaging: a narrative review
.
Healthcare (Basel)
.
2022
;
10
:
1511
.
22.
Zheng
H,
Chen
Y,
Yue
X,
et al.
Deep pancreas segmentation with uncertain regions of shadowed sets
.
Magn Reson Imaging
.
2020
;
68
:
45
52
.
23.
Panda
A,
Korfiatis
P,
Suman
G,
et al.
Two-stage deep learning model for fully automated pancreas segmentation on computed tomography: comparison with intra-reader and inter-reader reliability at full and reduced radiation dose on an external dataset
.
Med Phys
.
2021
;
48
:
2468
2481
.
24.
Kim
H,
Jung
J,
Kim
J,
et al.
Abdominal multi-organ auto-segmentation using 3D-patch-based deep convolutional neural network
.
Sci Rep
.
2020
;
10
:
6204
.
25.
Nainamalai
V,
Prasad
PJR,
Pelanis
E,
et al.
Evaluation of clinical applicability of automated liver parenchyma segmentation of multi-center magnetic resonance images
.
Eur J Radiol Open
.
2022
;
9
:
100448
.
26.
Fu
Y,
Mazur
TR,
Wu
X,
et al.
A novel MRI segmentation method using CNN-based correction network for MRI-guided adaptive radiotherapy
.
Med Phys
.
2018
;
45
:
5129
5137
.
27.
Chen
Y,
Ruan
D,
Xiao
J,
et al.
Fully automated multiorgan segmentation in abdominal magnetic resonance imaging with deep neural networks
.
Med Phys
.
2020
;
47
:
4971
4982
.
28.
Takeuchi
M,
Seto
T,
Hashimoto
M,
et al.
Performance of a deep learning-based identification system for esophageal cancer from CT images
.
Esophagus
.
2021
;
18
:
612
620
.
29.
Lin
C,
Guo
Y,
Huang
X,
et al.
Esophageal cancer detection via non-contrast CT and deep learning
.
Front Med
.
2024
;
11
:
1356752
.
30.
Xu
Y,
Zheng
B,
Liu
X,
et al.
Improving artificial intelligence pipeline for liver malignancy diagnosis using ultrasound images and video frames
.
Brief Bioinform
.
2022
;
24
.
31.
Ghoniem
R.
A novel bio-inspired deep learning approach for liver cancer diagnosis
.
Information
.
2020
;
11
:
80
.
32.
de Ridder
J,
de Wilt
JHW,
Simmer
F,
et al.
Incidence and origin of histologically confirmed liver metastases: an explorative case-study of 23,154 patients
.
Oncotarget
.
2016
;
7
:
55368
55376
.
33.
Hu
HT,
Wang
W,
Chen
LD,
et al.
Artificial intelligence assists identifying malignant versus benign liver lesions using contrast-enhanced ultrasound
.
J Gastroenterol Hepatol
.
2021
;
36
:
2875
2883
.
34.
Midya
A,
Chakraborty
J,
Srouji
R,
et al.
Computerized diagnosis of liver tumors from CT scans using a deep neural network approach
.
IEEE J Biomed Health Inform
.
2023
;
27
:
2456
2464
.
35.
Jia
W,
Li
F,
Cui
Y,
et al.
Deep learning radiomics model of contrast-enhanced CT for differentiating the primary source of liver metastases
.
Acad Radiol
.
2024
:
S1076-6332(24)00221-6
.
36.
Shi
W,
Kuang
S,
Cao
S,
et al.
Deep learning assisted differentiation of hepatocellular carcinoma from focal liver lesions: choice of four-phase and three-phase CT imaging protocol
.
Abdom Radiol (NY
).
2020
;
45
:
2688
2697
.
37.
Lee
JH,
Joo
I,
Kang
TW,
et al.
Deep learning with ultrasonography: automated classification of liver fibrosis using a deep convolutional neural network
.
Eur Radiol
.
2020
;
30
:
1264
1273
.
38.
Obaid
AM,
Turki
A,
Bellaaj
H,
et al.
Detection of gallbladder disease types using deep learning: an informative medical method
.
Diagnostics (Basel)
.
2023
;
13
:
1744
.
39.
Ponnoprat
D,
Inkeaw
P,
Chaijaruwanich
J,
et al.
Classification of hepatocellular carcinoma and intrahepatic cholangiocarcinoma based on multi-phase CT scans
.
Med Biol Eng Comput
.
2020
;
58
:
2497
2515
.
40.
Liu
KL,
Wu
T,
Chen
PT,
et al.
Deep learning to distinguish pancreatic cancer tissue from non-cancerous pancreatic tissue: a retrospective study with cross-racial external validation
.
Lancet Digit Health
.
2020
;
2
:
e303
e313
.
41.
Zhu
Z,
Xia
Y,
Xie
L,
et al.
Multi-scale coarse-to-fine segmentation for screening pancreatic ductal adenocarcinoma. In:
Shen
D,
Liu
T,
Peters
TM,
et al.
Eds.
Medical Image Computing and Computer Assisted Intervention – MICCAI 2019. Lecture Notes in Computer Science
.
Springer International Publishing
;
2019
:
3
12
.
42.
Alves
N,
Schuurmans
M,
Litjens
G,
et al.
Fully automatic deep learning framework for pancreatic ductal adenocarcinoma detection on computed tomography
.
Cancers (Basel)
.
2022
;
14
:
376
.
43.
Cao
K,
Xia
Y,
Yao
J,
et al.
Large-scale pancreatic cancer detection via non-contrast CT and deep learning
.
Nat Med
.
2023
;
29
:
3033
3043
.
44.
Kim
D,
Wit
H,
Thurston
M,
et al.
An artificial intelligence deep learning model for identification of small bowel obstruction on plain abdominal radiographs
.
Br J Radiol
.
2021
;
94
:
20201407
.
45.
Cheng
PM,
Tejura
TK,
Tran
KN,
Whang
G.
Detection of high-grade small bowel obstruction on conventional radiography with convolutional neural networks
.
Abdom Radiol (NY
).
2018
;
43
:
1120
1127
.
46.
Cheng
PM,
Tran
KN,
Whang
G,
Tejura
TK.
Refining convolutional neural network detection of small-bowel obstruction in conventional radiography
.
Am J Roentgenol
.
2019
;
212
:
342
350
.
47.
Oh
S,
Ryu
J,
Shin
HJ,
et al.
Deep learning using computed tomography to identify high-risk patients for acute small bowel obstruction: development and validation of a prediction model: a retrospective cohort study
.
Int J Surg
.
2023
;
109
:
4091
4100
.
48.
Wesp
P,
Grosu
S,
Graser
A,
et al.
Deep learning in CT colonography: differentiating premalignant from benign colorectal polyps
.
Eur Radiol
.
2022
;
32
:
4749
4759
.
49.
Zeng
Y,
Xu
S,
Chapman
WC,
et al.
Real-time colorectal cancer diagnosis using PR-OCT with deep learning
.
Theranostics
.
2020
;
10
:
2587
2596
.
50.
He
M,
Chen
Zi-fan,
Liu
S,
et al.
Deep learning model based on multi-lesion and time series CT images for predicting the benefits from anti-HER2 targeted therapy in stage IV gastric cancer
.
Insights Imaging
.
2024
;
15
:59:
1363812
.
51.
Bao
Z,
Du
J,
Zheng
Y,
et al.
Deep learning or radiomics based on CT for predicting the response of gastric cancer to neoadjuvant chemotherapy: a meta-analysis and systematic review
.
Front Oncol
.
2024
;
14
.
52.
Watson
MD,
Baimas-George
MR,
Murphy
KJ,
et al.
Pure and hybrid deep learning models can predict pathologic tumor response to neoadjuvant therapy in pancreatic adenocarcinoma: a pilot study
.
Am Surg
.
2021
;
87
:
1901
1909
.
53.
Shao
Y,
Dang
Y,
Cheng
Y,
et al.
Predicting the efficacy of neoadjuvant chemotherapy for pancreatic cancer using deep learning of contrast-enhanced ultrasound videos
.
Diagnostics (Basel)
.
2023
;
13
:
2183
.
54.
Zhang
XY,
Wang
L,
Zhu
HT,
et al.
Predicting rectal cancer response to neoadjuvant chemoradiotherapy using deep learning of diffusion kurtosis MRI
.
Radiology
.
2020
;
296
:
56
64
.
55.
Shi
L,
Zhang
Y,
Nie
K,
et al.
Machine learning for prediction of chemoradiation therapy response in rectal cancer using pre-treatment and mid-radiation multi-parametric MRI
.
Magn Reson Imaging
.
2019
;
61
:
33
40
.
56.
Lu
L,
Dercle
L,
Zhao
B,
Schwartz
LH.
Deep learning for the prediction of early on-treatment response in metastatic colorectal cancer from serial medical imaging
.
Nat Commun
.
2021
;
12
:
6654
.
57.
Yang
CK,
Yeh
JCY,
Yu
WH,
et al.
Deep convolutional neural network-based positron emission tomography analysis predicts esophageal cancer outcome
.
J Clin Med
.
2019
;
8
:
844
:
1548
.
58.
Wu
L,
Yang
X,
Cao
W,
et al.
Multiple level CT radiomics features preoperatively predict lymph node metastasis in esophageal cancer: a multicentre retrospective study
.
Front Oncol
.
2020
;
9
.
59.
Lin
Z,
Cai
W,
Hou
W,
et al.
CT-guided survival prediction of esophageal cancer
.
IEEE J Biomed Health Inform
.
2022
;
26
:
2660
2669
.
60.
Jiang
Y,
Jin
C,
Yu
H,
et al.
Development and validation of a deep learning CT signature to predict survival and chemotherapy benefit in gastric cancer: a multicenter, retrospective study
.
Ann Surg
.
2021
;
274
:
e1153
.
61.
Huang
Z,
Liu
D,
Chen
X,
et al.
Deep convolutional neural network based on computed tomography images for the preoperative diagnosis of occult peritoneal metastasis in advanced gastric cancer
.
Front Oncol
.
2020
;
10
:
601869
.
62.
Liu
F,
Liu
D,
Wang
K,
et al.
Deep learning radiomics based on contrast-enhanced ultrasound might optimize curative treatments for very-early or early-stage hepatocellular carcinoma patients
.
Liver Cancer
.
2020
;
9
:
397
413
.
63.
Vezakis
I,
Vezakis
A,
Gourtsoyianni
S,
et al.
An automated prognostic model for pancreatic ductal adenocarcinoma
.
Genes (Basel)
.
2023
;
14
:
1742
.
64.
Chang
J,
Liu
Y,
Saey
SA,
et al.
Machine-learning based investigation of prognostic indicators for oncological outcome of pancreatic ductal adenocarcinoma
.
Front Oncol
.
2022
;
12
:
895515
.
65.
Chen
Z,
Wang
Y,
Zhang
H,
et al.
deep learning models for severity prediction of acute pancreatitis in the early phase from abdominal nonenhanced computed tomography images
.
Pancreas
.
2023
;
52
:
e45
e53
.
66.
Meng
J,
Luo
Z,
Chen
Z,
et al.
Intestinal fibrosis classification in patients with Crohn’s disease using CT enterography-based deep learning: comparisons with radiomics and radiologists
.
Eur Radiol
.
2022
;
32
:
8692
8705
.
67.
Ding
L,
Liu
GW,
Zhao
BC,
et al.
Artificial intelligence system of faster region-based convolutional neural network surpassing senior radiologists in evaluation of metastatic lymph nodes of rectal cancer
.
Chin Med J (Engl)
.
2019
;
132
:
379
387
.
68.
Lu
Y,
Yu
Q,
Gao
Y,
et al.
Identification of metastatic lymph nodes in MR imaging with faster region-based convolutional neural networks
.
Cancer Res
.
2018
;
78
:
5135
5143
.
69.
Zhao
X,
Xie
P,
Wang
M,
et al.
Deep learning-based fully automated detection and segmentation of lymph nodes on multiparametric-mri for rectal cancer: a multicentre study
.
EBioMedicine
.
2020
;
56
:
102780
.
70.
Wang
D,
Xu
J,
Zhang
Z,
et al.
Evaluation of rectal cancer circumferential resection margin using faster region-based convolutional neural network in high-resolution magnetic resonance images
.
Dis Colon Rectum
.
2020
;
63
:
143
151
.
71.
Yuan
Z,
Xu
T,
Cai
J,
et al.
Development and validation of an image-based deep learning algorithm for detection of synchronous peritoneal carcinomatosis in colorectal cancer
.
Ann Surg
.
2022
;
275
:
e645
e651
.
72.
He
K,
Liu
X,
Li
M,
et al.
Noninvasive KRAS mutation estimation in colorectal cancer using a deep learning method based on CT imaging
.
BMC Med Imaging
.
2020
;
20
:
59
.
73.
Cao
W,
Hu
H,
Guo
J,
et al.
CT-based deep learning model for the prediction of DNA mismatch repair deficient colorectal cancer: a diagnostic study
.
J Transl Med
.
2023
;
21
:
214
.
74.
Zhang
W,
Yin
H,
Huang
Z,
et al.
Development and validation of MRI‐based deep learning models for prediction of microsatellite instability in rectal cancer
.
Cancer Med
.
2021
;
10
:
4164
4173
.
75.
Li
X,
Gao
H,
Zhu
J,
et al.
3D deep learning model for the pretreatment evaluation of treatment response in esophageal carcinoma: a prospective study (ChiCTR2000039279)
.
Int J Radiat Oncol Biol Phys
.
2021
;
111
:
926
935
.
76.
Ahn
JC,
Qureshi
TA,
Singal
AG,
et al.
Deep learning in hepatocellular carcinoma: current status and future perspectives
.
World J Hepatol
.
2021
;
13
:
2039
2051
.
77.
Ma
H,
Liu
ZX,
Zhang
JJ,
et al.
Construction of a convolutional neural network classifier developed by computed tomography images for pancreatic cancer diagnosis
.
World J Gastroenterol
.
2020
;
26
:
5156
5168
.
78.
Shi
YJ,
Zhu
HT,
Li
XT,
et al.
Histogram array and convolutional neural network of DWI for differentiating pancreatic ductal adenocarcinomas from solid pseudopapillary neoplasms and neuroendocrine neoplasms
.
Clin Imaging
.
2023
;
96
:
15
22
.
79.
Ferrari
R,
Mancini-Terracciano
C,
Voena
C,
et al.
MR-based artificial intelligence model to assess response to therapy in locally advanced rectal cancer
.
Eur J Radiol
.
2019
;
118
:
1
9
.

Competing Interests

Source of Support: None; Conflict of Interest: None.

This work is published under a CC-BY-NC-ND 4.0 International License.