The current state of artificial intelligence in robotic esophageal surgery
Abstract
Artificial intelligence (AI) is becoming increasingly utilized as a tool for physicians to optimize medical care and patient outcomes. The multifaceted approach to managing esophageal cancer provides a perfect opportunity for machine learning to support clinicians in all stages of management. Preoperatively, AI may aid gastroenterologists and surgeons in diagnosing and prognosticating premalignant or early-stage lesions. Intraoperatively, AI may also aid surgeons in identifying anatomic structures or minimize the learning curve for new learners. Postoperatively, machine learning algorithms can help predict complications and guide high-risk patients through recovery. While still evolving, AI holds promise in enhancing the efficiency and efficacy of multidisciplinary esophageal cancer care.
Keywords
INTRODUCTION
The treatment of esophageal malignancy is complex and requires coordination among a multidisciplinary team including gastroenterologists, medical and radiation oncologists, and thoracic surgeons. Similar to its applications in lung cancer, artificial intelligence (AI) and artificial neural networks (ANNs) are increasingly used to guide clinicians in all stages of esophageal cancer management, from diagnosis to non-operative management and surgical intervention[1-4]. The algorithms are trained on large amounts of clinical data and customized to solve unique clinical problems in all stages of cancer care. From a screening perspective, AI has been employed to supplement thorough endoscopic esophageal exams in order to identify premalignant and malignant conditions[4]. The utilization of machine learning (ML) with Barrett’s esophagus is in its infancy, but the growth of this application is imperative to avoid missed high-risk lesions and interval cancers. As more surgeons adopt the robotic-assisted minimally invasive esophagectomy (RAMIE) technique, AI is increasingly integrated into the operating room to reduce the surgical learning curve, enhance safety and efficiency, and improve postoperative outcomes. In this review, we will explore the current applications for AI in the realm of robotic esophageal surgery, which, while still in its infancy, has shown great potential as a tool for esophageal surgeons.
SCREENING AND PREOPERATIVE DECISION MAKING
Screening and early detection of esophageal premalignant lesions is imperative for early endoscopic and surgical treatments. AI has been applied to diagnostic upper endoscopies in aiding endoscopists in early detection and screening for premalignant and malignant esophageal diseases. Applications include computer-aided detection based on endoscopic images, deep learning algorithms of histologic specimens, and real-time video analysis. These AI algorithms demonstrate high sensitivity in identifying high-risk esophageal lesions and may enhance the traditional esophageal exam using high-definition white light and narrow band imaging[5,6]. Miss rates for esophageal adenocarcinoma and Barrett’s esophagus with existing biopsy strategies are estimated to be as high as > 20% and 50%, respectively. Narrow band imaging and more advanced imaging techniques such as chromoendoscopy have improved diagnostic accuracy for endoscopists. However, they require greater expertise[6]. Deep learning algorithms have shown promise as a novel adjunct to endoscopists in identifying high-risk esophageal lesions.
Computer-aided detection systems trained on large subsets of white light endoscopy images were able to identify Barrett’s esophagus with exceptionally high sensitivity and specificity with near-perfect localization of the disease[7-9]. A similar approach has been applied using ANN analysis of endoscopic videos and yielded similar detection rates of esophageal dysplasia[10]. ML algorithms applied to histologic specimen slides can identify and differentiate non-dysplastic Barrett’s esophagus, low-grade dysplasia, and high-grade dysplasia with > 90% sensitivity and specificity[11,12]. Similar approaches have been applied to the detection of malignancy with some neural network systems shown to detect gastroesophageal junction cancer on traditional white light endoscopy images with 66% accuracy, compared to an accuracy of 63% when analyzed by board-certified expert endoscopists[13]. AI-based tools have also been employed to ensure quality and consistency during screening endoscopy by providing automatic image capture and blind-spot recognition during screening routine endoscopy. In a randomized controlled trial, 153 patients were randomized to include real-time quality improvement system during routine screening endoscopy compared to 150 routine controls. In the experimental group, the quality of screening endoscopy was significantly improved including an approximate 15% reduction in blind spot rate [95% confidence interval (CI) -19.23 to -11.54][14]. This quality improvement system, called WISENSE, employed deep convolutional neural networks (DCNN) and deep reinforcement learning (DRL) trained on over 34,000 EGD images to classify gastric images into specific sites. Training this model involved first testing DCNN on still images and eventually integrating DCNN and DRL for testing on real EGD videos. After this initial training process, the system was formally tested in a randomized control trial.
As medical care has become increasingly individualized, AI has been integrated into the medical decision-making process for staging, prognosticating and treating premalignant and malignant esophageal lesions. Current estimations for progression of Barrett’s esophagus and surveillance intervals have been based on previously published large studies. In part due to interobserver variability in endoscopic and histologic evaluation, interval cancers have been seen with what would typically be considered “low risk” lesions. As a result of this alarming issue, a tissue systems pathology test (TissueCypher Barrett’s Esophagus Test, TSP-9) was developed and incorporates AI with biomarkers in an effort to predict the individual’s true risk of Barrett’s progression[15]. As patient response to therapy is variable, AI tools can guide multidisciplinary treatment plans and optimize patient outcomes. Rice et al., in their study using ML to analyze optimal treatment for esophageal cancer, found that 61% of patients who underwent esophagectomy alone received optimal treatment. In contrast, only 36% of patients who received neoadjuvant therapy were deemed to have received optimal therapy. These estimates were based on random forest technology in a sequential analysis. Optimal therapy was classified as the treatment modality that, after survival analysis using patient and cancer characteristics, maximized lifetime as measured by restricted mean survival time[16]. Another large study on 418 patients with esophageal adenocarcinoma undergoing surgery with curative intent utilized 65 variables from individualized patient data and used ANNs to predict one- and five-year survival. These variables spanned the full spectrum of care from symptoms at presentation to postoperative data such as tumor gene expression. Two ANNs were developed and compared against a linear discriminant analysis (LDA) to assess accuracy. Their results showed that the ANNs were more accurate than the LDA models and were superior to a model based solely on the tumor-node-metastasis (TNM) staging criteria when predicting survival[17].
Other groups have used AI to predict a patient’s responsiveness to chemotherapy. Using ANNs, real time polymerase chain reaction assays of pre- and post-treatment esophageal cancer specimens were analyzed for 17 genes in an attempt to predict histopathologic tumor response to chemoradiation. Not only did the analysis identify specific independent risk factors in the study population, it also outperformed univariate and multivariate analysis in terms of predicting response to treatment. The analysis of these gene expression arrays could predict tumor response to traditional neoadjuvant therapy with 85% accuracy[18]. Alternatively, ANNs trained on 18-fluorodeoxyglucose positron emission tomography (PET) scans from 107 patients with esophageal cancer were able to predict chemotherapy non-responders with > 80% sensitivity/specificity. Ypsilantis et al. used an “radiomics” approach where large amounts of quantitative data were extracted from pretreatment PET images to compile a tumor phenotype and employed an ANN to learn from intra-tumor slices seen on PET scans. Their model was able to extract PET image representations that could predict non-responders to treatment with 80.7% sensitivity and 81.6% specificity[19]. Oftentimes, non-responders to neoadjuvant chemotherapy have a worse prognosis compared with those individuals treated with upfront surgery[19]. This study highlights the potential use of AI in non-invasively predicting cancer treatment response which would allow for a more personalized approach to esophageal cancer patients.
INTRAOPERATIVE AND POSTOPERATIVE SURGICAL CARE
From a surgeon’s perspective, there are a number of hypothesized uses for AI applications to current surgical techniques including intraoperative support, surgical training, and postoperative care[20]. AI systems have potential for intraoperative utility by way of anatomic structure identification. As outlined in a meta-analysis by Anteby et al., ANNs have shown the ability to harness unlabeled laparoscopic footage to achieve precise tasks such as anatomy detection, instrument identification, action recognition, and surgical phase categorization[21]. Analysis of ANN performance in quantifying data from laparoscopic videos showed sensitivity as high as 95%, with a limitation being the heterogeneous nature of their pooled data from multiple different procedure types[22].
Similar techniques have been applied to RAMIEs where ANNs were trained to identify key anatomic structures including the azygos vein, superior vena cava, aorta, lung parenchyma, or the recurrent laryngeal nerve[22,23]. Sato et al. created an AI model to identify the recurrent laryngeal nerve after training on 2,000 static images extracted from 20 operative recordings. In each image, the recurrent laryngeal nerve was labeled by expert thoracic surgeons. Using the Dice coefficient to assess performance, the AI model outperformed general surgeons in identifying the recurrent laryngeal nerve and was only slightly less accurate than expert thoracic surgeons (Dice coefficient 0.58 vs. 0.62)[23]. The previous studies highlight real-time, autonomous intraoperative anatomical segmentation as a promising application for AI systems. While further research is undoubtedly necessary, existing literature suggests a strong potential for the expansion of AI and deep learning applications in the realm of robotic thoracic surgery.
AI has been utilized to minimize the steep learning curve of robotic surgery and bridge the gap between inexperienced and experienced surgeons. One study, which used Temporal Convolutional Networks for the Operating room (TeCNO), sought to develop an AI-based phase recognition system for RAMIE[24]. Video was incorporated from 31 RAMIE procedures and κ-fold cross-validation to train their model to recognize nine pre-determined surgical phases. By analyzing intraoperative recordings, their model was able to identify RAMIE phases with 84% accuracy. For trainees, automatic phase recognition provides objective data about surgical timing and efficiency for streamlined review and teaching. Intraoperatively, this information can be used to alert support staff to an operation’s current goals or needs. Automatic phase recognition is likely the first step in making truly autonomous platforms and sets the foundation for innovation and future robotic applications.
Postoperatively, there are multiple ML models that have been developed to predict patients at risk for complications following esophagectomy. ML algorithms can predict anastomotic leak rates with high sensitivity based on various patient characteristics with an area-under-the-receiver-operator curve (AUC) of 0.72-0.87[25,26]. Another model based on over 2,000 esophagectomy patients could predict early readmission with AUC of 0.72-0.74[27]. In a study of 864 patients with distal esophageal adenocarcinoma undergoing Ivor-Lewis Esophagectomy, an ANN was developed to predict clinically significant complications based on Clavien-Dindo classification. Based on 96 variables encompassing all phases of care, this model could predict Clavien-Dindo IIIa and above complications with an AUC of 0.67. It was also capable of discriminating between medical and surgical complications with AUCs of 0.70 and 0.66, respectively[28].
CONCLUSION
AI has emerged as a compelling tool for endoscopists, medical oncologists, and thoracic surgeons in the management of premalignant esophageal conditions and esophageal malignancy. While still in its infancy, it is clear that it will play an important role in assisting surgeons to complete RAMIEs safely and efficiently. Additionally, the multidisciplinary nature of esophageal malignancy offers multiple avenues for AI and ANN implementation including computer-aided detection on screening/surveillance endoscopy, predicting rates of progression of premalignant esophageal lesions or forecasting responses to medical-oncologic interventions. AI will be an important adjunct in optimizing patient outcomes by implementing predictive algorithms regarding preoperative and postoperative care. In concert with ongoing developments in robotic surgical platforms, current ML systems have set the foundation for future surgical innovation, which will continue to shape the field of thoracic surgery.
DECLARATIONS
Acknowledgments
The authors would like to thank the Departments of Surgery and Gastroenterology at Tufts Medical Center.
Authors’ contributions
Conceived the idea: Schumacher L, Holzwanger E
Performed the literature search and organized results: Poulos CM, Cassidy R, Khatibifar E
All authors discussed and contributed to the final manuscript.
Availability of data and materials
Not applicable.
Financial support and sponsorship
None.
Conflicts of interest
All authors declared that there are no conflicts of interest.
Ethical approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Copyright
© The Author(s) 2025.
REFERENCES
1. Bellini V, Valente M, Del Rio P, Bignami E. Artificial intelligence in thoracic surgery: a narrative review. J Thorac Dis. 2021;13:6963-75.
2. Seastedt KP, Moukheiber D, Mahindre SA, et al. A scoping review of artificial intelligence applications in thoracic surgery. Eur J Cardiothorac Surg. 2022;61:239-48.
3. Zhang YH, Guo LJ, Yuan XL, Hu B. Artificial intelligence-assisted esophageal cancer management: now and future. World J Gastroenterol. 2020;26:5256-71.
4. Kröner PT, Engels MM, Glicksberg BS, et al. Artificial intelligence in gastroenterology: a state-of-the-art review. World J Gastroenterol. 2021;27:6794-824.
5. Patel A, Arora GS, Roknsharifi M, Kaur P, Javed H. Artificial intelligence in the detection of Barrett’s esophagus: a systematic review. Cureus. 2023;15:e47755.
6. Meinikheim M, Messmann H, Ebigbo A. Role of artificial intelligence in diagnosing Barrett’s esophagus-related neoplasia. Clin Endosc. 2023;56:14-22.
7. de Groof J, van der Sommen F, van der Putten J, et al. The Argos project: the development of a computer-aided detection system to improve detection of Barrett’s neoplasia on white light endoscopy. United European Gastroenterol J. 2019;7:538-47.
8. Fockens KN, Jukema JB, Boers T, et al. Towards a robust and compact deep learning system for primary detection of early Barrett’s neoplasia: initial image-based results of training on a multi-center retrospectively collected data set. United European Gastroenterol J. 2023;11:324-36.
9. Abdelrahim M, Saiko M, Maeda N, et al. Development and validation of artificial neural networks model for detection of Barrett’s neoplasia: a multicenter pragmatic nonrandomized trial (with video). Gastrointest Endosc. 2023;97:422-34.
10. Hussein M, González-Bueno Puyal J, Lines D, et al. A new artificial intelligence system successfully detects and localises early neoplasia in Barrett’s esophagus by using convolutional neural networks. United European Gastroenterol J. 2022;10:528-37.
11. Faghani S, Codipilly DC, David Vogelsang, et al. Development of a deep learning model for the histologic diagnosis of dysplasia in Barrett’s esophagus. Gastrointest Endosc. 2022;96:918-25.e3.
12. Guleria S, Shah TU, Pulido JV, et al. Deep learning systems detect dysplasia with human-like accuracy using histopathology and probe-based confocal laser endomicroscopy. Sci Rep. 2021;11:5086.
13. Iwagami H, Ishihara R, Aoyama K, et al. Artificial intelligence for the detection of esophageal and esophagogastric junctional adenocarcinoma. J Gastroenterol Hepatol. 2021;36:131-6.
14. Wu L, Zhang J, Zhou W, et al. Randomised controlled trial of WISENSE, a real-time quality improving system for monitoring blind spots during esophagogastroduodenoscopy. Gut. 2019;68:2161-9.
15. Duits LC, Khoshiwal AM, Frei NF, et al; Barrett’s SURF LGD Study Pathologists Consortium. An automated tissue systems pathology test can standardize the management and improve health outcomes for patients with Barrett’s esophagus. Am J Gastroenterol. 2023;118:2025-32.
16. Rice TW, Lu M, Ishwaran H, Blackstone EH; Worldwide Esophageal Cancer Collaboration Investigators. Precision surgical therapy for adenocarcinoma of the esophagus and esophagogastric junction. J Thorac Oncol. 2019;14:2164-75.
17. Sato F, Shimada Y, Selaru FM, et al. Prediction of survival in patients with esophageal carcinoma using artificial neural networks. Cancer. 2005;103:1596-605.
18. Warnecke-Eberz U, Metzger R, Bollschweiler E, et al. TaqMan low-density arrays and analysis by artificial neuronal networks predict response to neoadjuvant chemoradiation in esophageal cancer. Pharmacogenomics. 2010;11:55-64.
19. Ypsilantis PP, Siddique M, Sohn HM, et al. Predicting response to neoadjuvant chemotherapy with PET imaging using convolutional neural networks. PLoS One. 2015;10:e0137036.
20. Gupta A, Singla T, Chennatt JJ, David LE, Ahmed SS, Rajput D. Artificial intelligence: a new tool in surgeon’s hand. J Educ Health Promot. 2022;11:93.
21. Anteby R, Horesh N, Soffer S, et al. Deep learning visual analysis in laparoscopic surgery: a systematic review and diagnostic test accuracy meta-analysis. Surg Endosc. 2021;35:1521-33.
22. den Boer RB, Jaspers TJM, de Jongh C, et al. Deep learning-based recognition of key anatomical structures during robot-assisted minimally invasive esophagectomy. Surg Endosc. 2023;37:5164-75.
23. Sato K, Fujita T, Matsuzaki H, et al. Real-time detection of the recurrent laryngeal nerve in thoracoscopic esophagectomy using artificial intelligence. Surg Endosc. 2022;36:5531-9.
24. Takeuchi M, Kawakubo H, Saito K, et al. Automated surgical-phase recognition for robot-assisted minimally invasive esophagectomy using artificial intelligence. Ann Surg Oncol. 2022;29:6847-55.
25. Zhao Z, Cheng X, Sun X, Ma S, Feng H, Zhao L. Prediction model of anastomotic leakage among esophageal cancer patients after receiving an esophagectomy: machine learning approach. JMIR Med Inform. 2021;9:e27110.
26. van de Beld JJ, Crull D, Mikhal J, et al. Complication prediction after esophagectomy with machine learning. Diagnostics. 2024;14:439.
27. Bolourani S, Tayebi MA, Diao L, et al. Using machine learning to predict early readmission following esophagectomy. J Thorac Cardiovasc Surg. 2021;161:1926-39.e8.
Cite This Article

How to Cite
Poulos, C. M.; Cassidy, R.; Khatibifar, E.; Holzwanger, E.; Schumacher, L. The current state of artificial intelligence in robotic esophageal surgery. Mini-invasive. Surg. 2025, 9, 6. http://dx.doi.org/10.20517/2574-1225.2024.42
Download Citation
Export Citation File:
Type of Import
Tips on Downloading Citation
Citation Manager File Format
Type of Import
Direct Import: When the Direct Import option is selected (the default state), a dialogue box will give you the option to Save or Open the downloaded citation data. Choosing Open will either launch your citation manager or give you a choice of applications with which to use the metadata. The Save option saves the file locally for later use.
Indirect Import: When the Indirect Import option is selected, the metadata is displayed and may be copied and pasted as needed.
About This Article
Special Issue
Copyright
Data & Comments
Data

Comments
Comments must be written in English. Spam, offensive content, impersonation, and private information will not be permitted. If any comment is reported and identified as inappropriate content by OAE staff, the comment will be removed without notice. If you have any queries or need any help, please contact us at [email protected].