Preprint
Review

A Comprehensive Guide to Artificial Intelligence in Endoscopic Ultrasound

Altmetrics

Downloads

180

Views

64

Comments

0

A peer-reviewed article of this preprint also exists.

Submitted:

01 May 2023

Posted:

02 May 2023

You are already at the latest version

Alerts
Abstract
Background: Endoscopic Ultrasound (EUS) is widely used for the diagnosis of bilio-pancreatic and gastrointestinal (GI) tract diseases, for the evaluation of subepithelial lesions, and for sampling of lymph nodes and solid masses lo-cated next to the GI tract. The role of Artificial Intelligence in healthcare in growing. This review aimed to provide an overview of the current state of AI in EUS from imaging to pathological diagnosis and training. Methods: AI algorithms can assist in lesion detection and characterization in EUS by analysing EUS images and identifying suspicious areas that may require further clinical evaluation or biopsy sampling. Deep learning tech-niques, such as convolutional neural networks (CNNs), have shown great potential for tumour identification and subepithelial lesion (SEL) evaluation by extracting important features from EUS images and using them to classify or segment the images. Results: AI models with new features can increase the accuracy of diagnoses, provide faster diagnoses, identify subtle differences in disease presentation that may be missed by human eyes, and provide more information and in-sights into disease pathology. Conclusions: The integration of AI in EUS images and biopsies has the potential to improve the diagnostic accura-cy, leading to better patient outcomes and to a reduction of repeated procedures in case of non-diagnostic biopsies.
Keywords: 
Subject: Medicine and Pharmacology  -   Gastroenterology and Hepatology

1. Introduction

Endoscopic ultrasound (EUS) has revolutionized the field of gastrointestinal (GI) endoscopy by providing high-resolution imaging of the gastrointestinal tract and adjacent anatomical structures. EUS has been widely used for the diagnosis of bilio-pancreatic diseases, staging of GI tract tumours, evaluation of subepithelial lesions, and sampling of lymph nodes and solid masses [1]. EUS-guided fine-needle aspiration (FNA) and biopsy (FNB) have enabled the diagnosis of various malignancies and have greatly improved patient outcomes [2].
However, the accuracy of EUS-guided FNA and FNB largely depends on the skills and experience of the endoscopist and of the pathologist. In recent years, artificial intelligence (AI) has emerged as a promising tool for improving the accuracy and efficiency of EUS-guided tissue sampling and pathological diagnosis [3].
AI refers to the use of computer algorithms to analyse large amounts of data and identify patterns or make predictions. In healthcare, AI has been applied to various tasks, including image recognition, natural language processing, and clinical decision-making [4]. AI algorithms can analyse EUS images and assist with the interpretation of findings, as well as predict the pathological diagnosis of tissue samples obtained by EUS-guided FNA and FNB.
The objective of this review article is to provide an overview of the current state of AI in EUS imaging and pathology on the final pathological diagnosis. We will discuss the various AI techniques used for EUS image analysis and pathological diagnosis, their strengths and limitations, and their potential impact on clinical practice. A session will also be dedicated to the application of AI in training program to improve the knowledge of EUS and the recognition of anatomical structures.

2. AI Algorithms and Image Acquisition

The use of artificial intelligence (AI) in EUS image interpretation has shown great potential for improving the accuracy and efficiency of the diagnostic process. AI algorithms can be divided into two main categories: deep learning techniques and machine learning techniques.
Deep learning techniques involve the use of neural networks to learn and recognize patterns in EUS images (Figure 1). These networks are composed of multiple layers of interconnected nodes that allow for the processing of large amounts of data. Convolutional neural networks (CNNs) are a commonly used deep learning technique for image analysis in healthcare [5]. They are designed to identify and extract important features from EUS images and use them to classify or segment the images.
Figures, Tables and Schemes
Figure 1. By combining recognized EUS-image features for pancreatic lesion diagnosis with measurements of non-Euclidean anatomical features, significant progress can be made in distinguishing diverse sub-types that have varying outcomes, A, B, and C. The utilization of fractal geometry, specifically the surface fractal dimension, as a measure of the space-filling property of a irregularly shaped structure, can be effectively merged as a feature within an AI-based neuronal network classification system, to achieve a more precise anatomical classifier system.
Figure 1. By combining recognized EUS-image features for pancreatic lesion diagnosis with measurements of non-Euclidean anatomical features, significant progress can be made in distinguishing diverse sub-types that have varying outcomes, A, B, and C. The utilization of fractal geometry, specifically the surface fractal dimension, as a measure of the space-filling property of a irregularly shaped structure, can be effectively merged as a feature within an AI-based neuronal network classification system, to achieve a more precise anatomical classifier system.
Preprints 72390 g001
Machine learning techniques, on the other hand, involve the use of algorithms that can learn from data and make predictions based on that learning. These techniques can be supervised, unsupervised, or semi-supervised. Supervised learning involves the use of labelled data to train an algorithm to recognize patterns in EUS images. Unsupervised learning involves the use of unlabelled data to discover patterns and relationships in the data. Semi-supervised learning combines both supervised and unsupervised learning [6]. In addition to image interpretation, AI can also be used to enhance the acquisition of EUS images. This includes automatic segmentation and image quality improvement.
Automatic segmentation involves the use of AI algorithms to identify and separate different structures in EUS images [7]. This can help to improve the accuracy and efficiency of EUS-guided procedures by providing better visualization of the target area. For example, AI algorithms can be used to automatically segment the pancreas or the lymph nodes in EUS images, allowing for more precise targeting during EUS-guided biopsy [8].
Image quality improvement involves the use of AI algorithms to enhance the clarity and resolution of EUS images. This can help to improve the accuracy of image interpretation and diagnosis. For example, AI algorithms can be used to reduce noise, improve contrast, and sharpen edges in EUS images. AI-enhanced image quality can also help to reduce the variability in image quality between different endoscopists and ultrasound machines, improving the consistency of diagnosis and treatment [9].

3. Lesion Detection and Characterization

3.1. Tumour identification

Tumour identification is a crucial step in the diagnosis and staging of GI neoplasia, especially in case of pancreatic cancers that may be isoechoic with the surrounding parenchyma or may be hidden by signs of chronic pancreatitis. AI algorithms can assist with tumour identification by analysing EUS images and identifying suspicious areas that may require biopsy sampling and microscopy observation or further clinical evaluation [10]. Deep learning techniques, such as CNNs, have shown great potential for tumour identification by extracting important features from EUS images and using them to classify or segment the images [11]. Supervised machine learning techniques can also be used to train AI algorithms to recognize specific tumour features, such as shape, size, and vascularity [12,13]. A recent meta-analysis of 10 studies, which involved 1871 patients, evaluated the diagnostic accuracy of AI applied to EUS in detecting pancreatic cancer, (Table 1). The results showed that AI had a high diagnostic sensitivity of 0.92 and specificity of 0.9, with an area under the summary receiver operating characteristics (SROC) curve of 0.95 and a diagnostic odds ratio of 128.9 [14]. These findings suggest that AI-assisted EUS could become an essential tool for the computer-aided diagnosis of pancreatic cancer. However, a relatively small number of studies and enrolled patients make generalizing difficult, and further research is needed to validate these results on a larger scale.
It is known that there are several advantages of incorporating new features in classifying pathological changes using AI. Among these: a) increased accuracy: the addition of new features to AI models can increase their accuracy in diagnosing and classifying pathological changes; b) faster diagnosis: AI models with new features can analyse large amounts of data quickly and accurately, allowing for earl diagnosis and better patient outcomes with reduced healthcare costs; c) personalized treatment: AI models with new features can identify subtle differences in disease presentation that may be missed by human eyes. This can lead to more personalized treatment plans that are tailored to the specific needs of each patient. d) improved decision-making: AI models with new features can provide clinicians with more information and insights into disease pathology, allowing them to make more informed decisions about patient care; and scalability: AI models can be easily scaled to analyse large amounts of data, making them ideal for analysing large datasets or monitoring patient health over time. This can lead to improved population health management and disease surveillance.
Diseases of various origins, such as inflammatory disorders, tumours, and functional diseases, can result in changes in the structural complexity and dynamic activity patterns. One way to quantify this structural complexity is by measuring the fractal dimension, among other parameters.
The human body is composed of intricate systems and networks, including its most complex structures. It is now widely accepted that the architecture of anatomical entities and their activities exhibit non-Euclidean properties. Natural fractals, including those found in anatomy, possess four distinct characteristics: a) irregular shape, b) statistical self-similarity, c) non-integer or fractal dimension, and d) scaling properties that depend on the scale of measurement. As anatomical structures do not conform to regular Euclidean shapes, their dimensions are expressed as non-integer values between two integer topological dimensions. Fractal geometry has been shown to be useful in evaluating the geometric complexity of anatomic and imaging patterns observed in both benign and malignant masses (Figure 1).
Recently, Carrara et al., have introduced a new estimator, called the surface fractal dimension, to evaluate the complexity of EUS-Elastography images in differentiating solid pancreatic lesions [15]. The study showed that the surface fractal dimension can distinguish malignant tumours from NETs, unaffected tissues surrounding malignant tumours from NETs and NETs from inflammatory lesions. This study highlights the importance of incorporating fractal analysis into AI algorithms for the diagnosis and categorization of the diverse array of pancreatic lesions.

3.2. Subepithelial lesion evaluation

Subepithelial lesions (SELs) are a common indication to perform EUS, and their diagnosis can be challenging. AI algorithms can assist with SELs evaluation by analysing EUS images and identifying suspicious lesions that may require biopsy. Deep learning techniques, such as CNNs, have shown great potential for SELs evaluation by extracting important features from EUS images and using them to classify or segment the images. The study by Hirai et al suggests that an AI system has higher diagnostic performance than experts in differentiating SELs on EUS images [16]. The AI system's accuracy for classifying five different types of SELs was 86.1%, which was significantly better than that of all endoscopists. In particular, the sensitivity and accuracy of the AI system for detecting gastrointestinal stromal tumours (GISTs) were higher than those of all endoscopists. These findings suggest that AI technology can be a valuable tool to assist in the diagnosis of SELs on EUS images, and may help improve clinical decision-making [16].

3.3. Diagnostic accuracy

The diagnostic accuracy of AI algorithms may be affected by several factors, such as the quality of the EUS images, the size and location of the lesion, and the expertise of the endoscopists and the pathologists. The results of previous studies have been controversial. In recent meta-analysis Xiao et al. identified seven studies to assess the diagnostic accuracy of AI-based EUS in distinguishing GISTs from other SELs [17]. The combined sensitivity and specificity of AI-based EUS were 0.93 and 0.78, respectively, with an overall diagnostic odds ratio of 36.74 and an area under the summary receiver operating characteristic curve (AUROC) of 0.94. These results suggest that AI-based EUS showed high diagnostic ability in differentiating GISTs from other SELs and could potentially set a premise for adapting diagnostic capabilities of other disease under EUS.

3.4. Clinical impact and limitations

The clinical impact of AI algorithms for lesion detection and characterization in EUS-guided pathological diagnosis is still under investigation. However, studies have reported that AI algorithms can improve the accuracy and efficiency of EUS-guided procedures, reduce the need for unnecessary biopsies, and assist with treatment planning [18]. AI algorithms can also help to reduce the inter-observer variability in lesion detection and characterization between different endoscopists [19]. However, the implementation of AI algorithms in clinical practice may be limited by several factors, such as the availability and cost of AI software, the need for specialized training, concerns about data privacy and security and most importantly the need for larger studies to establish the accuracy of such systems.

4. Digital Histopathological Diagnosis

The advancement of digital pathology has revolutionized the field of pathology by enabling the acquisition, management, and interpretation of pathological information in a digital format. This transition has been fuelled by advances in whole slide imaging (WSI) technology, which allows for the digitization of glass slides at high resolution [20]. The adoption of digital pathology offers numerous advantages, such as improved efficiency, reduced turnaround times, remote consultation, and easy access to archived cases [20]. The ability to stores pictures from tissue acquisition, as it happens for radiological imaging, puts the basis to share and use a lot of knowledge from pathological anatomy. Moreover, it sets the stage for the application of AI algorithms to facilitate and enhance diagnostic accuracy in the field of EUS.
WSI involves the scanning of entire histological glass slides to create high-resolution digital images. These digital images can be zoomed in or out and navigated as easily as a glass slide under a microscope. WSI technology has been instrumental in overcoming the challenges of data management in digital pathology, as it allows for efficient storage, retrieval, and sharing of massive amounts of image data [20]. Furthermore, WSI facilitates the standardization of image quality and provides an ideal platform for the application of AI algorithms to analyse the digital images, thereby supporting the development of novel diagnostic tools in endoscopic ultrasound.
In the context of histopathological image analysis, CNNs can be trained to automatically detect and classify the multifarious tissue structures, cellular patterns, and pathological alterations. CNNs are composed of multiple layers of interconnected “neurons”, including convolutional, pooling, and fully connected layers. The hierarchical structure of CNNs allows them to learn complex, high-level features from raw image data, thereby making them particularly suitable for the analysis of intricate histopathological images in endoscopic ultrasound [21,22,23].
In addition to CNNs, other machine learning techniques have been employed in histopathological image analysis for endoscopic ultrasound. These include support vector machines (SVM), random forests, and decision trees, among others. These algorithms can be used to extract and analyse peculiar features from histopathological images, such as texture, shape, and colour. By leveraging the strengths of multiple machine learning techniques, ensemble models can be created to improve overall performance and address potential limitations of individual algorithms [24].
The integration of AI algorithms, particularly CNNs, into the field of EUS has the potential to revolutionize the diagnosis and management of various GI tract disorders. As research progresses and these techniques become more refined, AI-based tools are expected to play an increasingly prominent role in the field of EUS and digital pathology.

5. AI in Pathological Diagnosis

5.1. AI applications in Anatomical Pathology

Tumor grading and staging are critical steps in the management of GI malignancies. Accurate tumor grading and staging are necessary for determining the appropriate treatment plan and predicting patient outcomes. AI algorithms can aid in tumor grading and staging by analyzing EUS images and identifying features that correspond to different tumor stages and grades (Figure 1). Deep learning techniques, such as CNNs, can identify subtle differences in tissue structure and morphology that may not be apparent to the human eye. For example, CNNs can potentially analyze EUS images of pancreatic cancer and differentiate between early-stage and advanced-stage tumors based on changes in tissue texture and vascularity [25,26,27]. AI algorithms can also predict the presence of lymph node metastasis by analyzing EUS images and identifying characteristic features, such as size, shape, and echogenicity, (Table 2). In a study by Săftoiu et al., contrast-enhanced harmonic EUS (CEH-EUS) with time-intensity curve (TIC) analysis and artificial neural network (ANN) processing were used to differentiate pancreatic carcinoma (PC) and chronic pancreatitis (CP) cases [28]. Parameters obtained through TIC analysis were able to differentiate between PC and CP cases and showed good diagnostic results in an automated computer-aided diagnostic system.
Prognostic and predictive biomarker analysis is essential for predicting patient outcomes and determining the most appropriate treatment plan. Prognostic biomarkers are associated with patient outcomes, such as survival or recurrence, while predictive biomarkers are associated with response to specific therapies [29]. Kurita et al., investigated the diagnostic ability of carcinoembryonic antigen (CEA), cytology, and AI using cyst fluid in differentiating malignant from benign pancreatic cystic lesions [30]. AI using deep learning showed higher sensitivity and accuracy in differentiating malignant from benign pancreatic cystic lesions than CEA and cytology.

5.2. Integrating AI in EUS-guided tissue acquisition

EUS-guided fine needle aspiration (EUS-FNA) and EUS-guided fine needle biopsy (EUS-FNB) are commonly used techniques for obtaining tissue samples for pathological diagnosis. The accuracy of EUS-guided tissue acquisition largely depends on the skills and experience of the operator and the quality and size of the tissue samples obtained can vary. The integration of AI in EUS-guided tissue acquisition has the potential to improve the accuracy and efficiency of the procedure, leading to better patient outcomes, (Table 3).
In a study by Inoue et al., an automatic visual inspection method based on supervised machine learning was proposed to assist rapid on-site evaluation (ROSE) for endoscopic ultrasound-guided fine needle aspiration (EUS-FNA) biopsy. The proposed method was effective in assisting on-site visual inspection of cellular tissue in ROSE for EUS-FNA, indicating highly probable areas including tumor cells [31].
Hashimoto et al., evaluated the diagnostic performance of their computer-aided diagnosis system using deep learning in EUS-FNA cytology of pancreatic ductal adenocarcinoma. The deep learning system showed promising results in improving diagnostic performance by step-by-step learning, with higher training volume and more efficient system development required for optimal CAD performance in ROSE of EUS-FNA cytology [32].
Ishikawa et al., developed a new AI-based method for evaluating EUS-FNB specimens in pancreatic diseases using deep learning and contrastive learning. The AI-based evaluation method using contrastive learning was comparable to macroscopic on-site evaluation (MOSE) performed by EUS experts and can be a novel objective evaluation method for EUS-FNB [18].
AI algorithms can potentially assist in EUS-FNA and EUS-FNB by providing real-time feedback to the endoscopist during the procedure. AI algorithms can analyze EUS images in real-time and provide guidance on the optimal location and depth of the needle insertion, as well as feedback on the quality of the tissue sample obtained. AI algorithms can also assist in the selection of the appropriate needle size and type based on the characteristics of the target lesion, such as diameter and location. The expectations of the limit of how AI can help improve variability in an endoscopic procedure can be the forefront need to find applicability from these systems to aid the outcome of a procedure [3].

6. Clinical validation of AI-enhanced pathological diagnosis

The accuracy of AI-enhanced pathological diagnosis has not been evaluated to a certain degree in the EUS setting. Traditional pathology approaches have been crucial in diagnosing diseases. Additionally, several qualitative and semi-quantitative grading and staging scoring systems have been widely proposed with even more accepted limitations. Semi-quantitative scores are not “measures” but only “labels” of severity [33]. An observer assigns "semiquantitative" scores to tissue changes based on predefined morphologic criteria. These scores are whole numbers and are less precise than quantitative scores because they only approximate relative changes. However, the advantage of semiquantitative scoring is that it can be applied to both macroscopic and microscopic tissue changes, generating strong data that can be statistically analyzed and used to evaluate experimental groups [34].
However, the development of digital pathology and AI solutions have allowed for more quantitative pathologic assessments, which are particularly useful in translational research [35]. These approaches provide invaluable opportunities for biomarker discovery and patient selection, aiding in the identification of optimal treatment regimens based on patient profiles [36]. Despite these benefits, challenges still exist in implementing AI-based methods in clinical settings. Specifically, in endoscopic setting, the incorporation of datasets and patient profiles to enhance the pathological diagnosis of a set disease is an expectation of many clinicians with the ever-growing boom of AI into clinical practice, that does not exist as of today.
The clinical impact of AI-enhanced pathological diagnosis has yet to be fully realized. However, the potential benefits include improved diagnostic accuracy, reduced inter-observer variability, and more efficient use of healthcare resources. AI-enhanced pathological diagnosis may also lead to the development of new biomarkers and treatment strategies for gastrointestinal malignancies. Limitations to the use of AI in pathological diagnosis exist, such as the accuracy of AI algorithms depends on the quality and quantity of the data used for training. The development of AI algorithms requires large amounts of data, which may be difficult to obtain for rare or uncommon gastrointestinal malignancies. Additionally, the use of AI in pathological diagnosis may raise ethical concerns regarding the role of technology in healthcare decision-making, a stage unexplored yet in many aspects of AI’s massive incorporation into healthcare.

7. EUS Training

The learning of EUS requires time and practice in a high-volume center, with an experienced endosonographer as teacher. This is due to the need to learn not only endoscopy but also to show excellent knowledge of ultrasound anatomy and of different district diseases [37].
As in other areas, in EUS AI ideally aims to improve the quality of the examination, helping to distinguish the type of lesions found, reducing procedural times, and providing real-time decision support and guidance in the execution of operative procedures [38]. It could also support the training of beginners, speeding up the learning process and reducing the need for a mentor. Furthermore, AI could provide quality control, standardizing performance between trainees and experts, (Table 4).
Often one of the first obstacles for the trainee in approaching EUS is the recognition of anatomical structures, as these are visualized in an unusual perspective, that varies according to the position of the endoscope and the station being examined. In 2021 at the ESGE days, a pilot study was awarded as the "best procedural innovation of the year": it proposed an AI system based on two convolutional neuronal networks that recognizes anatomical structures in both radial and linear EUS. This software has already achieved a recognition accuracy of 85% during the development phase [39].
Similarly, another study proposed the use of a CNN consisting of two branches, one for voice data and one for image data. EUS image labels were assigned based on simple verbal inputs indicating anatomical landmarks provided by experienced operators during the procedures. The prediction accuracy after the first system training reached 76% at the image level on a data set with five different labels. Moreover, voice tagging, instead of manual annotation, is very convenient in saving time [40]. These results are encouraging from the point of view of providing support to beginners, however data on actual improvement in the learning curve are scarce.
BP MASTER (pancreaticobiliary master) system was specifically designed by a joint collaboration (by Renmin Hospital of Wuhan University, Wuhan Union Hospital of Huazhong University of Science and Technology, Wuhan Puai Hospital and Wuhan EndoAngel Medical Technology Company) for training in EUS and examination quality control. The system includes a station classification model and a pancreas/abdominal aorta/portal confluence segmentation model. It was validated both internally and externally, reaching in the latter an accuracy of 82.4% in station classification and 0.72 Dice score in segmentation. The results of accuracy in classification and Dice score in segmentation were also comparable to that of experienced operators. In a crossover study, it was tested if the system could increase the accuracy of station recognition in trainees, showing an improvement from 67.2% to 78.4% (P < 0.01).
The same research group, subsequently, implemented the BP MASTER incorporating 4 deep convolutional neural networks (DCNN) in order to obtain additional functions: transducer location information, real-time operating instructions and to annotate the common bile duct anatomy and measure its caliber on freeze frame. At internal and external validations, the model confirmed its accuracy values, comparable to that of an expert.
Another crossover study was performed, aiming to evaluate the trainees’ accuracy improvement in interpreting the images when assisted by AI, which raised from 60.8% to 76.3% (P < 0.01) [41]. Finally, another study proposed to use AI to improve learning of the CH-EUS technique, particularly useful in identifying pancreatic masses and notoriously difficult to learn.
The system (CH-EUS MASTER), which includes a real-time acquisition and segmentation model, was adequately validated. A cross-trial was then conducted to assess the impact on trainees’ learning curve, using intersection over union (IoU) and time to lesion finding as indicators. Beginners who were supported by CH-EUS MASTER reported an improvement in mean IoU from 0.80 to 0.87 (P = 0.002) and a reduction in mean lesion identification times: from 22.75 to 17.98 seconds (P < 0.01) and from 34.21 to 25.92 seconds (P < 0.01) in the pancreatic body-tail and head-uncinate process, respectively [42].
CH-EUS MASTER seems also a valid tool in guiding EUS-FNA, with improvement in the first-pass diagnostic yield [43]. The development of more effective and articulate AI systems is desirable to allow trainees to speed up the training process and improve their performance. Ideally, integrating AI assistance systems with the use of simulators, up to virtual reality, could almost make the mentor unnecessary, but further data are necessary [44].

8. Future Directions and Conclusions

The integration of AI in EUS images and biopsy microscopy analysis has the potential to improve the diagnostic accuracy, leading to better patient outcomes and to a reduction of repeated procedures in case of non-diagnostic biopsies. AI algorithms can also aid in tumor grading, staging, and prognostic analysis. However, the clinical impact of AI-enhanced pathological diagnosis has yet to be established. Further research is needed to evaluate the long-term benefits and limitations of AI in EUS-imaging and biopsies.

Author Contributions

Conceptualization, S.C; data curation, K.K., K.P., M.T., F.G.; writing—original draft preparation, K.K., K.P., M.T., F.G.; writing—review and editing, S.C., F, G; supervision, S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

None.

Conflicts of Interest

Silvia Carrara is a consultant for Olympus. All other authors declare no conflict of interest.

References

  1. Friedberg, S.R.; Lachter, J. Endoscopic ultrasound: Current roles and future directions. World J Gastrointest Endosc 2017, 9, 499–505. [Google Scholar] [CrossRef] [PubMed]
  2. Sooklal, S.; Chahal, P. Endoscopic Ultrasound. Surg Clin North Am 2020, 100, 1133–1150. [Google Scholar] [CrossRef] [PubMed]
  3. Liu, E.; Bhutani, M.S.; Sun, S. Artificial intelligence: The new wave of innovation in EUS. Endosc Ultrasound 2021, 10, 79–83. [Google Scholar] [CrossRef] [PubMed]
  4. Yu, K.-H.; Beam, A.L.; Kohane, I.S. Artificial intelligence in healthcare. Nat Biomed Eng 2018, 2, 719–731. [Google Scholar] [CrossRef]
  5. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: an overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef]
  6. Erickson, B.J.; Korfiatis, P.; Akkus, Z.; Kline, T.L. Machine Learning for Medical Imaging. Radiogr Rev Publ Radiol Soc N Am Inc 2017, 37, 505–515. [Google Scholar] [CrossRef]
  7. Iwasa, Y.; Iwashita, T.; Takeuchi, Y.; Ichikawa, H.; Mita, N.; Uemura, S.; et al. Automatic Segmentation of Pancreatic Tumors Using Deep Learning on a Video Image of Contrast-Enhanced Endoscopic Ultrasound. J Clin Med 2021, 10, 3589. [Google Scholar] [CrossRef]
  8. Sinkala, M.; Mulder, N.; Martin, D. Machine Learning and Network Analyses Reveal Disease Subtypes of Pancreatic Cancer and their Molecular Characteristics. Sci Rep 2020, 10, 1212. [Google Scholar] [CrossRef]
  9. Simsek, C.; Lee, L.S. Machine learning in endoscopic ultrasonography and the pancreas: The new frontier? Artif Intell Gastroenterol 2022, 3, 54–65. [Google Scholar] [CrossRef]
  10. Bi, W.L.; Hosny, A.; Schabath, M.B.; Giger, M.L.; Birkbak, N.J.; Mehrtash, A.; et al. Artificial intelligence in cancer imaging: Clinical challenges and applications. CA Cancer J Clin 2019, 69, 127–157. [Google Scholar] [CrossRef]
  11. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  12. Murali, N.; Kucukkaya, A.; Petukhova, A.; Onofrey, J.; Chapiro, J. Supervised Machine Learning in Oncology: A Clinician’s Guide. Dig Dis Interv 2020, 4, 73–81. [Google Scholar] [CrossRef]
  13. Shao, D.; Dai, Y.; Li, N.; Cao, X.; Zhao, W.; Cheng, L.; et al. Artificial intelligence in clinical research of cancers. Brief Bioinform 2022, 23, bbab523. [Google Scholar] [CrossRef]
  14. Dumitrescu, E.A.; Ungureanu, B.S.; Cazacu, I.M.; Florescu, L.M.; Streba, L.; Croitoru, V.M.; et al. Diagnostic Value of Artificial Intelligence-Assisted Endoscopic Ultrasound for Pancreatic Cancer: A Systematic Review and Meta-Analysis. Diagn Basel Switz 2022, 12, 309. [Google Scholar] [CrossRef]
  15. Carrara, S.; Di Leo, M.; Grizzi, F.; Correale, L.; Rahal, D.; Anderloni, A.; et al. EUS elastography (strain ratio) and fractal-based quantitative analysis for the diagnosis of solid pancreatic lesions. Gastrointest Endosc 2018, 87, 1464–1473. [Google Scholar] [CrossRef]
  16. Hirai, K.; Kuwahara, T.; Furukawa, K.; Kakushima, N.; Furune, S.; Yamamoto, H.; et al. Artificial intelligence-based diagnosis of upper gastrointestinal subepithelial lesions on endoscopic ultrasonography images. Gastric Cancer Off J Int Gastric Cancer Assoc Jpn Gastric Cancer Assoc 2022, 25, 382–391. [Google Scholar] [CrossRef] [PubMed]
  17. Ye, X.H.; Zhao, L.L.; Wang, L. Diagnostic accuracy of endoscopic ultrasound with artificial intelligence for gastrointestinal stromal tumors: A meta-analysis. J Dig Dis 2022, 23, 253–261. [Google Scholar] [CrossRef] [PubMed]
  18. Ishikawa, T.; Hayakawa, M.; Suzuki, H.; Ohno, E.; Mizutani, Y.; Iida, T.; et al. Development of a Novel Evaluation Method for Endoscopic Ultrasound-Guided Fine-Needle Biopsy in Pancreatic Diseases Using Artificial Intelligence. Diagn Basel Switz 2022, 12, 434. [Google Scholar] [CrossRef] [PubMed]
  19. Parasher, G.; Wong, M.; Rawat, M. Evolving role of artificial intelligence in gastrointestinal endoscopy. World J Gastroenterol 2020, 26, 7287–7298. [Google Scholar] [CrossRef] [PubMed]
  20. Rodriguez, J.P.M.; Rodriguez, R.; Silva, V.W.K.; Kitamura, F.C.; Corradi, G.C.A.; de Marchi, A.C.B.; et al. Artificial intelligence as a tool for diagnosis in digital pathology whole slide images: A systematic review. J Pathol Inform 2022, 13, 100138. [Google Scholar] [CrossRef] [PubMed]
  21. Marya, N.B.; Powers, P.D.; Chari, S.T.; Gleeson, F.C.; Leggett, C.L.; Abu Dayyeh, B.K.; et al. Utilisation of artificial intelligence for the development of an EUS-convolutional neural network model trained to enhance the diagnosis of autoimmune pancreatitis. Gut 2021, 70, 1335–1344. [Google Scholar] [CrossRef] [PubMed]
  22. Marya, N.B.; Powers, P.D.; Fujii-Lau, L.; Abu Dayyeh, B.K.; Gleeson, F.C.; Chen, S.; et al. Application of artificial intelligence using a novel EUS-based convolutional neural network model to identify and distinguish benign and malignant hepatic masses. Gastrointest Endosc 2021, 93, 1121–1130.e1. [Google Scholar] [CrossRef] [PubMed]
  23. Oh, C.K.; Kim, T.; Cho, Y.K.; Cheung, D.Y.; Lee, B.-I.; Cho, Y.-S.; et al. Convolutional neural network-based object detection model to identify gastrointestinal stromal tumors in endoscopic ultrasound images. J Gastroenterol Hepatol 2021, 36, 3387–3394. [Google Scholar] [CrossRef] [PubMed]
  24. Zhang, M.-M.; Yang, H.; Jin, Z.-D.; Yu, J.-G.; Cai, Z.-Y.; Li, Z.-S. Differential diagnosis of pancreatic cancer from normal tissue with digital imaging processing and pattern recognition based on a support vector machine of EUS images. Gastrointest Endosc 2010, 72, 978–985. [Google Scholar] [CrossRef]
  25. Cazacu, I.M.; Udristoiu, A.; Gruionu, L.G.; Iacob, A.; Gruionu, G.; Saftoiu, A. Artificial intelligence in pancreatic cancer: Toward precision diagnosis. Endosc Ultrasound 2019, 8, 357–359. [Google Scholar] [CrossRef]
  26. Muhammad, W.; Hart, G.R.; Nartowt, B.; Farrell, J.J.; Johung, K.; Liang, Y.; et al. Pancreatic Cancer Prediction Through an Artificial Neural Network. Front Artif Intell 2019, 2. [Google Scholar] [CrossRef]
  27. Corral, J.E.; Hussein, S.; Kandel, P.; Bolan, C.W.; Bagci, U.; Wallace, M.B. Deep Learning to Classify Intraductal Papillary Mucinous Neoplasms Using Magnetic Resonance Imaging. Pancreas 2019, 48, 805–810. [Google Scholar] [CrossRef]
  28. Săftoiu, A.; Vilmann, P.; Dietrich, C.F.; Iglesias-Garcia, J.; Hocke, M.; Seicean, A.; et al. Quantitative contrast-enhanced harmonic EUS in differential diagnosis of focal pancreatic masses (with videos). Gastrointest Endosc 2015, 82, 59–69. [Google Scholar] [CrossRef]
  29. Echle, A.; Rindtorff, N.T.; Brinker, T.J.; Luedde, T.; Pearson, A.T.; Kather, J.N. Deep learning in cancer pathology: a new generation of clinical biomarkers. Br J Cancer 2021, 124, 686–696. [Google Scholar] [CrossRef]
  30. Kurita, Y.; Kuwahara, T.; Hara, K.; Mizuno, N.; Okuno, N.; Matsumoto, S.; et al. Diagnostic ability of artificial intelligence using deep learning analysis of cyst fluid in differentiating malignant from benign pancreatic cystic lesions. Sci Rep 2019, 9, 6893. [Google Scholar] [CrossRef]
  31. Inoue, H.; Ogo, K.; Tabuchi, M.; Yamane, N.; Oka, H. An automatic visual inspection method based on supervised machine learning for rapid on-site evaluation in EUS-FNA. 2014 Proc. SICE Annu. Conf. SICE; 2014; pp. 1114–1119. [Google Scholar] [CrossRef]
  32. Hashimoto, Y.; Ohno, I.; Imaoka, H.; Takahashi, H.; Mitsunaga, S.; Sasaki, M.; et al. Mo1296 RELIMINARY RESULT OF COMPUTER AIDED DIAGNOSIS (CAD) PERFORMANCE USING DEEP LEARNING IN EUS-FNA CYTOLOGY OF PANCREATIC CANCER. Gastrointest Endosc 2018, 87, AB434. [Google Scholar] [CrossRef]
  33. Bencze, J.; Szarka, M.; Kóti, B.; Seo, W.; Hortobágyi, T.G.; Bencs, V.; et al. Comparison of Semi-Quantitative Scoring and Artificial Intelligence Aided Digital Image Analysis of Chromogenic Immunohistochemistry. Biomolecules 2021, 12, 19. [Google Scholar] [CrossRef] [PubMed]
  34. Meyerholz, D.K.; Beck, A.P. Fundamental Concepts for Semiquantitative Tissue Scoring in Translational Research. ILAR J 2018, 59, 13–17. [Google Scholar] [CrossRef]
  35. Baxi, V.; Edwards, R.; Montalto, M.; Saha, S. Digital pathology and artificial intelligence in translational medicine and clinical practice. Mod Pathol Off J U S Can Acad Pathol Inc 2022, 35, 23–32. [Google Scholar] [CrossRef] [PubMed]
  36. Bera, K.; Schalper, K.A.; Rimm, D.L.; Velcheti, V.; Madabhushi, A. Artificial intelligence in digital pathology - new tools for diagnosis and precision oncology. Nat Rev Clin Oncol 2019, 16, 703–715. [Google Scholar] [CrossRef] [PubMed]
  37. Johnson, G.; Webster, G.; Boškoski, I.; Campos, S.; Gölder, S.K.; Schlag, C.; et al. Curriculum for ERCP and endoscopic ultrasound training in Europe: European Society of Gastrointestinal Endoscopy (ESGE) Position Statement. Endoscopy 2021, 53, 1071–1087. [Google Scholar] [CrossRef] [PubMed]
  38. Kuwahara, T.; Hara, K.; Mizuno, N.; Haba, S.; Okuno, N.; Koda, H.; et al. Current status of artificial intelligence analysis for endoscopic ultrasonography. Dig Endosc Off J Jpn Gastroenterol Endosc Soc 2021, 33, 298–305. [Google Scholar] [CrossRef] [PubMed]
  39. Robles-Medranda, C.; Oleas, R.; Valle, R.S.D.; Mendez, J.C.; Alcívar-Vásquez, J.M.; Puga-Tejada, M.; et al. ID: 3521955 APPLICATION OF ARTIFICIAL INTELLIGENCE FOR REAL-TIME ANATOMICAL RECOGNITION DURING ENDOSCOPIC ULTRASOUND EVALUATION: A PILOT STUDY. Gastrointest Endosc 2021, 93, AB221. [Google Scholar] [CrossRef]
  40. Bonmati, E.; Hu, Y.; Grimwood, A.; Johnson, G.J.; Goodchild, G.; Keane, M.G.; et al. Voice-Assisted Image Labeling for Endoscopic Ultrasound Classification Using Neural Networks. IEEE Trans Med Imaging 2022, 41, 1311–1319. [Google Scholar] [CrossRef]
  41. Yao, L.; Zhang, J.; Liu, J.; Zhu, L.; Ding, X.; Chen, D.; et al. A deep learning-based system for bile duct annotation and station recognition in linear endoscopic ultrasound. EBioMedicine 2021, 65, 103238. [Google Scholar] [CrossRef]
  42. Tang, A.; Gong, P.; Fang, N.; Ye, M.; Hu, S.; Liu, J.; et al. Endoscopic ultrasound diagnosis system based on deep learning in images capture and segmentation training of solid pancreatic masses. Med Phys 2023. [Google Scholar] [CrossRef] [PubMed]
  43. Tang, A.; Tian, L.; Gao, K.; Liu, R.; Hu, S.; Liu, J.; et al. Contrast-enhanced harmonic endoscopic ultrasound (CH-EUS) MASTER: A novel deep learning-based system in pancreatic mass diagnosis. Cancer Med 2023. [Google Scholar] [CrossRef] [PubMed]
  44. Khan, R.; Plahouras, J.; Johnston, B.C.; Scaffidi, M.A.; Grover, S.C.; Walsh, C.M. Virtual reality simulation training in endoscopy: a Cochrane review and meta-analysis. Endoscopy 2019, 51, 653–664. [Google Scholar] [CrossRef] [PubMed]
  45. Zhang, J.; Zhu, L.; Yao, L.; Ding, X.; Chen, D.; Wu, H.; et al. Deep learning-based pancreas segmentation and station recognition system in EUS: development and validation of a useful training tool (with video). Gastrointest Endo’s 2020, 92, 874–885.e3. [Google Scholar] [CrossRef]
Table 1. AI’s role in pancreatic cancer.
Table 1. AI’s role in pancreatic cancer.

Study

Sinkala M et al. 2020, South Africa [8]

Wazir M. et al.
2019, USA
[26]

Corral JE et al.
2019, USA
[27]

AI type

ANN

ANN

ANN

Topic

differentiation of molecular-genomic profile of PDAC subtypes (mRNA and DNA methylation models)

PDAC risk prediction based on personal health data

IPMN characterization on MRI

Study population

45 pancreatic cancer samples

898 patients diagnosed with pancreatic cancer

139 patients with histologically characterized IPMNs (due to pancreatectomy) and previous MRI images


Main results

AI in differentiation of two PDAC subtypes: overall classification Ac 100% for the mRNA-based model, 99% for the DNA methylation-model); model provides predictions of clinical response to chemotherapy

AI Sn and Sp in testing cohort: 80.7%, 80.7%; AUROC curve 0.85.

AI in detect dysplasia Sn and Sp: 92%, 52%.
Identification of high-grade dysplasia/ cancer: Sn and Sp 75% and 78%.
AI AUROC curves 0,78 (P = 0.90) vs
AUROC base on AGA criteria 0.76, AUROC based on Fukuoka criteria 0.77.
Table 2. Studies assessing diagnostic capabilities of AI systems in EUS.
Table 2. Studies assessing diagnostic capabilities of AI systems in EUS.

Study

Hirai K. et al.
2022, Japan 
[16]

Marya NB. et al. 2021, USA
[21]

Marya NB. et al.
2021, USA
[22]

Oh CK. et al.
2021, South Korea
[23]

Săftoiu A. et al.
2015, Denmark
[28]

Zhang MM. et al.
2010, China
[24]

AI type

CNN

CNN

CNN

CNN

ANN

SVM

Topic

differential diagnosis of SELs (five-category: GIST, leiomyoma, schwannoma, NET, ectopic pancreas)

enhance the diagnosis of AIP

identify and classify FLLs

differential diagnosis of SELs (GISTs and leiomyomas)

Differential diagnosis of PDAC and CP using CH-EUS and TIC analysis

differential diagnosis of PDAC from normal tissue (based on 29 pattern features)

Study population

16110 images, 631 examinations

583 patients, 1174461 still images from videos

256 patients, 210685 still images from videos

114 patients (with histologically confirmed gastric GIST), 376 still images

167 patients with PDAC or CP

216 patients (153 with PDAC, 63 without)


Main results

overall Ac: AI 86.1% vs expert endoscopy 68.0% (p<0,001); Sn, Sp, Ac of AI
in differentiating GISTs from non-GIST: 98.8%, 67.6%, 89.3% (better than expert endoscopist: Sn, Ac p< 0,001)

AI processed 955 frames/sec.
Sn and Sp for distinguishing AIP from PDAC: 90%, 93%.
Sn and Sp for distinguishing AIP from all studied conditions (PDAC, CP, NP): 90%, 85%.

AI autonomously locates FLLs in 92.0% of videos. Sn and Sp in classifying malignant FLLs on random still images: 90%, 71%.
Sn and Sp in classifying malignant FLLs on full-length videos: 100% and 80%.

AI Sn, Sp, Ac in per-image analysis: 95.6%, 82.1%, 91.2%
Sn, Sp, Ac in per-patient analysis 100.0%, 85.7%, 96.3% (better than expert endoscopist: Sn, Ac p<0,001)

AI Sn Sp PPV and NPV using TIC analysis on CH-EUS: 94.64%, 94.44%, 97.24%,89.47%

AI Ac, Sn, Sp, PPV and NPV for the diagnosis of pancreatic cancer: 97.98%, 94.32%, 99.45%, 98.65%, 97.77%.
Table 3. Studies assessing AI’s pathological diagnosis.
Table 3. Studies assessing AI’s pathological diagnosis.

Study

Ishikawa T. et al.
2022, Japan
[18]

Kurita Y et al.
2019, Japan,
[30]

Hashimoto Y. et al.
2018, Japan
[32]

Inoue H. et al.
2014, Japan
[31]

AI type

CNN

ANN

ANN

GMM

Topic

MOSE in pancreatic diseases

analysis of cyst fluid, cytology and EUS characteristics in differentiating malignant from benign pancreatic cysts.

ROSE in PDAC

AI automatic visual inspection method is proposed to assist ROSE

Study population

96 patients, 173 specimens

85 patients (59 surgical specimens, 26 EUS-guided FNA specimens)

500 images of cytology specimen (stained and in high definition)

\


Main results

Initial study: AI Ac 71.8% (vs MOSE performed by EUS experts 81.6%). Using contrastive learning: AI Sn, Sp, Ac: 90.34%, 53.5%, 84.39%, (vs 88.97%, 53.5%, 83.24% of EUS experts)

AI diagnostic ability in malignant cystic lesions: AUROC curve 0.966 (vs 0.719 for CEA, 0.739 for cytology).
AI Sn, Sp, Ac: 95.7%, 91.9%, 92.9% (vs. CEA Sn 60.9%, p = 0.021; cytology Sn 47.8% p = 0.001; CEA Ac 71.8%, p < 0.00; cytology Ac 85.9%, p =0.210).

AI Sn, Sp, Ac at the first learning stage: 78%, 60% 69%.
AI Sn, Sp, Ac at the second learning stage: 80%, 80%, 80%.

The AI method is reported as helpful for EUS-FNA in aiding ROSE, indicating areas highly likely to include tumor cells
Table 4. AI’s role in EUS training.
Table 4. AI’s role in EUS training.

Study

Tang A. et al. 2023, China
[43]

Tang A. et al.
2023, China
[42]

Bonmati E. et al.
2022, UK
[40]

Robles-Medranda C. et al.
2021, Ecuador
[39]

Yao L. et al.
2021, China
[41]

Iwasa Y et al.
2021, Japan.
[7]

Zhang j. et al.
2020, China
[45]

AI type

CNN

CNN

CNN

CNN

CNN

CNN

CNN


Topic


pancreatic mass diagnosis with CH-EUS and AI guided EUS-FNA in real time
(CH-EUS MASTER)

real-time capture and segmentation of solid pancreatic masses with CH-EUS
(CH-EUS MASTER)

voice-assisted image labeling for AI image classification

real time recognition and characterization of anatomical structures (radial and linear EUS)

bile duct scanning augmentation

CH-EUS for pancreatic tumors, automatic segmentation (U-Net system)

pancreatobiliary segmentation and station recognition system (BP MASTER)


Study population

39 patients (randomized to EUS FNA with or without CH-EUS MASTERuse)

4530 images of pancreatic masses; 8 trainees

12 patients, 3,575 images (143 sets of 25 frames each)

8113 still images from EUS videos

13210 images (10681 for station classification model, 2529 for segmentation model), 781 videos clips (264, 517); 12 treinees

100 patients

4155 images (19,486 for station classification model, 2207 for segmentation model); 8 trainees




Main results

AI for diagnosing pancreatic masses Ac, Sn, Sp, PPV, NPV: 93.8%, 90.9%, 100%, 100%, 83.3%, AI AUC vs endoscopists 0.923 vs 0.865 (p < 0,05); AI guided EUS-FNA Ac, Sn, Sp, PPV, NPV: 93.8%, 90.9%, 100%, 100%, 83.3% ; AI AUC 0.955 vs control 0.933 (p>0,05)

AI real-time capture and segmentation: dice coefficient of 0.763, recall rate 0.941, precision rate 0.642; Ac 0.842. Median IoU of all cases 0.731. Average IoU of trainees: from 0.80 to 0.87 (p = 0.002), average time for identifying lesions (pancreatic body/ tail 22.75 vs 17.98 s p < 0.01; pancreatic head/uncinate 34.21 vs 25.92 s p < 0.01).

AI prediction Ac 76% at image level on a dataset with 5 different landmarks

Radial model: mAP 69.67%, F1-score 92%, average IoU (overall between model prediction and expert marking) 79.08 %, loss of 0.13.
Linear model: mAP 83.43%, F1-score 89%, average IoU 73.48%, total loss 0.16.

AI external validation: Ac in station classification 83.9%, in video set 90.1%.
AI segmentation: dice 0.77 in image set; Sn and Sp 89.48% and 82.3% in video set. Trainees aided by AI: Ac in station recognition 76,3% from 60.8% (p< 0.01).

concordance rate using IoU 0.77

AI external validation: Ac in station classification 82.4%; dice in segmentation 0.715.
AI station classification model: per-frame Ac 86.2% in videos. Trainees aided by the AI: Ac in station recognition 78,4% from 67.2% (p < .01)
Abbreviations: EUS endoscopic ultrasound AI artificial intelligence ANN artificial neural network CNN convolutional neural network GMM Gaussian mixture model SVM support-vector machines SEL subepithelial lesion PDAC pancreatic adenocarcinoma AIP autoimmune pancreatitis CP chronic pancreatitis NP normal pancreas FLL focal liver lesions NET neuroendocrine tumor GIST gastrointestinal stromal tumor IPMN intraductal papillary mucinous neoplasia Ac accuracy Sp specificity Sn sensitivity PPV positive predictive value NPV negative predictive value TIC time-intensity curve AUROC: area under the receiver operating characteristic AUC area under the curve IoU intersection over union mAP mean average precision F1-score FNA fine needle aspiration FNB fine needle biopsy ROSE rapid on-site evaluation MOSE macroscopic on-site evaluation.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions, or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated