Preprint
Review

Computer-Aided Bleeding Detection Algorithms for Capsule Endoscopy: A Systematic Review

Altmetrics

Downloads

152

Views

39

Comments

0

A peer-reviewed article of this preprint also exists.

Submitted:

30 May 2023

Posted:

01 June 2023

You are already at the latest version

Alerts
Abstract
Capsule endoscopy (CE) has been a widely used medical imaging tool for the diagnosis of gastrointestinal tract abnormalities like bleeding. But the CE captures a huge number of image frames that are time-consuming and tedious tasks for medical experts to diagnose manually. To address this issue, researchers focused on the computer-aided bleeding detection system to identify bleeding automatically in real-time. This paper presents a systematic review of the available state-of-the-art computer-aided bleeding detection algorithms for capsule endoscopy. The review was carried out by searching five different repositories: Scopus, PubMed, IEEE Xplore, ACM Digital Library, and ScienceDirect for all original publications on computer-aided bleeding detection published between 2001 and 2021. The PRISMA methodology was used to perform the review, and 112 full-texts of scientific papers were reviewed. The contributions of this paper are I) a taxonomy for computer-aided bleeding detection algorithms for capsule endoscopy is identified, II) the available state-of-the-art computer-aided bleeding detection algorithms including various color spaces (RGB, HSV, etc.), feature extraction techniques, and classifiers are discussed, and III) identify the most effective algorithm for practical use cases. Finally, the paper is concluded by providing future direction for computer-aided bleeding detection research.
Keywords: 
Subject: Computer Science and Mathematics  -   Artificial Intelligence and Machine Learning

1. Introduction

The small bowel (SB) of human is formed with a complex looped shape, configuration, and an extremely large length (around 6 m). For the SB diseases diagnosis, endoscopy can be used to decide the causes of the tumor, cancer, bleeding, and Crohn’s disease [1]. In 2001, Capsule Endoscopy (CE, also called Wireless Capsule Endoscopy, WCE) was approved by the Food and Drug Administration in the United States. It is a noninvasive technology that was designed to provide diagnostic imaging of the SB primarily, as through instrumental examinations this part of the human body is difficult to inspect. It represents the latest endoscopic technique that has revolutionized the treatment and diagnosis of diseases of the upper gastrointestinal (GI) tract, SB, and colon. CE device consists of a CMOS camera sensor with a microchip, LED, a RF transmitter, and battery. The clinical examination which involves the use of CE can be executed in an ambulatory or hospital setting on an outpatient basis. After fasting overnight (8−12 hours), a small capsule is swallowed by the patient. The capsule provides a wireless circuit and micro-imaging video technology for the acquisition and transmission of images. Software that gives localization of the device during its passage through the intestine is encompassed in the system. The capsule is propelled by peristaltic movements when it goes through the SB. While moving along the GI tract, images are captured at a fixed frame rate (2 frames per second - fps) even though the newest model of CE manufactured by Given Imaging (PillCam SB 3 capsule) is able to achieve a frame rate of 2–6 fps based on capsule speed as it travels through the SB [2]. These images are transferred to a data recorder, worn on a belt outside the patient’s body, and about eight hours after swallowing, the patient returns to the clinic where data and images are downloaded. Within 24−48 hours, the capsule is passed into the patient’s stools.
CE has been considered a first-line examination tool for diagnosing various kinds of diseases, including ulcers, polyps, bleeding, and Crohn’s disease [3]. A single scan may include up to ten thousand of images of the GI tract for each patient, but in only a few of them, evidence of abnormalities may appear. A very common abnormality found in the GI tract is bleeding [4]. To detect this many researchers have contributed with high-performance classifiers. Detection of bleeding at an early age is critical since it is a precursor for inflammatory bowel diseases such as Crohn’s disease and Ulcerative colitis (UC). Bleeding is not only limited to the stomach, even anywhere in the GI tract can occur [5]. It can be considered a common abnormality detected by CEs. Which is often defined as "bleeding of unknown origin that recurs or persists or is visible after an upper endoscopy and/or negative endoscopy result" [6]. The major challenge is that residual traces and blood spots don't have any typical shapes or textures, and their colors can range from light red to dark intense red and brown, making it difficult to distinguish the blood from other digestive contents. This diversity of color might depend on the position of the camera, the bleeding timing [7], and the neighboring condition of the intestinal content [8]. Bleeding is not a single pathology, and it may be caused by a variety of small intestinal diseases, such as open wounds, vascular lesions, angiodysplasia, ulcers, Crohn's disease, and tumors. To discriminate pathology, both texture and color features have been used.
Since the CE diagnostic process captured over 57,000 images, manual reviewing is a labor-intensive task for physicians and time-consuming in order to detect bleeding regions[9], which may involve several challenges including complex background and low contrast, variations in the lesion, and color. This may affect the accuracy of subsequent classification and segmentation [10,11]. These issues complicate objective disease diagnosis and necessitate the opinions of many specialists to avoid misdiagnosis
As a result, there is a strong need for an alternate technique to detect bleeding automatically in the GI tract. Some research has been put into the automated inspection and analysis of CE images. Software suites that use computational techniques are often made available with the brand of a particular capsule and are used by a lot of people. The benefit is represented by the efficiency and availability of a tool that can detect bleeding regions automatically and improve diagnostic accuracy. Commercial software built by Given Imaging aims to recognize active blood spontaneously, although the reported sensitivity and specificity are not satisfying [8]. Although having many advantages, the research area in CE technology is not widespread. For instance, at present, it is challenging for physicians to go through the entire collection of more than 50,000 frames in order to find a disease. Due to visual fatigue and the relatively small size of the lesion region, the disease may go undetected in its early stages. The fact that software packages already available on the market are based on low-level, hand-crafted feature extraction algorithms that have poor generalizability should not be ignored. Additionally, because feature extraction and classification phases are separated in hand-crafted feature-based techniques, it is difficult to make reliable diagnostic decisions.
There has been published several informative, original articles and reviews on bleeding detection from CE images over the last 15 years. The authors of [12] reviewed the clinical applications and developments of small bowel CE, i.e., small bowel tumors, Celiac disease, and Crohn’s disease. They gave an insight into the potential future prospects of small bowel CE. In [13] authors discussed different imaging methods: signal-processing, color- and image-processing, artificial intelligence for representing, analyzing, and evaluating CE images. The study [14] calculated the performance metrics: accuracy, positive and negative predictive values, sensitivity, specificity, and compared the diagnostic accuracy of video CE and double-balloon enteroscopy in cases of obscure GI bleeding of vascular origin. Another study [15] discussed the market's available CE models, diagnostic yield, safety profile image quality and technical evolution of small-bowel CE. In [16], authors reviewed and analyzed the computational methods from the literature that can be applied in software, which can improve the diagnostic yield of video CE. Another research group [17] reviewed deep learning based approaches for CE which is used to solve a variety of issues, e.g., detection of polyp/ulcer/cancer, bleeding/hemorrhage/angiectasia and hookworms. In [18], authors reviewed the state-of-the-art approaches of machine-vision-based analysis of CE video. It mainly concentrated on the study of shot boundaries and GI pathology detection. In the literature [19] have assessed the accuracy of video CE to identify active hemorrhage in upper GI. In another article [20] discussed the deep learning methods for WCE, in which only PubMed repository was used for article selection. Moreover, none of these review articles particularly focused on only bleeding detection algorithms for CE. The foremost contributions of this paper are summarized as follows:
  • A taxonomy for computer-aided bleeding detection algorithms for capsule endoscopy was identified.
  • Various color spaces and feature extraction techniques were used to boost the bleeding detection performance, which was discussed in depth.
  • From the observation of existing literature, a direction to the computer-aided bleeding detection research community was provided.
This work only emphasizes on only state-of-the-art bleeding detection algorithms using CE, that is why this paper differentiates from various recent review papers. This review is done by gathering the required information from the recent research works and organizing it according to taxonomy, and analyzing the performance of bleeding detection methods, and providing a path for future research. Moreover, in order to improve the current acceptance of computer aided bleeding detection algorithms with the help of CE, it is hoped that this effort will capture advanced techniques which will be more acceptable in real life applications.

2. Review Methodology

Preferred Reporting Items for Systematic Review and Meta-Analyses (PRISMA) [12] was used as a guide for this review. The authors separately screened the titles and abstracts of the publications retrieved through the database search and then carried out a full-text review of all relevant studies. This methodology used the following processes.

2.1. Identifying Research Questions

To conduct this systematic review, one research question was selected:
Available state-of-the-art computer-aided bleeding detection algorithms for capsule endoscopy: How accurate and suitable they are for practical use?
The answer to the question will help in improving computer-aided bleeding detection algorithms for CE and identifying the research gap in the current methodology.

2.2. Database

Comprehensive searches for suitable literature were performed across five repositories: Scopus, PubMed, IEEE Xplore, ACM Digital Library, and ScienceDirect.

2.3. Search Strategy

To cover all the remaining approaches for bleeding detection in CE, the considered keywords were: ‘Bleeding’, ‘Hemorrhage’, ’Blood’, ‘Detection’, ‘Segmentation’, ‘Recognition’, ‘Classification’, ‘Capsule Endoscopy’. The search query string was: (“Bleeding” OR “Hemorrhage” OR “Blood”) AND (“Detection” OR “Segmentation” OR “Recognition” OR “Classification”) AND (“Capsule Endoscopy”).
The search results were confined to the English language. The primary selected full-text article’s references were analyzed for related publications. Articles related to non-humans, poster, and book chapters were excluded. Articles that fulfilled the exclusion standard were filtered out which is shown in Table 1.

2.4. Results

The initial search results in a total of 1901 publications. A total of 513 duplicates, 34 non-English articles, one abstract, and one incomplete article were removed before the screening. Based on abstracts and titles 1352 articles were set aside for screening. Articles excluded through title skimming were 866 and abstract skimming were 323. And 4 articles were not retrieved. As a result, 159 articles were included as eligible articles from which 38 review articles, 1 dataset article, 1 non-bleeding, 3 non-human articles, 1 poster, and 3 book chapters were excluded. Through the process, 121 articles were included in the systematic review after including 9 published review articles. Figure 1. shows the methodology and results of the systematic review.

3. Review Findings

3.1. Taxonomy

From the current literature, the CE databases are two types according to domain analysis: image and video. The literature on image and video databases is included in this review. To answer the research question, a taxonomy is generated from the literature findings, which is shown in Figure 2. The image and video domains are further divided into three categories: classification, segmentation, and combined (classification + segmentation). Classification is a task that can perform to classify bleeding images from non-bleeding images using classification algorithms. To detect a bleeding zone, segmentation is another important technique that can cluster a bleeding image into several coherent subregions. A few articles proposed both classification and segmentation to classify bleeding images and also detects bleeding region at the same time for the development of an advanced computer-aided diagnosis system. Different types of feature extraction techniques for example color space, texture attributes, region of interest (ROI), pixel or block-wise contourlet transform, etc. are used to tune the machine learning algorithms for identification of bleeding images accurately. Also, various convolutional neural networks such as VGG, ResNet, DenseNet, etc. are used to extract features from the CE images. Most of the literature is used color space for feature extraction like RGB, HSV, HSI, YCbCr, CMYK, CIE Lab color space, etc. From the review finding, the state-of-the-art computer-aided bleeding detection algorithms are categorized into two types: conventional machine learning algorithms and deep learning algorithms. The available bleeding detection algorithms in the literature will cover the research-identified question.

3.2. Analysis Domain

Two types of CE databases of bleeding are available. One is the image domain, and another is the video domain database. Most of the literature reviews proposed bleeding detection systems based on the image domain.

3.2.1. Image

An image is a matrix of pixels organized in columns and rows. The greyscale image represents a one-dimensional matrix and the RGB image represents a three-dimensional matrix. Due to the characteristics of the image, a digital machine or processor can analyze the medical image to detect the abnormality. During an eight-hour-long, a CE captured approximately 55,000-57,000 images throughout the GI tract in one experiment. The article [21], proposed a CE image dataset obtained from 10 patients at the University of Malay Medical Center (UMMC). The dataset consists of 100 bleeding and 300 normal images which has a resolution of 288×288 pixels. 1131 bleeding lesions and 2164 normal non-sequential images were utilized in an algorithm to detect bleeding. The size of the images was 320×320 or 512×512 pixels [22]. In [23], authors suggested a KID Dataset 2 that contains a total of 2352 CE images (bleeding: 303) with a resolution of 360×360 pixels and a MICCAI 2017 that contains a total of 3895 CE images (bleeding: 1570) with 320×320 or 512×512 resolutions. A total of 1200 CE images with a resolution of 576×576 pixels were utilized in the study [24]. According to [25], the model was trained using 2000 images that extract from 20 different videos. The images dataset was available on capsuleendoscopy.com. An OMOM capsule produced 3,596 CE images with 256×240 pixels size which were extracted from 5 subjects [26]. The size of 240×240×3 CE image dataset was utilized in the study that contains 148 bleeding and 152 inactive images from 60 CE video snippets of 12 subjects [27]. The size of 240×240×3 CE image dataset was utilized in the study that contains 148 bleeding and 152 inactive images from 60 CE video snippets of 12 subjects [27].

3.2.2. Video

Few studies utilized the video domain of CE for automatic bleeding detection. The frames of the video are sequentially arranged and applied to an algorithm for identifying bleeding frames or regions. Time is an important parameter of the video domain. The video domain has the best likelihood of detecting bleeding by using the previous and next frames of video as well as the time information. However, it is difficult for long-length videos that require high processing power. The study of [28] utilized a video CE dataset with a sequence of 600 frames which was collected from a PillCam SB3 video. Among them, 73% of the total frames were red lesions. In [29,30], if the last frame of the CE video was not found, the authors presented a classification method in which the model continued the process to the next frame of the video. Videos of 15 patients ranging in duration from 12,000 to 20,000 frames were used for automatic bleeding detection methods [31]. A semi-automatic method was proposed to extract bleeding regions from successive video frames containing bleeding. Three video files were used that include 589, 500, and 428 frames [32]. Deeba et al. improved the model by skipping one or a few frames from the sequence of bleeding frames [33]. Using 84 full-length videos, the authors proposed an algorithm that compares with the Suspected Blood Indicator (SBI) [34].
Five publicly accessible CE bleeding videos were tested by [35,36] for the automatic bleeding detection model to identify bleeding. Ten real patient video files were used in [37], which consisted of 200 frames and about 40 seconds in duration. Instead of dealing with a complete video at a time, consecutive small portions of videos were used. Each video consisted of several frames which were tested sequentially [38]. Another research used time domain information of CE videos for a bleeding localization technique. The approach used 8 different videos collected from 8 subjects at West China hospital [39]. 4166 third-generation small-bowel CE videos were applied in [40], which were collected from the Computer-Assisted Diagnosis for Capsule Endoscopy database (CAD-CAP) endorsed by Société Française d’Endoscopie Digestive, French.

3.2.3. Task

In this review, all kinds of literature on the computer-aided bleeding detection approach are divided into three categories based on the task. The first category is the bleeding images or frames classification task. The second is the segmentation task for bleeding zone identification. And the third one is the combination of classification and segmentation.

Classification

Classification is a supervised learning technique in both machine learning and deep learning that is used to categorize a given set of data into classes. A classification model learns from the given dataset and then classifies new observations into a number of classes or groups. Such as 0 or 1, yes or no, bleeding image or non-bleeding image, etc. The study of [23], presented a machine learning algorithm to classify CE images into two categories named bleeding and non-bleeding images. Rustam et al. developed a deep neural network for the classification of bleeding CE images [41]. A video frame classification model was presented by [30], to classify bleeding frames from CE videos.

Segmentation

Segmentation is a technique in which an image is broken down into different subregions according to the extracted features. An image is a collection or set of different pixels. Similar types of pixels are grouped according to image segmentation. This technique helps in minimizing the complexity of the image to make further processing or analysis of the image simpler. In [42], presented an automated segmentation technique based on a machine learning algorithm to detect bleeding regions in CE images. In another research for bleeding zone segmentation of CE images, a deep learning-based model was proposed by [43]. A time domain-based segmentation approach was provided by [39] for locating bleeding in CE videos.

Classification + Segmentation

This category includes the classification and segmentation technique a time to classify images and locates the region of similar types of pixels. In this review, several articles proposed to classify bleeding images and also detects bleeding region at a time using both classification and segmentation technique for the development of advanced computer-aided diagnosis system. Rathnamala et al. presented a model for automatic bleeding detection using CE images. Firstly, the model classifies bleeding and non-bleeding image and then applied the post-segmentation technique to detect the bleeding region of the bleeding image [44]. Two deep-learning models were presented in [45], to classify bleeding images and zone in CE images. In [35], a computer-aided scheme was presented for the classification of bleeding frames from CE videos and then the post-segmentation technique was applied for the localization of the bleeding zone.

3. Feature Extraction

Features are a major part of any pattern recognition task. Feature extraction is the process of converting raw data (like an image) into a set of features. It aids in reducing the number of resources required to explain big amounts of data. Color and texture are two common and crucial image recognition properties. Both are also highly beneficial to extract features in CE images to identify bleeding because bleeding areas are given more or fewer color differences and/or texture compared with their neighborhood. Feature extraction domain is a selection process that selects a region of the image that is used to efficiently extract features. Various feature extraction domains are used to accurately extract bleeding features such as region of interest (ROI), specific block, pixel level, or image level.

Color Space

Color space is a specific color arrangement. There are many different color spaces such as RGB (red-R, green-G, blue-B), HSV (hue-H, saturation-S, value-V), YIQ (luminance-Y, chrominance-IQ: in phase-I and quadrature-Q), YCbCr, CMYK (cyan-C, magenta-M, yellow-Y, and key (black)), CIE-Lab, CIE XYZ, etc. Form the literature review, the color spaces used for feature extraction can be categorized into four groups: RGB, HSV, Other and Combined color space.
RGB: Images are represented in the RGB color space as an m-by-n-by-3 numeric array whose components indicate the intensity levels of red, green, and blue color channels. The range of numeric values is determined by the image's data type. Different types of RGB color space are available such as linear RGB, sRGB (standard red, green, blue), adobe RGB, and so on. In CE images or videos, a bleeding zone is distinguished by the presence of a brightly red or dark red zone. Many studies utilized RGB color to extract features for the identification of bleeding images or regions from CE images. The studies [46,47] presented an automated obscure bleeding detection technique on the GI tract based on statistical RGB color features that can classify bleeding and non-bleeding images from CE images. By using the same RGB components of each pixel of CE images, the study [48] presented a system to automatically detect bleeding zones in the CE images. From the first-order histogram in RGB planes, the approach extracted bleeding color information from CE image zones by calculating mean, standard deviation, skew, and energy [49]. Zhao et al. presented two-dimensional color coordinate systems in RGB color space to segment abnormality in the CE video. The approach combined two descriptors to extract features: the first was based on image color content, while the second was based on image edge information [37]. Another research group proposed using color vector similarity coefficients to evaluate the color similarity in RGB color space in order to detect bleeding in CE images [50]. Yun et al. presented a method using color spectrum transformation (CST) for the identification of bleeding in CE images. This approach included a parameter compensation step that used a color balance index (CBI) in RGB color space to compensate for irregular image conditions [51]. The study [52] suggested an automatic bleeding image detection technique utilizing RGB color histogram as a feature extractor and bit-plane slicing to detect bleeding and non-bleeding images from CE videos. In [53] utilized superpixel segmentation in RGB color format to extract bleeding information for an automatic obscure bleeding detection technique. According to [21], an automated bleeding detection approach was presented by using color-based per-pixel feature extraction techniques. Ghosh et al. [36] presented an automatic bleeding detection approach based on an RGB color histogram of block statistics to extract features from CE videos. For reducing computational complexity and flexibility, the approach utilized blocks of surrounding pixels rather than individual pixel values. The traditional machine learning model used single pixels for training and testing data. So, the model was unable to eliminate a few very small-judged bleeding zones which are not bleeding. To address this issue, a cluster of pixels-based feature extraction techniques was used in some research to extract features from bleeding CE images. A cluster of pixels in RGB color space was utilized instead of single pixels in an automatic bleeding classification system which improved the sensitivity [54].
Instead of directly using RGB color space, a G/R composite color plane was utilized to extract features from the CE images [35]. Another research [38] extracted statistical features from the overlapping spatial blocks of the CE images based on the G/R color plane. A transform color plane R/G pixel intensity ratio was utilized for the extraction of bleeding information from the CE images [55]. Rather than considering individual pixels, Ghosh et al. considered the surrounding neighborhood block of that individual pixel and the R/G ratio plane for bleeding feature extraction from CE images [56]. Shi et al. [39] used a temporal red-to-green ratio (R/G) feature value to detect bleeding regions.
According to [57], the authors presented an average pixel intensity ratio in RGB color space to extract features in CE images for an automatic bleeding detection approach. The study [58] presented rapid bleeding detection from CE video. The red ratio (RR) in RGB space was used to extract a feature from each superpixel of CE images. Also, the RR feature for individual pixels was utilized for feature extraction of bleeding in CE images [25]. The various coefficient of RGB color space for bleeding and non-bleeding superpixel blocks are RG, RB, and GB two-dimensional space. Liu et al. [59] presented an automatic detection gastric hemorrhage system based on the coefficient of variation in RG two-dimensional color space coefficient of a different super-pixel block of CE images. Another transformation form of RGB color space is OHTA color space. In [60], OHTA color space was utilized to extract the feature of bleeding from CE images.
A custom RGB color space was proposed by [31], which was similar to the CMYK color space and was used to extract features for automatic blood detection in CE video. Kundu et al. [61], presented a normalized RGB color space histogram-based feature extraction method to identify bleeding in CE images. Another research group employed a two-stage saliency map extraction method to localize the bleeding areas in the CE images. The first-stage saliency map was constructed using a color channel mixer, and the second-stage saliency map was derived from the RGB color space's visual contrast [62].
Some studies applied algorithms in RGB color space to extract features from CE images or videos. Using the advanced pattern recognition techniques, a MapReduce framework was presented for the identification of bleeding frames and segmentation of the bleeding zones. For classification, the system encodes RGB color information from raw data of CE images using a K-means clustering algorithm. And for the segmentation of the bleeding zone, a density-based algorithm (DBSCAN) was utilized [63]. Hwang et al. presented an automatic bleeding region detection system by using the Expectation Maximization (EM) clustering algorithm in RGB color space [64].
HSV: HSV (hue-H, saturation-S, value-V) is an alternative representation of the RGB color space that correlates better to the human perception system. The HSV color space generated from cartesian RGB primaries, and its components and colorimetry are related to the color space from which it is derived. Several studies employed the HSV color space histogram as the color feature descriptor to extract features from CE images or videos in an automated bleeding detection system [65,66]. In [67], the RGB image was transformed into HSV color and calculated several statistical parameters like variance, kurtosis, skewness, entropy, etc. from the histogram of CE images. The extracted features were applied for a bleeding detection method. The contrast, cluster shade, and cluster prominence, entropy were computed to extract bleeding features from Gray Level Co-occurrence Matrix (GLCM) in HSV color space [68]. Usman et al. suggested a pixel-based method for the detection of bleeding regions in CE videos. The HSV color space was utilized to compute the bleeding information [29]. The study [69], presented a color-based segmentation using HSV color space to detect the bleeding regions in CE images. Giritharan et al. presented a bleeding detection method based on the HSV color space with dominant color, and co-occurrence of dominant colors for feature extraction to classify bleeding lesions [70]. In [71], suggested HSV color moments to extract bleeding features of CE images achieved the highest accuracy compared to the local binary pattern (LBP), local color moments, and gabor filter. Using a block-based color saturation approach in HSV color space, the CE images were classified as bleeding or non-bleeding [72]. A two-stage analysis system was proposed in [73], where both blocks and pixels-based color saturation methods used HSV color space to extract bleeding features. Color saturation and hue were obtained for the study by converting the input videos or images to the HSV color space. In [74], authors compared texture features with color features extracted from HSV color space for the classification of bleeding and showed color features provided better results. A fuzzy logic edge detection technique was applied in HSV color space by [1], to extract features of bleeding and non-bleeding from the CE images. According to [9], a feature selection strategy was proposed based on HSV color transformation to extract geometric features from CE images for classifying bleeding images. To extract the color features, another research group [75] applied the HSV color scale-invariant feature transform (HSV-SIFT) for CE abnormality detection.
Various statistical features are computed from hue, saturation, and value channels of HSV color spaces. Such as the hue space (H) provided a useful feature for color objects or surfaces. The hue space was utilized by [76], to extract features of CE images for an automatic bleeding detection approach. Another strategy used the combination of hue-saturation (HS) color histogram with relevant features (64 bins) for extraction of information to identify suspected blood abnormality [77]. An HSI (hue-H, saturation-S, intensity-I) color space is another variation of HSV color space. A binary feature vector in HSI color space was more effective to extract features in a bleeding detection approach [78]. A segmentation approach in which the average saturation from the HSI color space, as well as the skewness and kurtosis of the uniform LBP histogram, were used as features for automated segmentation to detect bleeding in CE images [42]. Another research group suggested an HSI color histogram to follow a moving background and bleeding color distributions through time in the first stage. Cui et al. [79] presented six color features in HSI color space to classify bleeding and normal CE images.
Other color spaces: Besides the most popular RGB and HSV color space, a few articles utilized different color spaces to extract bleeding features from CE images such as YIQ (luminance-Y, chrominance-IQ: in phase-I and quadrature-Q), YCbCr, CIE Lab, CIE XYZ and K-L (Karhunen-Loeve transform) color space. The study [80] analyzed only the Q value of the YIQ color scheme to determine the ROI section. Then a composite space Y.I/Q of YIQ color space was presented to extract bleeding by computing the mean, median, skewness, and minima of the pixel values. Based on the YIQ color histogram, another article proposed an automatic bleeding detection scheme from CE images [81]. A YCbCr color space was presented to collect information from CE images in order to identify images with lesions [82]. Yuan et al. investigated several color histograms, including RGB, HSV, YCbCr, and LAB, and proposed YCbCr color space to extract bleeding features for discrimination of bleeding images from normal CE images [83]. The study of [84], suggested a second component I of CIE Lab color space for localization of bleeding region in CE images. Mathew et al. proposed a bleeding zone detection system based on contourlet transform in CIE XYZ color space [85]. Another feature extraction color space was Karhunen-Loeve (K-L) transformation which was utilized for fuzzy region segmentation of CE images [86]. The study [87], proposed a computer-aided bleeding and ulcer detection approach based on the covariance of second-order statistical features in K-L color space. A K-means color group was suggested as a color feature extractor for super-pixel segmentation to find bleeding regions from CE videos [88].
Combined multiple color spaces: To detect bleeding, a group of color features was computed using multiple color spaces of CE images. The method of [26] employed two distinct enhancing operations: the first is for RGB, and the second is for grayscale color space for identifying bleeding images in CE. The study of [89] determined the ROI of bleeding CE images using YIQ color space and extract features from the ROI using CMYK color space. Based on RGB and HSV color spaces, the CE images were defined using statistical characteristics to extract bleeding features [33,90]. A 9-D feature was extracted at the superpixel level from the RGB and HSV color spaces during the segmentation stage of the study [91]. Some studies utilized a combination of RGB and HSI color spaces to extract features for the bleeding detection approach in CE images [92,93,94,95]. Five color spaces (RGB, HSV, Lab, YCbCr, and CMYK) were used to extract the features in [96]. In the study [7], authors proposed the R channel with respect to the G and B channels and the ratio of G and B channels as features in RGB color space, and HSV color space was chosen for the saturation feature. The color features were extracted by using the color components X = {R, G, B, L, a, b, H, S, V, F1, F2, F3} in RGB, CIE-Lab, and HSV color space from each super-pixel of CE images [97]. In [98] the color components H, S, a, b from HSV, Lab color space, and Ros (Rosenfeld-Troy) metric were used. Ten features including normalized excessive red (NER), Hue, sum RGB, chroma, etc., were used for analyzed CE video frames in [99]. For the segmentation of bleeding region from bleeding CE images, deltaE color differences were used to extract features which applied 9 color shades (red, orange, brown, maroon, purple, pink, mahogany, brown, bittersweet) for characterizing different types of bleeding [44]. The recommended probability density function (PDF) fitting-based feature extraction technique which used in YIQ, HSV, and CIE-LAB color spaces [100]. In [101], 40 features were extracted from five different channels including R in RGB, V in HSV, Cr in YCbCr, and a, L in Lab color spaces. The article [102] investigated 21 color components for feature extraction of CE images such as U/Y, V/Y, I/Y, and Q/Y in RGB, YUV, YIQ, HSB, CIE XYZ, CIE L*a*b* color spaces. In [103], a combination of HSV and YIQ color spaces was demonstrated using normal PDF to detect GI diseases from CE videos. According to [30], HSV color space was used for threshold analysis of the classification model, and CIE-Lab color space was used in the trainable model for edge detection of images.

Texture

Texture feature is used to partition images into ROI and to classify those regions as bleeding and non-bleeding. It provides information about an image in the spatial pattern of colors or intensities that repeats. In [104,105] a traditional texture representation model named uniform local binary pattern (LBP) was used to differentiate bleeding and normal regions. The study of [34], extracted the texture features (LBP) from the suspicious areas of images and their surroundings for classifying bleeding. Zhao et al. [37] extracted a LBP based on contourlet transforms as texture features to segment the abnormality in WCE. As a color texture feature, Li et al. [12] integrate chrominance moments and uniform LBP to discriminate bleeding region from normal region. Charfi et al.[69] also extract texture features (LBP) for segmentation of WCE images in order to prevent false detections. For recognizing bleeding regions, a 6D color texture feature vector {x = (R, G, B, H, S, I)} was developed in the article [94]. Pogorelov et al. [106] presented a bleeding detection system computed texture features to extract additional information from the captured image frames. By using a histogram on the index image, a distinguishable color texture feature is developed in [52] for automatic bleeding image detection.
The gray-level co-occurrence matrix (GLCM) is a statistical approach to assessing texture that considers the spatial interaction of pixels. The GLCM functions describe the texture of CE images by computing how frequently pairs of pixels with given values and in a specified spatial relationship appear in an image, generating a GLCM, and afterward extracting statistical measures from this matrix. In the study [107], the authors proposed an efficient normalized GLCM for extracting the bleeding features of the CE images. In [108] proposed a texture feature descriptor based algorithm that operates on the normalized GLCM of the magnitude spectrum of the images for a real-time computerized GI hemorrhage detection system. The study of [77] compared two types of texture feature which are GLCM and homogeneous texture descriptor (HTD) with various numbers of color histogram bins. Rathnamala et al. [44] extracted texture attributes from the Gaussian mixture model superpixels of the WCE images.

Extraction Domain

According to the taxonomy, the extraction domain is the process of extracting bleeding features from CE images. All the review studies are categorized into three parts depending on the extraction domain. Global feature: when the features are extracted from the whole frames or images. Local feature: when the features are extracted from pixel-level or a portion of an image (specific block size, ROI, POI). Combined local and global features: when the features are extracted from both pixel level and image level.
Global feature: The entire image information was used in the global feature extraction technique. Using statistical features (such as mean, mode, variance, moment, entropy, energy, skewness, kurtosis, etc.), several articles [1,67,76] extracted bleeding features from whole CE images. In [49], statistical color features of bleeding images were extracted from the RGB plane's first-order histogram. In another study [68], statistical features were measured from Gray Level Co-occurrence Matrix after applying an Undecimated Double Density Dual Tree Discrete Wavelet Transform on CE images. Cui et al. [79] applied six color statistical features to identify bleeding features from the full image feature. Zhou et al. [78] utilized color information to extract the bleeding features from the image feature. In [9], color, shape, and surf were used for feature extraction from whole images.
Local feature: A pixel-level feature extraction approach was proposed to several studies in order to accurately identify bleeding images [29,35,50,55,57]. Instead of computing different features from each pixel, a few researchers proposed block-based local feature extraction techniques for reducing time and computational cost [36,46]. The study [38] investigated various overlapping block sizes: 3×3, 5×5, 7×7 and 9×9 and proposed a 7×7 block size to extract features of CE images. Maghsoudi et al. [109] divided the original image of 512×512 pixels into 256 sub-images with a resolution of 32×32 pixels for feature extraction. In [74], 576×576 pixels input was sliced into nine non-overlapping blocks with 64×64 pixels. Another research group divided each CE image into blocks of 64×64 pixels and analyzed the 64×64 = 4096 pixels in each block to recognition of hemorrhage [73]. The CE images are surrounded by a large black background which provided unwanted features. As a result, it reduces the performance of the model. To address this, a few studies introduced regions of interest (ROI) for proper feature extraction. In [83], and [88] authors selected an ROI from a maximum square inside the circular CE images without loss of main information. The ROI was 180×180 pixels in size, chosen from a total of 256×256 pixels. An elliptical ROI was selected inside the image to extract local features [33]. According to [89], an ROI of the bleeding CE image is determined using YIQ color space. After that, CMYK (cyan-C, magenta-M, yellow-Y, and black-K) values are computed within the ROI pixels which was applied to discriminate the bleeding and non-bleeding pixels. An ROI was selected based on the Q value of the YIQ color space and a composite space Y.I/Q was used to capture bleeding information from the ROI section of CE images [80]. The pixels of interest (POI) are another technique for extracting local features that depend on the intensity values of pixels. In [100,103] utilized POI instead of whole CE image to extract features for the classification of bleeding images.
Combined local and global features: Studies used local and global features to develop a robust and accurate computer-aided bleeding detection system. The studies of [64,94,95] proposed a bleeding detection application software that was tested on pixel level and image level. A region-level block-based and an image-level global feature extraction technique were applied in [38] to identify bleeding images. There were two stages presented in [106], the first used only local color features to categorize bleeding images, while the second included global texture and color features to classify bleeding pixels. In [36], a block-based local feature extraction technique was presented, and then global features were extracted using a color histogram to classify bleeding and non-bleeding images. Ghosh et al. [56] presented an article that used the maximum pixel value of each proposed spatial block and global feature of an image to extract the bleeding features. Another study [58] used pixels to remove the edge zone and group pixels adaptively based on the red ratio in RGB color space for superpixel segmentation. Another study proposed a global feature descriptor based on magnitude spectrum entropy and a local textural descriptor based on contrast, sum entropy, sum variance, difference variance, and difference average that operating on the normalized GLCM [108]. Few researchers are proposed various machine learning and deep learning algorithms to extract both global and local features of CE images. The study of [22] proposed a bleeding detection method by using a genetic algorithm for the feature selection of CE images. Using an unsupervised K-means clustering algorithm, some studies extracted features for automatic bleeding detection in CE [38], [63]. Three pre-trained deep convolutional neural networks (CNNs) named ResNet50, VGG19, and InceptionV3 models were used to extract features of CE images suggested by [23].

Algorithm

Initially, researchers proposed various threshold values to detect bleeding [50,51]. Various machine learning (ML) and deep learning (DL) algorithms are currently being used for accurate bleeding detection.
Machine Learning (ML): Several ML algorithms are applied for computer-aided bleeding detection systems to detect bleeding effectively from CE images or videos. Such as Support Vector Machine (SVM), K-Nearest Neighbors (KNN), K-means clustering, Naïve Bayes, Random Tree, Random Forest, Artificial Neural Networks (ANN), Probabilistic Neural Networks (PNN), Multilayer Perceptron (MLP), etc.
SVM is one of the most popular supervised ML algorithms that is used to detect bleeding and non-bleeding images or zones from CE images or videos. The majority of studies used SVM based on the information of extracted features of input images including color space and texture. In 2008, Liu et al. [47] developed an automated obscure bleeding detection technique on the GI tract that could classify bleeding and non-bleeding images from CE images by using SVM algorithm. An automated bleeding detection approach was presented by [46] which provided an accuracy of 97.67%. Another research group suggested an automatic bleeding image detection technique utilizing an SVM classifier to detect bleeding and non-bleeding frames from CE videos. The approach reported 94.50% accuracy, 93.00% sensitivity and 94.88% specificity [52]. The study [35], utilized the SVM classifier to train with 200 bleeding and 200 non-bleeding CE images and achieved 97.96%, 97.75%, and 97.99% of accuracy, sensitivity, and specificity respectively. Studies in [48,110] suggested a system to automatically detect bleeding in the CE images by using the SVM classifier. A recent study suggested a quadratic support vector machine (QSVM) classifier for an automated bleeding detection approach which was proposed in [1]. A fuzzy logic technique was applied to extract the feature of the images. The model achieved 98.2%, 98%, and 98% of accuracy, sensitivity, and specificity respectively. Joshi et. al. [111] presented an SVM classification model based on an improved Bag of Visual Words to detect bleeding from CE images
Different kernel functions of the SVM algorithm including linear, polynomial (homogeneous), and radial basis function (RBF) were used in bleeding detection in CE image research. The SVM classifier with linear kernel was utilized in a real-time computerized gastrointestinal hemorrhage detection method in CE videos. The results obtained 99.19%, 99.41%, and 98.95% of accuracy, sensitivity, and specificity respectively [108]. Liu et al. [59] presented an automatic detection gastric hemorrhage system SVM classifier with RBF as the kernel function which achieved 95.8% of accuracy, 87.5% of sensitivity, 98.1% of specificity, and 12.5% of miss detection rate, 1.9% of false detection rate respectively. Another research [21] also used the RBF kernel of the SVM classifier to discriminate between bleeding and non-bleeding images that achieved 98.0%, 97.0%, and 98.0% of accuracy, specificity, and sensitivity respectively. In addition [93], the SVM classifier model was used to detect bleeding using chi-square kernel and histogram intersection. The combination of spatial pyramids with a robust hue histogram improved the accuracy by about 8%.
KNN is the second most popular supervised ML algorithm to detect bleeding from CE images or videos. The algorithm processes all existing pixels of CE images and classifies new pixels based on similarities. In [81], the KNN classifier was also used to train CE videos that achieved 97.50% of accuracy with 94.33% of sensitivity and 98.21% specificity. Kundu et al. employed a KNN model for detecting bleeding from CE images which achieved an accuracy of 98.12%, sensitivity of 94.98%, and specificity of 98.55%. The article [91] presented a bleeding detection approach in CE images that was compared to various ML algorithms such as SVM, AdaBoost, and KNN and, the KNN achieved the best results with 99.22% accuracy. In another research, a KNN classifier was employed by [76] to distinguish the characteristics of bleeding and non-bleeding images. The classifier was trained with 200 color CE images and achieved 99.0% of accuracy.
Artificial Neural Network (ANN) or Neural Network (NN) is a ML algorithm of computer systems that are designed inspired by biological neural networks. The approach [49], employed an artificial neural network (ANN) with 3 input neurons, 22 hidden neurons, and 2 output neurons with a minimum squared error loss function. The ANN classifier also was applied in preprocessing step to analyze the pixels in the CE images [25]. ANN also known as neural network (NN). A NN cell classifier was applied in [65] to categorize bleeding and non-bleeding patches from CE images. Another research proposed back propagation NN to detect bleeding regions that achieved 97% of sensitivity and 90% of specificity. A Probabilistic Neural Network (PNN) is a radial basis function and Bayesian Theory-based feedforward neural network. One such study [95], applied PNN to detect bleeding zone in CE images that achieved 93.1% of sensitivity and 85.6% of specificity. Multilayer Perceptron (MLP) is a type of fully connected feedforward ANN that is widely used in statistical pattern recognition. Several articles [32,105,109,112,113], employed MLP neural networks to classify bleeding images and bleeding regions from CE images. The article [102] proposed Vector Supported Convex Hull classification algorithm which was compared to that of SVM configured with two alternative feature selection approaches. The model achieved a 98% of sensitivity and specificity ratio for bleeding detection. Another research [90] suggested a computer-aided color feature-based bleeding detection technique using a modified ant colony optimization algorithm. The model achieved 98.82%, 99.66%, and 98.01% accuracy, sensitivity, and specificity respectively.
Other ML: Naive Bayes classifier is another ML classification algorithm based on Bayes Theorem. The studies [53], and [88] used a Naive Bayes classifier to detect bleeding from CE images. In [37], another ML algorithm K-means clustering was applied to extract important features for summarizing CE video clips. Random Tree and Random Forest is a tree-based ML algorithm for making decisions. In the study [67], a Random Tree classifier was trained with 100 bleeding and 100 non-bleeding images for a computer-aided bleeding detection system. The classifier achieved an accuracy of 99%, sensitivity of 98%, and specificity of 99%. In [68], both Random Tree and Random Forest classifier models outperformed for bleeding detection compared with MLP, and Naive Bayes models. Both models provided 99.5%, 99%, and 100% accuracy, sensitivity, and specificity. The Random Forest model was also used in [66] which achieved 95.7% of sensitivity and 92.3% of specificity.
Combined multiple ML: In the study [98] a block-based segmentation technique using local features was presented and several ML algorithms like linear discriminant analysis; SVM, random decision forests, and ADA boost were applied for discriminating between bleeding and non-bleeding images. Using SVM and K-means algorithm, a GI bleeding detection approach was presented in [63] to detect bleeding images and regions of bleeding images which reported less computation time with 98.04% of accuracy and 84.88% of precision. According to [77], an automatic detection system is designed to identify suspected blood indicators from CE images. The authors compared various ML classifiers like SVM, and NN. ISVM, which trained with 136 normal with 214 abnormal images and achieved a maximum of 98.13% accuracy for the TAF-SVM algorithm. In addition, an automatic bleeding detection approach in the CE video was suggested by Ghosh et al. by using a cluster-based feature. The clustering information of the CE images was applied to the SVM classifier to detect bleeding zone from CE images and obtained a precision of 97.05%, FPR of 1.1%, and FNR of 22.38% [38].
Deep Learning (DL): The most widely used deep learning approach for image classification and segmentation is the convolutional neural network (CNN). The CNN model has been utilized in a number of studies to detect bleeding in CE images [40,114,115,116,117,118]. The study of [27], presented a Fully Convolutional Neural Network (FCN) model for automatic blood region segmentation systems. Another study [119], proposed a Look-Behind FCN algorithm for abnormalities detection (polyps, ulcers, and blood) in CE images that achieved an accuracy of 97.84%. In [120], a LeNet model was trained and adopted pre-trained AlexNet, VGG-Net, and GoogLeNet models to identify intestinal hemorrhage. In [45], the authors applied a pre-trained AlexNet model for identification and a SegNet model for the segmentation of intestinal bleeding. Another CNN algorithm named U-Net architecture was proposed in [28,43] to segment bleeding areas from CE images and videos. Xing et al. [24] proposed a Saliency-aware Hybrid Network algorithm based on two densely connected convolutional networks (DenseNets) for an automatic bleeding detection system. The authors of [121] developed a CNN model for detecting bleeding zones that were trained using SegNet layers with bleeding, non-bleeding, and background classes. A blood content detection using ResNet architecture with 50 layers was suggested in [122] which achieved an accuracy of 99.89%, sensitivity of 96.63%, and specificity of 99.96%. Hwang et al. used a CNN model based on VGGNet to identify lesions with a 96.83% accuracy [123]. Another CNN model for classifying bleeding images from CE was provided in another work [41]. The model was created utilizing MobileNet and a custom-built CNN. To identify small-bowel angioectasia, the authors of [124] used a 16-layer Single Shot MultiBox Detector (SSD) deep CNN method.
Combined ML & DL algorithms: In the study [125], deep CNNs (VGG16 and VGG19) were applied to extract features from CE images. A KNN algorithm was proposed to classify bleeding images that achieved 99.42% and 99.51% of accuracy and precision rate. An automatic bleeding region segmentation technique was presented by [126] using MLP and CNN models individually. In another study [23], authors proposed pre-trained deep CNNs (VGG19, InceptionV3, and ResNet50) models to extract bleeding features, and ML algorithms (SVM, KNN, Linear Regression) were utilized to distinguish the bleeding and non-bleeding images. In [127] applied DenseNet for feature extraction and the features were trained with an MLP algorithm to classify GI track abdominal infections. In addition, the study [128] applied the CNN model to extract bleeding features and the SVM classifier was used to detect bleeding.

4. Discussion

A CE device typically recorded video in the GI tract around 8 hours long. Few studies utilize video to detect bleeding abnormalities. Figure 3. shows the overview of used domain and algorithms of the papers that are reviewed in this study. To detect bleeding from CE images, researchers used three tasks which include classification, segmentation and combined (classification + segmentation). Articles those proposed classification task showed an average accuracy of 95.89% ± 3.11, sensitivity of 94.91% ± 4.46 and specificity of 94.95% ± 4.35. For the segmentation algorithms, achieved average accuracy is obtained around 94.95% ± 2.91, sensitivity of 95.18% ± 3.99, and specificity of 96.30% ± 1.98. Articles that proposed combined task achieved an average accuracy of 97.17% ± 2.08, sensitivity of 96.10% ± 4.89, and specificity of 96.62% ± 4.15. Based on the above literature analysis, the combined task performed better. One significant benefit of the current methods is their ability to identify bleeding in images/frames from CE and pinpoint the specific bleeding region. However, a limitation of these methods is their inability to measure the extent or depth of the bleeding area.
Features extraction is an essential part of bleeding detection from CE images. The feature values are extracted from the color channels of CE images. The performance of the bleeding detection algorithms directly depends on the feature values. To identify bleeding, lots of color spaces were presented. RGB is the popular color space to extract features. Because it is the default color space. Apart from RGB, there are several studies proposed individual color channels (R or G or B, etc.) or color channel pixel ratios (R/G, G/R, etc.), or various color spaces (HSV, YIQ, YCbCr, CIE Lab, CIE XYZ, K-L, etc.) to extract appropriate bleeding features. A few researchers applied two or more color spaces together to extract features. According to the taxonomy, all the suggested color spaces were categorized into four groups: RGB, HSV, Combined (multiple color spaces), and Other (YIQ, YCbCr, CIE Lab, CIE XYZ, etc.). It is important to acknowledge that the performance results presented in this article are directly extracted from the original paper. The box plots in all the figures compare the performance between groups of methods, rather than individual algorithms. The statistical measures such as mean, median, 25th percentile, and 75th percentile is utilized for each group. The performance results to detect bleeding for different color spaces are shown in Figure 4. using a box plot. According to the figure, color space does not provide any performance benefits. All the color spaces provided similar results except the ‘Other’ color space. When comparing all color spaces, the RGB space has a slightly higher recall value. It should be noted that the recall performance criterion is the most important in the detection of bleeding. On top of that, the RGB color space achieved lower variance for accuracy, recall, and specificity. The current methods make a significant contribution by investigating all potential color spaces to detect bleeding in capsule endoscopy
A typical approach for extracting bleeding features from whole CE images is the global feature extraction domain. Because the approach analyzes the entire image at once, the complexity and processing time are increased. It provides an average result which is shown in Figure 5. using a box plot. To address the problem, various studies proposed a pixel-level feature extraction domain, where the technique analyzed each pixel of the CE image. The technique improved the result but did not reduce the complexity and computation time. Several authors selected a portion of CE images (specific block size, ROI, POI) in the preprocessing for feature extraction as name as local feature extraction domain. The technique improved the result as well as reduced the complexity and processing time. Recently, a few researchers applied both global and local feature extraction domains in a computer-aided bleeding detection system, which significantly enhanced the detection accuracy compared to the individual domain. Texture and statistical values (mean, mode, variance, moment, entropy, energy, skewness, kurtosis, etc.) are calculated using the feature values. Finally, a classification or segmentation algorithm was used by using the extracted values to detect bleeding from CE images. According to Figure 5., the combined feature extraction domain outperformed the other domains in terms of accuracy, sensitivity, and specificity because it was tested at both the pixel and image levels. Also, a CNN model is used to train the model at the pixel level, and after that, a classification model is applied to detect bleeding. The current feature extraction methods have certain limitations such as introducing bias (which is influenced by the chosen algorithm), increased complexity, overfitting, and reduced generalizability
The majority of the literature review proposed various ML algorithms that were trained with the texture and statistical features to identify bleeding in CE images. Before introducing advanced ML algorithms, researchers set a threshold value in the feature extraction values to detect bleeding in CE. The review found from the literature that the maximum number of articles used KNN, SVM, MLP, NN, and CNN algorithms. Besides these algorithms, a few ML algorithms (PCA, Random Tree, Random Forest, fuzzy means, Expectation Maximization clustering, and Supported Convex Hull) were used, which are called in this study "Other ML". Also, a few DL algorithms (FCN, SegNet, U-Net, DenseNet, ResNet50, VGGNet, and MobileNet) were used, which are referred to as "Other DL". But the performance of the ML technique is based on the extracted features of color channels. The color channel intensity values overlapped between bleeding and non-bleeding pixels. As a result, utilizing ML approaches to distinguish between bleeding and non-bleeding is problematic. In the last few years, researchers proposed a DL technique to identify bleeding from CE images. DL is an end-to-end classification and segmentation approach that extracted features automatically at the pixel level. Unlike ML, the DL approach does not require a separate feature extraction stage and it extracts features automatically to provide more efficient outcomes. The performance results for different state-of-the-art ML and DL algorithms are shown in Figure 6. using a box plot. From the figure, it is observed that both the KNN and DL algorithms outperformed compared to the other algorithms. While the existing methods have made significant contributions to the development of classification and segmentation algorithms for detecting bleeding with satisfactory performance, they have often been tested on a limited number of test samples such as images. Furthermore, deep learning algorithms have not yet incorporated attention mechanisms to further enhance their performance
The most commonly used color spaces for available state-of-the-art bleeding detection algorithms from CE images are RGB, HSV, YIQ, YCbCr, CIE Lab, CIE XYZ, K-L, etc. Among them from the above review we can see, the RGB space has a slightly higher recall value and achieved lower variance for accuracy, recall, and specificity. A computer-aided system also improves execution speed because no color conversion operations are required when using the RGB color space. So, for practical use RGB color space is the best option there is no need to convert them into other color domains. For the feature extraction method, the combined global and local feature extraction domain showed greater detection accuracy compared to the individual domain. Which makes them more suitable for practical use. For bleeding detection from CE images, most of the literature proposed ML algorithms, which include SVM, KNN, PCA, MLP, NN, Random Tree, Random Forest, etc. Also, the maximum number of articles for DL, used CNN, FCN, SegNet, U-Net, DenseNet, ResNet50, VGGNet, and MobileNet. From the above review of the literature, KNN and DL algorithms outperformed compared to the other algorithms. For ML algorithms the color channel intensity values overlapped between bleeding and non-bleeding pixels. While DL approach does not require a separate feature extraction stage and it extracts features automatically, which is more effective for practical use.

5. Future Direction

To be suitable for practical use and deliver robust performance on unknown test data, the current methods require improvement. This improvement should include minimizing computational requirements and representing the outcomes visually to ensure that clinicians can trust the results. In the above review, color space does not provide any performance benefits for classification or segmentation tasks. The RGB color space provides higher execution speed with lower performance variance, and it should be used in the future. The DL algorithms provide a promising path for practical computer-aided bleeding detection systems in CE. The DL algorithms depend on the number of medical image datasets. However, the insufficient quantity of data is a limitation. Recently, a new technique called generative adversarial networks (GAN) has been introduced to generate synthetic images [129,130,131,132]. So, in the future, the dataset limitation can be overcome by collecting more data and with synthetic CE images. And DL algorithms may be a better alternative for detecting bleeding in CE images. Several articles introduced ML and DL algorithms concurrently, such as DL for feature extraction and ML for bleeding image categorization. The integrated models considerably improved the performance of the computer-aided bleeding detection system. From the literature review, the majority of articles used the same patient or subject data for training and testing datasets as they randomly selected images for training and testing. It introduces bias in an experimental setup. Same-patient’s CE images should not be mixed with training and test folders for classifying bleeding.

6. Conclusion

A systematic review of the available state-of-the-art computer-aided bleeding detection algorithms for capsule endoscopy (CE) was conducted, and the most accurate and suitable algorithms for practical use was identified. This review suggested a taxonomy for computer-aided bleeding detection systems. Researchers used various color spaces and feature extraction techniques to boost the bleeding detection algorithm's performance. The analysis revealed that the color space offers no additional benefits. Considering simplicity RGB color space is preferred. For the feature extraction, combining (both global and local feature extraction domains) in a computer-aided bleeding detection system significantly enhanced the detection accuracy compared to the individual domain. The k-nearest neighbor (KNN) and deep learning algorithms outperformed compared to the other algorithms for computer-aided bleeding detection systems. However, the KNN faces a few limitations, like overfitting problems and required hand-crafted feature extraction. Recently, the computer-aided bleeding detection system focuses on deep learning algorithms. The performance of the deep learning bleeding detection algorithm is improving day by day. In the future, the deep learning algorithm will be a promising path for computer-aided bleeding detection systems in capsule endoscopy.

Author Contributions

Conceptualization, Tonmoy Ghosh and Abdullah Al Mamun; methodology, Ahmmad Musha and Rehnuma Hasnat; validation, Tonmoy Ghosh; formal analysis, Ahmmad Musha; investigation, Abdullah Al Mamun; data curation, Ahmmad Musha; writing—original draft preparation, Ahmmad Musha and Abdullah Al Mamun; writing—review and editing, Rehnuma Hasnat and Abdullah Al Mamun; visualization, Ahmmad Musha; supervision, Tonmoy Ghosh and Em Poh Ping; project administration, Tonmoy Ghosh; funding acquisition, Em Poh Ping. All authors have read and agreed to the published version of the manuscript.

Funding

MMUE/220021 (TM R&D) and FRGS/1/2022/TK0/MMU/02/13 (FRGS)

Conflicts of Interest

The authors state that there are no potential conflicts of interest.

References

  1. Al Mamun, P. P. Em, T. Ghosh, M. M. Hossain, M. G. Hasan, and M. G. Sadeque, “Bleeding recognition technique in wireless capsule endoscopy images using fuzzy logic and principal component analysis,” Int. J. Electr. Comput. Eng., vol. 11, no. 3, pp. 2689–2696, 2021. [CrossRef]
  2. S. Monteiro, F. D. De Castro, P. B. Carvalho, M. J. Moreira, B. Rosa, and J. Cotter, “PillCam® SB3 capsule: Does the increased frame rate eliminate the risk of missing lesions?,” World J. Gastroenterol., vol. 22, no. 10, pp. 3066–3068, Mar. 2016. [CrossRef]
  3. S. Fan, L. Xu, Y. Fan, K. Wei, and L. Li, “Computer-aided detection of small intestinal ulcer and erosion in wireless capsule endoscopy images,” Phys. Med. Biol., vol. 63, no. 16, Aug. 2018. [CrossRef]
  4. K. Pogorelov et al., “Bleeding detection in wireless capsule endoscopy videos — Color versus texture features,” J. Appl. Clin. Med. Phys., vol. 20, no. 8, pp. 141–154, Aug. 2019. [CrossRef]
  5. Karargyris and N. Bourbakis, “Detection of small bowel polyps and ulcers in wireless capsule endoscopy videos,” IEEE Trans. Biomed. Eng., vol. 58, no. 10 PART 1, pp. 2777–2786, Oct. 2011. [CrossRef]
  6. G. R. Zuckerman, C. Prakash, M. P. Askin, and B. S. Lewis, “AGA technical review on the evaluation and management of occult and obscure gastrointestinal bleeding,” Gastroenterology, vol. 118, no. 1, pp. 201–221, 2000. [CrossRef]
  7. Y. Lee, G. Yoon, and A. O. Architecture, “for Capsule Endoscopy,” pp. 672–677, 2011.
  8. S. C. Park et al., “Sensitivity of the suspected blood indicator: An experimental study,” World J. Gastroenterol., vol. 18, no. 31, pp. 4169–4174, 2012. [CrossRef]
  9. Liaqat, M. A. Khan, J. H. Shah, M. Sharif, M. Yasmin, and S. L. Fernandes, “AUTOMATED ULCER and BLEEDING CLASSIFICATION from WCE IMAGES USING MULTIPLE FEATURES FUSION and SELECTION,” J. Mech. Med. Biol., vol. 18, no. 4, Jun. 2018. [CrossRef]
  10. Al Mamun, M. S. Hossain, P. P. Em, A. Tahabilder, R. Sultana, and M. A. Islam, “Small intestine bleeding detection using color threshold and morphological operation in WCE images,” Int. J. Electr. Comput. Eng., vol. 11, no. 4, pp. 3040–3048, Aug. 2021. [CrossRef]
  11. Al Mamun, M. S. Hossain, M. M. Hossain, and M. G. Hasan, “Discretion Way for Bleeding Detection in Wireless Capsule Endoscopy Images,” May 2019. [CrossRef]
  12. H. Neumann, L. C. Fry, A. Nägel, and M. F. Neurath, “Wireless capsule endoscopy of the small intestine: a review with future directions.,” Curr. Opin. Gastroenterol., vol. 30, no. 5, pp. 463–471, 2014. [CrossRef]
  13. Karargyris and N. Bourbakis, “Wireless Capsule Endoscopy and Endoscopic Imaging: A Survey on Various Methodologies Presented,” IEEE Eng. Med. Biol. Mag., vol. 29, no. 1, pp. 72–83, 2010. [CrossRef]
  14. H. P. Brito et al., “Video capsule endoscopy vs double-balloon enteroscopy in the diagnosis of small bowel bleeding: A systematic review and meta-analysis,” World J. Gastrointest. Endosc., vol. 10, no. 12, pp. 400–421, 2018. [CrossRef]
  15. Koulaouzidis, E. Rondonotti, and A. Karargyris, “Small-bowel capsule endoscopy: A ten-point contemporary review,” World J. Gastroenterol., vol. 19, no. 24, pp. 3726–3746, 2013. [CrossRef]
  16. D. K. Iakovidis and A. Koulaouzidis, “Software for enhanced video capsule endoscopy: challenges for essential progress.,” Nat. Rev. Gastroenterol. Hepatol., vol. 12, no. 3, pp. 172–186, 2015. [CrossRef]
  17. Y. Hwang, J. Park, Y. J. Lim, and H. J. Chun, “Application of artificial intelligence in capsule endoscopy: Where are we now?,” Clin. Endosc., vol. 51, no. 6, pp. 547–551, 2018. [CrossRef]
  18. Y. Chen and J. Lee, “A review of machine-vision-based analysis of wireless capsule endoscopy video,” Diagn. Ther. Endosc., vol. 2012, 2012. [CrossRef]
  19. N. Shah et al., “Video capsule endoscopy for upper gastrointestinal hemorrhage in the emergency department: A systematic review and meta-analysis,” Am. J. Emerg. Med., vol. 38, no. 6, pp. 1245–1252, 2020. [CrossRef]
  20. S. Soffer et al., “Deep learning for wireless capsule endoscopy: a systematic review and meta-analysis,” Gastrointest. Endosc., vol. 92, no. 4, pp. 831-839.e8, 2020. [CrossRef]
  21. S. Suman, F. A. B. Hussin, N. Walter, A. S. Malik, S. H. Ho, and K. L. Goh, “Detection and classification of bleeding using statistical color features for wireless capsule endoscopy images,” 2017. [CrossRef]
  22. Z. Amiri, H. Hassanpour, and A. Beghdadi, “Feature Selection for Bleeding Detection in Capsule Endoscopy Images using Genetic Algorithm,” 2019. [CrossRef]
  23. Caroppo, A. Leone, and P. Siciliano, “Deep transfer learning approaches for bleeding detection in endoscopy images.,” Comput. Med. Imaging Graph., vol. 88, p. 101852, 2021. [CrossRef]
  24. X. Xing, Y. Yuan, X. Jia, and Q. H. M. Max, “A saliency-aware hybrid dense network for bleeding detection in wireless capsule endoscopy images,” in Proceedings - International Symposium on Biomedical Imaging, 2019, vol. 2019-April, pp. 104–107. [CrossRef]
  25. Y. Fu, M. Mandal, and G. Guo, “Bleeding region detection in WCE images based on color features and neural network,” 2011. [CrossRef]
  26. Y. Xiong, Y. Zhu, Z. Pang, Y. Ma, D. Chen, and X. Wang, “Bleeding detection in wireless capsule endoscopy based on MST clustering and SVM,” in IEEE Workshop on Signal Processing Systems, SiPS: Design and Implementation, 2015, vol. 2015-Decem. [CrossRef]
  27. X. Jia and M. Q.-H. Meng, “A study on automated segmentation of blood regions in Wireless Capsule Endoscopy images using fully convolutional networks,” in Proceedings - International Symposium on Biomedical Imaging, 2017, pp. 179–182. [CrossRef]
  28. P. Coelho, A. Pereira, A. Leite, M. Salgado, and A. Cunha, “A Deep Learning Approach for Red Lesions Detection in Video Capsule Endoscopies,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 10882 LNCS, pp. 553–561, 2018. [CrossRef]
  29. M. A. Usman, G. B. Satrya, M. R. Usman, and S. Y. Shin, “Detection of small colon bleeding in wireless capsule endoscopy videos.,” Comput. Med. Imaging Graph., vol. 54, pp. 16–26, 2016. [CrossRef]
  30. M. A. Usman et al., “QI-BRiCE: Quality index for bleeding regions in capsule endoscopy videos,” Comput. Mater. Contin., vol. 67, no. 2, pp. 1697–1712, 2021. [CrossRef]
  31. Novozámský, J. Flusser, I. Tachecí, L. Sulík, J. Bureš, and O. Krejcar, “Automatic blood detection in capsule endoscopy video.,” J. Biomed. Opt., vol. 21, no. 12, p. 126007, 2016. [CrossRef]
  32. S. Sainju, F. M. Bui, and K. A. Wahid, “Automated bleeding detection in capsule endoscopy videos using statistical features and region growing.,” J. Med. Syst., vol. 38, no. 4, p. 25, 2014. [CrossRef]
  33. F. Deeba, M. Islam, F. M. Bui, and K. A. Wahid, “Performance assessment of a bleeding detection algorithm for endoscopic video based on classifier fusion method and exhaustive feature selection,” Biomed. Signal Process. Control, vol. 40, pp. 415–424, 2018. [CrossRef]
  34. M. Mackiewicz, M. Fisher, and C. Jamieson, “Bleeding detection in wireless capsule endoscopy using adaptive colour histogram model and support vector classification,” in Progress in Biomedical Optics and Imaging - Proceedings of SPIE, 2008, vol. 6914. [CrossRef]
  35. T. Ghosh, S. A. Fattah, and K. A. Wahid, “Automatic Computer Aided Bleeding Detection Scheme for Wireless Capsule Endoscopy (WCE) Video Based on Higher and Lower Order Statistical Features in a Composite Color,” J. Med. Biol. Eng., vol. 38, no. 3, pp. 482–496, 2018. [CrossRef]
  36. T. Ghosh, S. A. Fattah, and K. A. Wahid, “CHOBS: Color Histogram of Block Statistics for Automatic Bleeding Detection in Wireless Capsule Endoscopy Video.,” IEEE J. Transl. Eng. Heal. Med., vol. 6, p. 1800112, 2018. [CrossRef]
  37. Q. Zhao, M. Q.-H. Meng, and B. Li, “WCE video clips segmentation based on abnormality,” in 2010 IEEE International Conference on Robotics and Biomimetics, ROBIO 2010, 2010, pp. 442–447. [CrossRef]
  38. T. Ghosh, S. A. Fattah, K. A. Wahid, W.-P. Zhu, and M. O. Ahmad, “Cluster based statistical feature extraction method for automatic bleeding detection in wireless capsule endoscopy video.,” Comput. Biol. Med., vol. 94, pp. 41–54, 2018. [CrossRef]
  39. W. Shi, J. Chen, H. Chen, Q. Peng, and T. Gan, “Bleeding fragment localization using time domain information for WCE videos,” in Proceedings - 2015 8th International Conference on BioMedical Engineering and Informatics, BMEI 2015, 2016, pp. 73–78. [CrossRef]
  40. R. Leenhardt et al., “A neural network algorithm for detection of GI angiectasia during small-bowel capsule endoscopy.,” Gastrointest. Endosc., vol. 89, no. 1, pp. 189–194, 2019. [CrossRef]
  41. F. Rustam et al., “Wireless Capsule Endoscopy Bleeding Images Classification Using CNN Based Model,” IEEE Access, vol. 9, pp. 33675–33688, 2021. [CrossRef]
  42. E. Tuba, M. Tuba, and R. Jovanovic, “An algorithm for automated segmentation for bleeding detection in endoscopic images,” in Proceedings of the International Joint Conference on Neural Networks, 2017, vol. 2017-May, pp. 4579–4586. [CrossRef]
  43. S. Li, J. Zhang, C. Ruan, and Y. Zhang, “Multi-Stage Attention-Unet for Wireless Capsule Endoscopy Image Bleeding Area Segmentation,” in Proceedings - 2019 IEEE International Conference on Bioinformatics and Biomedicine, BIBM 2019, 2019, pp. 818–825. [CrossRef]
  44. S. Rathnamala and S. Jenicka, “Automated bleeding detection in wireless capsule endoscopy images based on color feature extraction from Gaussian mixture model superpixels.,” Med. Biol. Eng. Comput., 2021. [CrossRef]
  45. T. Ghosh and J. Chakareski, “Deep Transfer Learning for Automated Intestinal Bleeding Detection in Capsule Endoscopy Imaging.,” J. Digit. Imaging, 2021. [CrossRef]
  46. S. Suman et al., “Detection and classification of bleeding region in WCE images using color feature,” in ACM International Conference Proceeding Series, 2017, vol. Part F1301. [CrossRef]
  47. J. Liu and X. Yuan, “Obscure bleeding detection in endoscopy images using support vector machines,” Optim. Eng., vol. 10, no. 2, pp. 289–299, 2009. [CrossRef]
  48. J. Li, J. Ma, T. Tillo, B. Zhang, and E. G. Lim, “A training based Support Vector Machine technique for blood detection in wireless capsule endoscopy images,” in 2012 IEEE-EMBS Conference on Biomedical Engineering and Sciences, IECBES 2012, 2012, pp. 826–830. [CrossRef]
  49. S. Sainju, F. M. Bui, and K. Wahid, “Bleeding detection in wireless capsule endoscopy based on color features from histogram probability,” 2013. [CrossRef]
  50. G. Pan, F. Xu, and J. Chen, “Bleeding detection in wireless capsule endoscopy using color similarity coefficient,” Appl. Mech. Mater., vol. 195–196, pp. 307–312, 2012. [CrossRef]
  51. S. J. Yun, H. K. Young, H. L. Dong, H. L. Sang, J. S. Jeong, and H. K. Jong, “Automatic patient-adaptive bleeding detection in a capsule endoscopy,” in Progress in Biomedical Optics and Imaging - Proceedings of SPIE, 2009, vol. 7260. [CrossRef]
  52. T. Ghosh, S. A. Fattah, C. Shahnaz, and K. A. Wahid, “An automatic bleeding detection scheme in wireless capsule endoscopy based on histogram of an RGB-indexed image.,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Int. Conf., vol. 2014, pp. 4683–4686, 2014. [CrossRef]
  53. P. Sivakumar and B. M. Kumar, “A novel method to detect bleeding frame and region in wireless capsule endoscopy video,” Cluster Comput., vol. 22, pp. 12219–12225, 2019. [CrossRef]
  54. J. Ma, T. Tillo, B. Zhang, Z. Wang, and E. G. Lim, “Novel training and comparison method for blood detection in wireless capsule endoscopy images,” in International Symposium on Medical Information and Communication Technology, ISMICT, 2013, pp. 56–60. [CrossRef]
  55. T. Ghosh, S. K. Bashar, M. S. Alam, K. Wahid, and S. A. Fattah, “A statistical feature based novel method to detect bleeding in wireless capsule endoscopy images,” 2014. [CrossRef]
  56. T. Ghosh, S. A. Fattah, C. Shahnaz, A. K. Kundu, and M. N. Rizve, “Block based histogram feature extraction method for bleeding detection in wireless capsule endoscopy,” in IEEE Region 10 Annual International Conference, Proceedings/TENCON, 2016, vol. 2016-Janua. [CrossRef]
  57. T. Ghosh, S. A. Fattah, and K. A. Wahid, “Automatic bleeding detection in wireless capsule endoscopy based on RGB pixel intensity ratio,” 2014. [CrossRef]
  58. Y. Fu, W. Zhang, M. Mandal, and M. Q.-H. Meng, “Computer-aided bleeding detection in WCE video.,” IEEE J. Biomed. Heal. informatics, vol. 18, no. 2, pp. 636–642, 2014. [CrossRef]
  59. Z. Liu, C. Hu, and Z. Shen, “Research on a new feature detection algorithm for wireless capsule endoscope bleeding images based on super-pixel segmentation,” in IEEE International Conference on Robotics and Biomimetics, ROBIO 2019, 2019, pp. 1744–1749. [CrossRef]
  60. S. Kumar, I. N. Figueiredo, C. Graca, and G. Falcao, “A GPU accelerated algorithm for blood detection inwireless capsule endoscopy images,” Lect. Notes Comput. Vis. Biomech., vol. 19, pp. 55–71, 2015. [CrossRef]
  61. K. Kundu, M. N. Rizve, T. Ghosh, and S. A. Fattah, “A segmented color plane histogram based feature extraction scheme for automatic bleeding detection in wireless capsule endoscopy,” in 2016 IEEE Students’ Technology Symposium, TechSym 2016, 2017, pp. 245–249. [CrossRef]
  62. Y. Yuan and M. Q. H. Meng, “Automatic bleeding frame detection in the wireless capsule endoscopy images,” in Proceedings - IEEE International Conference on Robotics and Automation, 2015, vol. 2015-June, no. June, pp. 1310–1315. [CrossRef]
  63. X. Jia, L. Cai, J. Liu, W. Dai, and M. Q.-H. Meng, “GI bleeding detection in wireless capsule endoscopy images based on pattern recognition and a MapReduce framework,” in 2016 IEEE International Conference on Real-Time Computing and Robotics, RCAR 2016, 2016, pp. 266–271. [CrossRef]
  64. S. Hwang, J. Oh, J. Cox, S. J. Tang, and H. F. Tibbals, “Blood detection in wireless capsule endoscopy using expectation maximization clustering,” in Progress in Biomedical Optics and Imaging - Proceedings of SPIE, 2006, vol. 6144 I. [CrossRef]
  65. K. Poh et al., “Multi-level local feature classification for bleeding detection in Wireless Capsule Endoscopy images,” in 2010 IEEE Conference on Cybernetics and Intelligent Systems, CIS 2010, 2010, pp. 76–81. [CrossRef]
  66. P. Pons, R. Noorda, A. Nevárez, A. Colomer, V. P. Beltrán, and V. Naranjo, “Design and Development of an Automatic Blood Detection System for Capsule Endoscopy Images,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 11872 LNCS, pp. 105–113, 2019. [CrossRef]
  67. Dilna and V. P. Gopi, “A novel method for bleeding detection in Wireless Capsule Endoscopic images,” in 2015 International Conference on Computing and Network Communications, CoCoNet 2015, 2016, pp. 854–858. [CrossRef]
  68. K. R. Reeha, K. Shailaja, and V. P. Gopi, “Undecimated Complex Wavelet Transform based bleeding detection for endoscopic images,” in Proceedings - 2016 2nd International Conference on Cognitive Computing and Information Processing, CCIP 2016, 2016. [CrossRef]
  69. S. Charfi and M. El Ansari, “Gastrointestinal tract bleeding detection from wireless capsule endoscopy videos,” in ACM International Conference Proceeding Series, 2017. [CrossRef]
  70. Giritharan, X. Yuan, J. Liu, B. Buckles, J. Oh, and S. J. Tang, “Bleeding detection from capsule endoscopy videos.,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Int. Conf., vol. 2008, pp. 4780–4783, 2008. [CrossRef]
  71. Bchir, M. M. Ben Ismail, and N. AlZahrani, “Multiple bleeding detection in wireless capsule endoscopy,” Signal, Image Video Process., vol. 13, no. 1, pp. 121–126, 2019. [CrossRef]
  72. P. Y. Lau and P. L. Correia, “Detection of bleeding patterns in WCE video using multiple features.,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Int. Conf., vol. 2007, pp. 5601–5604, 2007. [CrossRef]
  73. Kukushkin et al., “Recognition of hemorrhage in the images of wireless capsule endoscopy,” in Proceedings of the Mediterranean Electrotechnical Conference - MELECON, 2012, pp. 899–902. [CrossRef]
  74. K. Poh, Z. Zhang, Z. Y. Liang, L. Li, and J. Liu, “Feature selection and classification for wireless capsule endoscopic frames,” 2009. [CrossRef]
  75. K. Poh, Z. Zhang, Z. Y. Liang, L. Li, and J. Liu, “Feature selection and classification for wireless capsule endoscopic frames,” 2009. [CrossRef]
  76. T. Ghosh, S. K. Bashar, S. A. Fattah, C. Shahnaz, and K. A. Wahid, “An automatic bleeding detection scheme in wireless capsule endoscopy based on statistical features in hue space,” in 2014 17th International Conference on Computer and Information Technology, ICCIT 2014, 2003, pp. 354–357. [CrossRef]
  77. K. Timotius, S.-G. Miaou, E. B. Valdeavilla, and Y.-H. Liu, “Abnormality detection for capsule endoscope images based on support vector machines,” Biomed. Eng. - Appl. Basis Commun., vol. 24, no. 1, pp. 71–83, 2012. [CrossRef]
  78. S. Zhou, X. Song, M. A. Siddique, J. Xu, and P. Zhou, “Bleeding detection in wireless capsule endoscopy images based on binary feature vector,” in 5th International Conference on Intelligent Control and Information Processing, ICICIP 2014 - Proceedings, 2015, pp. 29–33. [CrossRef]
  79. L. Cui, C. Hu, Y. Zou, and M. Q.-H. Meng, “Bleeding detection in wireless capsule endoscopy images by support vector classifier,” in 2010 IEEE International Conference on Information and Automation, ICIA 2010, 2010, pp. 1746–1751. [CrossRef]
  80. T. Ghosh et al., “An automatic bleeding detection technique in wireless capsule endoscopy from region of interest,” in International Conference on Digital Signal Processing, DSP, 2015, vol. 2015-Septe, pp. 1293–1297. [CrossRef]
  81. K. Kundu, M. N. Rizve, T. Ghosh, S. A. Fattah, and C. Shahnaz, “A histogram based scheme in YIQ domain for automatic bleeding image detection from wireless capsule endoscopy,” in 2015 IEEE International WIE Conference on Electrical and Computer Engineering, WIECON-ECE 2015, 2016, pp. 455–458. [CrossRef]
  82. D.-Y. Liu et al., “Identification of lesion images from gastrointestinal endoscope based on feature extraction of combinational methods with and without learning process.,” Med. Image Anal., vol. 32, pp. 281–294, 2016. [CrossRef]
  83. Y. Yuan, B. Li, and M. Q.-H. Meng, “Bleeding Frame and Region Detection in the Wireless Capsule Endoscopy Video.,” IEEE J. Biomed. Heal. informatics, vol. 20, no. 2, pp. 624–630, 2016. [CrossRef]
  84. N. Figueiredo, S. Kumar, C. Leal, and P. N. Figueiredo, “Computer-assisted bleeding detection in wireless capsule endoscopy images,” Comput. Methods Biomech. Biomed. Eng. Imaging Vis., vol. 1, no. 4, pp. 198–210, 2013. [CrossRef]
  85. M. Mathew and V. P. Gopi, “Transform based bleeding detection technique for endoscopic images,” in 2nd International Conference on Electronics and Communication Systems, ICECS 2015, 2015, pp. 1730–1734. [CrossRef]
  86. Karargyris and N. Bourbakis, “A methodology for detecting blood-based abnormalities in wireless capsule endoscopy videos,” 2008. [CrossRef]
  87. X. Liu, J. Gu, Y. Xie, J. Xiong, and W. Qin, “A new approach to detecting ulcer and bleeding in wireless capsule endoscopy images,” in Proceedings - IEEE-EMBS International Conference on Biomedical and Health Informatics: Global Grand Challenge of Health Informatics, BHI 2012, 2012, pp. 737–740. [CrossRef]
  88. Priyadharshini and T. Gomathi, “Navie bayes classifier for wireless capsule endoscopy video to detect bleeding frames,” Int. J. Sci. Technol. Res., vol. 9, no. 1, pp. 3286–3291, 2020, [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093909545&partnerID=40&md5=6215a2e2a1ae9c0f862eb0f35c40631c.
  89. T. Ghosh, S. K. Bashar, S. A. Fattah, C. Shahnaz, and K. A. Wahid, “A feature extraction scheme from region of interest of wireless capsule endoscopy images for automatic bleeding detection,” 2015. [CrossRef]
  90. S. K. Mohammed, F. Deeba, F. M. Bui, and K. A. Wahid, “Application of modified ant colony optimization for computer aided bleeding detection system,” in Proceedings of the International Joint Conference on Neural Networks, 2016, vol. 2016-Octob, pp. 4317–4324. [CrossRef]
  91. X. Xing, X. Jia, and M.-H. Q. Meng, “Bleeding Detection in Wireless Capsule Endoscopy Image Video Using Superpixel-Color Histogram and a Subspace KNN Classifier.,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Int. Conf., vol. 2018, pp. 1–4, 2018. [CrossRef]
  92. Y. S. Jung, Y. H. Kim, D. H. Lee, and J. H. Kim, “Active blood detection in a high resolution capsule endoscopy using color spectrum transformation,” in BioMedical Engineering and Informatics: New Development and the Future - Proceedings of the 1st International Conference on BioMedical Engineering and Informatics, BMEI 2008, 2008, vol. 1, pp. 859–862. [CrossRef]
  93. G. Lv, G. Yan, and Z. Wang, “Bleeding detection in wireless capsule endoscopy images based on color invariants and spatial pyramids using support vector machines.,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Int. Conf., vol. 2011, pp. 6643–6646, 2011. [CrossRef]
  94. G. Pan, G. Yan, X. Song, and X. Qiu, “BP neural network classification for bleeding detection in wireless capsule endoscopy.,” J. Med. Eng. Technol., vol. 33, no. 7, pp. 575–581, 2009. [CrossRef]
  95. G. Pan, G. Yan, X. Qiu, and J. Cui, “Bleeding detection in Wireless Capsule Endoscopy based on Probabilistic Neural Network.,” J. Med. Syst., vol. 35, no. 6, pp. 1477–1484, 2010. [CrossRef]
  96. S. K. Mohammed, F. Deeba, F. M. Bui, and K. A. Wahid, “Feature selection using modified ant colony optimization for wireless capsule endoscopy,” in 2016 IEEE 7th Annual Ubiquitous Computing, Electronics and Mobile Communication Conference, UEMCON 2016, 2016. [CrossRef]
  97. K. Iakovidis, D. Chatzis, P. Chrysanthopoulos, and A. Koulaouzidis, “Blood detection in wireless capsule endoscope images based on salient superpixels.,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Int. Conf., vol. 2015, pp. 731–734, 2015. [CrossRef]
  98. N. Obukhova, A. Motyko, B. Timofeev, and A. Pozdeev, “Method of endoscopic images analysis for automatic bleeding detection and segmentation,” in Conference of Open Innovation Association, FRUCT, 2019, vol. 2019-April, pp. 285–290. [CrossRef]
  99. S. Yi et al., “A clinically viable Capsule Endoscopy video analysis platform for automatic bleeding detection,” in Proceedings of SPIE - The International Society for Optical Engineering, 2013, vol. 8670. [CrossRef]
  100. K. Kundu and S. A. Fattah, “Probability density function based modeling of spatial feature variation in capsule endoscopy data for automatic bleeding detection.,” Comput. Biol. Med., vol. 115, p. 103478, 2019. [CrossRef]
  101. Z. Amiri, H. Hassanpour, and A. Beghdadi, “A Computer- Aided Method to Detect Bleeding Frames in Capsule Endoscopy Images,” in Proceedings - European Workshop on Visual Information Processing, EUVIP, 2019, vol. 2019-Octob, pp. 217–221. [CrossRef]
  102. P. Szczypiński, A. Klepaczko, M. Pazurek, and P. Daniel, “Texture and color based image segmentation and pathology detection in capsule endoscopy videos,” Comput. Methods Programs Biomed., vol. 113, no. 1, pp. 396–411, 2014. [CrossRef]
  103. K. Kundu, S. A. Fattah, and K. A. Wahid, “Least Square Saliency Transformation of Capsule Endoscopy Images for PDF Model Based Multiple Gastrointestinal Disease Classification,” IEEE Access, vol. 8, pp. 58509–58521, 2020. [CrossRef]
  104. Tuba, S. Tomic, M. Beko, D. Zivkovic, and M. Tuba, “Bleeding Detection in Wireless Capsule Endoscopy Images Using Texture and Color Features,” in 2018 26th Telecommunications Forum (TELFOR), 2018, pp. 1–4. [CrossRef]
  105. Li and M. Q.-H. Meng, “Computer aided detection of bleeding in capsule endoscopy images,” in Canadian Conference on Electrical and Computer Engineering, 2008, pp. 1963–1966. [CrossRef]
  106. Pogorelov et al., “Bleeding detection in wireless capsule endoscopy videos - Color versus texture features.,” J. Appl. Clin. Med. Phys., vol. 20, no. 8, pp. 141–154, 2019. [CrossRef]
  107. R. Ponnusamy and S. Sathiamoorthy, “An efficient gastrointestinal hemorrhage detection and diagnosis model for wireless capsule endoscopy,” Int. J. Recent Technol. Eng., vol. 8, no. 3, pp. 7549–7554, 2019. [CrossRef]
  108. R. Hassan and M. A. Haque, “Computer-aided gastrointestinal hemorrhage detection in wireless capsule endoscopy videos.,” Comput. Methods Programs Biomed., vol. 122, no. 3, pp. 341–353, 2015. [CrossRef]
  109. H. Maghsoudi, M. Alizadeh, and M. Mirmomen, “A computer aided method to detect bleeding, tumor, and disease regions in Wireless Capsule Endoscopy,” 2017. [CrossRef]
  110. Patel, K. Rani, S. Kumar, I. N. Figueiredo, and P. N. Figueiredo, “Automated bleeding detection in wireless capsule endoscopy images based on sparse coding,” Multimed. Tools Appl., 2020. [CrossRef]
  111. Joshi, S. Kumar, and I. N. Figueiredo, “Bag of visual words approach for bleeding detection in wireless capsule endoscopy images,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 9730, pp. 575–582, 2016. [CrossRef]
  112. Li and M. Q.-H. Meng, “Computer-aided detection of bleeding regions for capsule endoscopy images.,” IEEE Trans. Biomed. Eng., vol. 56, no. 4, pp. 1032–1039, 2009. [CrossRef]
  113. Li and M. Q.-H. Meng, “Computer-based detection of bleeding and ulcer in wireless capsule endoscopy images by chromaticity moments.,” Comput. Biol. Med., vol. 39, no. 2, pp. 141–147, 2009. [CrossRef]
  114. X. Jia and M. Q.-H. Meng, “Gastrointestinal bleeding detection in wireless capsule endoscopy images using handcrafted and CNN features.,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Int. Conf., vol. 2017, pp. 3154–3157, 2017. [CrossRef]
  115. M. D. Vasilakakis, D. Diamantis, E. Spyrou, A. Koulaouzidis, and D. K. Iakovidis, “Weakly supervised multilabel classification for semantic interpretation of endoscopy video frames,” Evol. Syst., vol. 11, no. 3, pp. 409–421, 2020. [CrossRef]
  116. M. Hajabdollahi, R. Esfandiarpoor, K. Najarian, N. Karimi, S. Samavi, and S. M. Reza Soroushmehr, “Low Complexity CNN Structure for Automatic Bleeding Zone Detection in Wireless Capsule Endoscopy Imaging.,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Int. Conf., vol. 2019, pp. 7227–7230, 2019. [CrossRef]
  117. M. Hajabdollahi, R. Esfandiarpoor, E. Sabeti, N. Karimi, S. M. R. Soroushmehr, and S. Samavi, “Multiple abnormality detection for automatic medical image diagnosis using bifurcated convolutional neural network,” Biomed. Signal Process. Control, vol. 57, 2020. [CrossRef]
  118. H. S. Pannu, S. Ahuja, N. Dang, S. Soni, and A. K. Malhi, “Deep learning based image classification for intestinal hemorrhage,” Multimed. Tools Appl., vol. 79, no. 29–30, pp. 21941–21966, 2020. [CrossRef]
  119. E. Diamantis, D. K. Iakovidis, and A. Koulaouzidis, “Look-behind fully convolutional neural network for computer-aided endoscopy,” Biomed. Signal Process. Control, vol. 49, pp. 192–201, 2019. [CrossRef]
  120. P. Li, Z. Li, F. Gao, L. Wan, and J. Yu, “Convolutional neural networks for intestinal hemorrhage detection in wireless capsule endoscopy images,” in Proceedings - IEEE International Conference on Multimedia and Expo, 2017, pp. 1518–1523. [CrossRef]
  121. T. Ghosh, L. Li, and J. Chakareski, “Effective Deep Learning for Semantic Segmentation Based Bleeding Zone Detection in Capsule Endoscopy Images,” in Proceedings - International Conference on Image Processing, ICIP, 2018, pp. 3034–3038. [CrossRef]
  122. T. Aoki et al., “Automatic detection of blood content in capsule endoscopy images based on a deep convolutional neural network.,” J. Gastroenterol. Hepatol., vol. 35, no. 7, pp. 1196–1200, 2020. [CrossRef]
  123. Y. Hwang et al., “Improved classification and localization approach to small bowel capsule endoscopy using convolutional neural network.,” Dig. Endosc., 2020. [CrossRef]
  124. Tsuboi et al., “Artificial intelligence using a convolutional neural network for automatic detection of small-bowel angioectasia in capsule endoscopy images.,” Dig. Endosc., vol. 32, no. 3, pp. 382–390, 2020. [CrossRef]
  125. M. Sharif, M. Attique Khan, M. Rashid, M. Yasmin, F. Afza, and U. J. Tanik, “Deep CNN and geometric features-based gastrointestinal tract diseases detection and classification from wireless capsule endoscopy images,” J. Exp. Theor. Artif. Intell., 2019. [CrossRef]
  126. M. Hajabdollahi et al., “Segmentation of bleeding regions in wireless capsule endoscopy for detection of informative frames,” Biomed. Signal Process. Control, vol. 53, 2019. [CrossRef]
  127. M. A. Khan, M. Sharif, T. Akram, M. Yasmin, and R. S. Nayak, “Stomach Deformities Recognition Using Rank-Based Deep Features Selection.,” J. Med. Syst., vol. 43, no. 12, p. 329, 2019. [CrossRef]
  128. X. Jia and M. Q.-H. Meng, “A deep convolutional neural network for bleeding detection in Wireless Capsule Endoscopy images.,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Int. Conf., vol. 2016, pp. 639–642, 2016. [CrossRef]
  129. P. Zhuang, A. G. Schwing, and O. Koyejo, “FMRI data augmentation via synthesis,” Proc. - Int. Symp. Biomed. Imaging, vol. 2019-April, pp. 1783–1787, 2019. [CrossRef]
  130. H. Zhao, H. Li, S. Maurer-Stroh, and L. Cheng, “Synthesizing retinal and neuronal images with generative adversarial nets,” Med. Image Anal., vol. 49, pp. 14–26, 2018. [CrossRef]
  131. H. Zhao, H. Li, and L. Cheng, “Synthesizing Filamentary Structured Images with GANs,” pp. 1–10.
  132. V. Bellemo, P. Burlina, L. Yong, T. Y. Wong, and D. S. W. Ting, “Generative Adversarial Networks (GANs) for Retinal Fundus Image Synthesis,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 11367 LNCS, no. March 2020, pp. 289–302, 2019. [CrossRef]
Figure 1. Flowchart of the inclusion and exclusion processes of this systematic review.
Figure 1. Flowchart of the inclusion and exclusion processes of this systematic review.
Preprints 75101 g001
Figure 2. Taxonomy.
Figure 2. Taxonomy.
Preprints 75101 g002
Figure 3. Overview of all studies in this review.
Figure 3. Overview of all studies in this review.
Preprints 75101 g003
Figure 4. Performance of different color spaces.
Figure 4. Performance of different color spaces.
Preprints 75101 g004
Figure 5. Performance of different feature extraction domains.
Figure 5. Performance of different feature extraction domains.
Preprints 75101 g005
Figure 6. Performance of different state-of-the-art algorithms.
Figure 6. Performance of different state-of-the-art algorithms.
Preprints 75101 g006
Table 1. ELIGIBILITY CRITERIA.
Table 1. ELIGIBILITY CRITERIA.
Inclusion Criteria Exclusion Criteria
Articles published in peer-reviewed venues. Articles that do not involve bleeding, lesion, or hemorrhage.
Articles published in the year: 2001-2021 Articles not written in English.
Articles must address a set of keywords: (Bleeding OR Hemorrhage OR blood) AND (Detection OR Segmentation OR Recognition OR Classification) AND (Capsule Endoscopy) Exclude articles for non-human.
Articles that describe an automatic computer-aided bleeding detection for capsule endoscopy. Exclude poster and book chapter.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated