|
1.IntroductionWireless capsule video endoscopy (WCE) is a noninvasive diagnostic tool for small bowel investigation that has been used in clinical praxis since 2001. WCE is very considerate of the patient. The patient swallows a cylindrical plastic capsule of a size about (depending on the manufacturer), which contains a digital video camera, LED light source, signal transmitter, and battery. The capsule travels through the gastrointestinal tract by peristaltic contractions, captures the images, and wirelessly transmits them in real time to an external console worn by the patient. The images are recorded and stored in the console memory and can be uploaded to a computer for a visual inspection or automatic analysis immediately after the monitoring has been completed. Current capsules take frames at a rate between two and six per second on average, which results in many thousands of images (typically up to 60,000) and of video per patient. (Fortunately, current batteries are powerful enough to supply the light source, camera, and transmitter all the time.) The primary use of the capsule endoscopy is to examine areas of the small intestine that are difficult to display by other types of endoscopy. WCE has been successfully applied in the detection of small bowel bleeding sources, Crohns disease, complications of coeliac disease, small bowel tumors, and nonsteroidal anti-inflammatory drug induced enteropathy. The main obstacle in a routine usage of WCE is that the visual evaluation of the video is very time consuming. It is supposed to be done by a trained endoscopist. Even if the software provided by the capsule producers offers certain supporting tools to simplify and speed up the process, it still takes at least 1 h of full concentration of the evaluator. Taking into account that the pathology can be visible only on very few frames and, hence, can be missed easily, the importance of a human factor is apparent. The goal of this paper is to propose a technique for the detection of frames that are suspected to have the presence of bleeding. We do not aim to develop a fully automatic tool for bleeding detection that would replace the doctor. Rather, the method should preprocess the video, identify and export suspected frames, and prepare them for visual inspection, whereas the other frames are skipped and not sent to the inspection at all. This may significantly reduce the evaluation time while the final decision is still left to the endoscopist. This intended goal predetermines the required properties of the method. It should be fast and should provide a high true-positive (TP) rate while keeping the false-positive (FP) rate reasonably low. In Sec. 2, we present a short survey of relevant literature. Section 3 describes the proposed algorithms. In Sec. 4, we present the WCE technical parameters and implementation details of the methods. Section 5 contains experimental evaluation on real data. 2.Literature SurveyThe first research articles on WCE video analysis appeared soon after the WCE had been introduced into clinical praxis. Since the bleeding detection is a very frequent requirement and appears to be easily achievable (which is, however, not generally true as we can see from the literature survey), the largest group of papers has dealt with this challenge. The main problem here is that the blood spots and traces do not have any typical texture and shape, and the blood color may vary widely from light red through dark red to brown, which makes the blood difficult to distinguish from other objects and from the intestinal wall. This diversity in color depends on the types of disease, the bleeding time,1 the position of the capsule, and the surrounding conditions.2 If the parameters have been trained on a certain data collection, the color-based methods perform well on this set, but their generalization to other patients or even to another capsule manufacturer is questionable. However, the color-based features remain the main indicators for the classification. The approaches, adopted by various authors, differ from one another only in the color space in which they are working. Traditional color spaces such as RGB and HSV have been commonly used, but some authors proposed specialized color spaces. Most of the methods extract the features at the pixel level,3–5 others use image blocks,1 and some methods work with the image as a whole.6 The first one who came up with the solution of the blood detection was one of the producers of the capsules, the Israeli medical technology company Given Imaging. Its second-generation capsule had an embedded software for detection of bleeding lesions called the suspected blood indicator (SBI). The SBI detected the “red areas” in the frames and marked the frames as potentially bloody. However, the study by Signorelli et al.7 shows a low performance of SBI, where the TP, true-negative, FP, and false-negative results, were 40.9%, 70.7%, 69.2%, and 42.6%, respectively. Other studies,2,8 reached similar conclusions and confirmed that the SBI does not reduce the time required for interpretation of WCE, which was the main goal. This insufficiency initiated designing of new algorithms for detecting blood presence. They can be categorized into three main groups according to the size of the frame area that they are working with. Pixel-wise methods do not work with any local context and process the individual pixels independently. Their main advantages are their simplicity and speed. Some of them work in the RGB color space with spectrum transformation4 or use a simple thresholding9 with the threshold found by a support vector machine. However, as we will discuss below in Sec. 3.1, RGB space does not exhibit a good discriminability in tests on more patients. Shahet al.3 proposed using only the hue component of the hue, saturation, intensity (HSI) representation, and other authors followed their approach. Penna et al.5 tried to improve the accuracy by edge masking with Mumford–Shah functional; Mohanapriya and Sangeetha10 used gray-level co-occurrence matrix and classified the pixels by neural networks. In Ref. 6, the color features are extracted with the help of -means clustering. Block-wise methods divide each frame into blocks of pixels and calculate the color features thereof. Each block is evaluated at first separately, then additional criteria for adjacent blocks may be applied.1 The performance of the block-wise methods is controlled by the block size . For large , the methods converge to image-wise approach and become robust but lose sensitivity to small blood spots. If approaches one, the methods converge to the pixel-wise approach—they become less stable but more sensitive. A choice of is recommended as a compromise. Image-wise methods provide global features that describe the whole frame. However, this approach cannot discover small bleeding areas. Still, some authors classified the frames in this way, e.g., Lv et al.11 used a spatial pyramid of color invariant histograms. There are of course other image processing methods that were designed for WCE data analysis. For instance, Gueye et al.12 used WCE for automatic detection of cancer and precancerous colon lesions. They consider polyps, inflammation, tumor, and bleeding areas as suspect regions they wanted to detect. Since the primary features are texture and shape, the authors employed the scale invariant feature transform (SIFT)13 as the descriptors. Although they reported promising results in polyp and inflammation detection, their approach cannot be adopted for blood detection in the small intestine. In that case, the texture and the distribution of the key points detected by the SIFT method are not distinctive features for classification. The same is true for the methods that use local binary patterns (LBP) for WCE image description.14,15 They capture the texture of the intestinal wall, which is not discriminative for bleeding detection. 3.Proposed TechniqueThe aim of this paper is to develop and implement methods for blood detection that recognize blood in the frame regardless of its particular color and size. Special attention is paid to detection of small blood spots (with a diameter around 5 pixels or less), which cannot be detected by most of the earlier methods. We propose two different methods for blood detection, which can be used either individually or their results can be fused by various fusion rules. Let us call these methods A and B in the sequel. 3.1.Method AMethod A works pixel-wise and is based solely on the color. However, it does not work in standard RGB space because it has been known both from the literature1 as well as from our experiments that the RGB space does not provide enough discriminability of blood pixels (see Fig. 1). We define our new color space such that the separability of blood pixels and intestinal wall should be maximized. The study we performed on 15 patients shows that an appropriate color space can be defined, as shown below in the first step. The complete algorithm can be summarized as follows:
The main advantage of method A is its speed because it does not contain any high-level operations. As we will see in the experimental section, the method provides a good TP rate. 3.2.Method BMethod B uses a more sophisticated approach that not only is based on pixel colors but also incorporates the assumption that the blood in the frame forms a continuous region (or a few such regions). In other words, it eliminates isolated pixels or small spots with a color similar to blood but that probably do not exhibit actual bleeding. Thanks to this, method B achieves a low FP rate; however, it is at the expense of the computational time. On the other hand, it yields generally higher false-negative rate than A. The algorithm consists of four main steps:
For illustration of the individual steps, see Fig. 2. 4.Technical Parameters and ImplementationIn our study, we were working with two WCE systems—PillCam SB3 by Given Imaging17 and EndoCapsule developed by Olympus.18 The former capsule produces a video of a spatial resolution pixels with variable sample frequency, where the frames per second (fps) rate fluctuates from 0.5 to 3; the latter one yields a video in the stable rate 2 fps. Both videos are stored in proprietary compression formats based on motion JPEG. We decoded these formats and extracted individual video frames. Both companies also provide a simple blood detection software along with the capsule. This software apparently works in RGB space, trying to detect the frames containing the “above-threshold” number of pixels of “blood color” (the details about the implemented algorithms are not available). As we already mentioned, the “blood color” varies significantly among the patients, which is the main drawback of these simple algorithms. Nevertheless, both Olympus as well as Given Imaging software are very fast, have comfortable graphical interface [see Figs. 3(b) and 3(a)], and may be useful for approximative view, but they are not suitable for a rigorous analysis. Both our methods were implemented in C language and contain a user-friendly graphical user interface (GUI) to provide the doctor with maximum comfort. The input parameter can be easily changed by a GUI slider (see Fig. 4 for a GUI screenshot). Note that becomes effective in the last step of the algorithm. Hence, the steps 1 to 3 are performed automatically and only once, regardless of particular value of . When changing , only a single-number comparison is performed for each frame, and the results are displayed in real time. This is a big advantage of method A. 5.ExperimentsWe tested the performance of the methods on real WCE data. We used videos of 15 patients of the length from 12,000 to 20,000 frames each. The patients suffered from bleeding into the small intestine of various intensities and scales. Some patients bled at more than one place. All videos were first annotated manually by an experienced endoscopist, who selected 390 frames altogether of bleeding of various types and extents (see Fig. 5 for some examples). Among them, he marked 339 frames as “serious bleeding.” Then, he selected other 1500 frames with no bleeding (see Fig. 6 for some examples). Some of them are of an appearance that could be misinterpreted as bleeding because of the presence of red spots. We did not use this knowledge for training but solely as a ground truth for evaluating the performance of the methods. The results are summarized in Tables 1Table 2–3. We can see the performance of A and B when used individually along with the performance of various fusion rules. The rule AB means that the frame must be classified as a blood frame by both A and B together. means that at least one method must classify the frame as positive. The rule requires the frame to be labeled as positive by A but as negative by B. In the second column of Table 1, we see the TP rate achieved on 339 significant blood frames. The third column shows the same calculated for all 390 blood frames. The last column shows the FP rate evaluated on the set of 1500 blood-free frames. Both A and B require few user-defined parameters. The main parameter is the number of suspected pixels in method A. Table 1 is for , Table 2 for , and Table 3 for . Table 1Sensitivity t=1 pixel.
Table 2Sensitivity t=5 pixels.
Table 3Sensitivity t=10 pixels.
Based on this experiment, we can deduce the following conclusions.
We performed a comparison with the state-of-the-art method.19 The TP rate on 390 frames was 58%, whereas the FP rate on 1500 negative frames was 41%. Clearly, our method performs much better in most parameter settings. In Fig. 7 we also show some examples of the frames which are difficult to detect. Tables 1Table 2–3 show our ability to detect bleeding in an isolated frame without taking the time context into account. However, in reality, the appearance of the blood in the intestine is always depicted on more consecutive frames than one. So, we evaluated the results of the same experiments once again with a modified methodology. The detection is marked as “TP” if a ground-truth positive frame lies in its 5-s neighborhood. This corresponds to how the method is actually used—the detected frames are displayed, and the doctor checks their short-time neighborhood. When recalculating the success rates (see Tables 4Table 5–6), an excellent performance of our method is clearly demonstrated. Table 4Sensitivity t=1 pixel—at least one frame within ±5 s.
Table 5Sensitivity t=5 pixels—at least one frame within ±5 s.
Table 6Sensitivity t=10 pixels—at least one frame within ±5 s.
The last statistics illustrates saving of time needed for visual inspection. In the third column of Table 7, we can see the time needed for observation of the original video. The last three columns show the time spent by a doctor (in percents of the original time) when checking the positive frames only depending on the choice of parameter (we assume that the frames are displayed in 3 fps rate). The fifth column with the bold values corresponds to the recommended choice of parameter .The time saving, of course, depends on the extent of bleeding, and in our study varied from 1% to almost 90% in the case of a patient with extensive bleeding throughout the digestive tract. Table 7Reducing the inspection time by the proposed method.
6.ConclusionIn this paper, we proposed two automatic methods for detecting bleeding in WCE video of the small intestine. The first one uses solely the color information, whereas the second one incorporates the assumptions about the blood spot shape and size. The original idea is namely the definition of a new color space that provides good separability of blood pixels and the intestinal wall. We tested both methods individually as well as in various combinations. We evaluated the results on a large test set, manually annotated by an endoscopist. The conclusion is that method A gives good results in terms of both TP and FP rates. The fusion rule further enhances this result because method B is based on different assumptions, but is significantly slower than the individual A. We compared the proposed method with the method published in Ref. 19, which proved our method superior. Objective comparison with other methods published in the literature is unfortunately not possible because each author uses his own dataset, and different hardware and software platforms; additionally, most authors have not made their codes publicly available. We believe the presented technique significantly saves the time of endoscopists, which are required for visual expert assessment. The method is currently in use at the University Hospital at Hradec Kralove of the Charles University, Czech Republic. AcknowledgmentsThe study was supported by the IGA NT 13532-4/2012 research grant provided by the Czech Ministry of Health Care. ReferencesY.-G. Lee and G. Yoon,
“Real-time image analysis of capsule endoscopy for bleeding discrimination in embedded system platform,”
Int. J. Med. Health Biomed. Bioeng. Pharm. Eng., 5
(11), 583
–587
(2011). Google Scholar
S. C. Park et al.,
“Sensitivity of the suspected blood indicator: an experimental study,”
World J. Gastroenterol., 18 4169
–4174
(2012). http://dx.doi.org/10.3748/wjg.v18.i31.4169 Google Scholar
S. K. Shah, J. K. Lee and M. E. Celebi,
“Classification of bleeding images in wireless capsule endoscopy using HSI color domain and region segmentation,”
in Proc. of 2007 New England American Society for Engineering Education Conf.,
(2007). Google Scholar
Y. S. Jung et al.,
“Active blood detection in a high resolution capsule endoscopy using color spectrum transformation,”
in 2008 Int. Conf. on BioMedical Engineering and Informatics,
859
–862
(2008). Google Scholar
B. Penna et al.,
“A technique for blood detection in wireless capsule endoscopy images,”
in 17th European Signal Processing Conf.,
1864
–1868
(2009). Google Scholar
Y. Yuan, B. Li and M. Q. H. Meng,
“Bleeding frame and region detection in the wireless capsule endoscopy video,”
IEEE J. Biomed. Health Inform., 20 624
–630
(2016). http://dx.doi.org/10.1109/JBHI.2015.2399502 Google Scholar
C. Signorelli et al.,
“Sensitivity and specificity of the suspected blood identification system in video capsule enteroscopy,”
Endoscopy, 37 1170
–1173
(2005). http://dx.doi.org/10.1055/s-2005-870410 ENDCAM Google Scholar
P.-N. D’Halluin et al.,
“Does the suspected blood indicator improve the detection of bleeding lesions by capsule endoscopy?,”
Gastrointestinal Endoscopy, 61
(2), 243
–249
(2005). http://dx.doi.org/10.1016/S0016-5107(04)02587-8 GAENBQ 0016-5107 Google Scholar
T. Ghosh, S. A. Fattah and K. A. Wahid,
“Automatic bleeding detection in wireless capsule endoscopy based on rgb pixel intensity ratio,”
in 2014 Int. Conf. on Electrical Engineering and Information Communication Technology (ICEEICT),
1
–4
(2014). Google Scholar
P. Mohanapriya and M. Sangeetha,
“An efficient approach to detect bleeding region in gi tract using segmentation and classification techniques,”
Int. J. Adv. Inf. Commun. Technol., 1
(1), 153
–159
(2014). http://dx.doi.org/01.0401/ijaict.2014.01.30 Google Scholar
G. Lv, G. Yan and Z. Wang,
“Bleeding detection in wireless capsule endoscopy images based on color invariants and spatial pyramids using support vector machines,”
in 2011 Annual Int. Conf. of the IEEE Engineering in Medicine and Biology Society,
6643
–6646
(2011). http://dx.doi.org/10.1109/IEMBS.2011.6091638 Google Scholar
L. Gueye et al.,
“Automatic detection of colonoscopic anomalies using capsule endoscopy,”
in 2015 IEEE Int. Conf. on Image Processing (ICIP),
1061
–1064
(2015). http://dx.doi.org/10.1109/ICIP.2015.7350962 Google Scholar
D. G. Lowe,
“Object recognition from local scale-invariant features,”
in The Proc. of the Seventh IEEE Int. Conf. on Computer Vision,
1150
–1157
(1999). Google Scholar
B. Li and M. Q. H. Meng,
“Computer-aided detection of bleeding regions for capsule endoscopy images,”
IEEE Trans. Biomed. Eng., 56 1032
–1039
(2009). http://dx.doi.org/10.1109/TBME.2008.2010526 IEBEAX 0018-9294 Google Scholar
Q. Zhao and M. Q. H. Meng,
“Polyp detection in wireless capsule endoscopy images using novel color texture features,”
in 2011 9th World Congress on Intelligent Control and Automation (WCICA),
948
–952
(2011). Google Scholar
J. Canny,
“A computational approach to edge detection,”
IEEE Trans. Pattern Anal. Mach. Intell., PAMI-8 679
–698
(1986). http://dx.doi.org/10.1109/TPAMI.1986.4767851 ITPIDJ 0162-8828 Google Scholar
Given Imaging, “Rapid for pillcam software,”
(2016) http://www.givenimaging.com/ May ). 2016). Google Scholar
Olympus, “Endocapsule 10 software client,”
(2016) http://www.olympus-europa.com/medical/ May ). 2016). Google Scholar
A. A. Al-Rahayfeh and A. A. Abuzneid,
“Detection of bleeding in wireless capsule endoscopy images using range ratio color,”
(2010). Google Scholar
BiographyAdam Novozámský received his MSc degree in informatics from the Faculty of Nuclear Sciences and Physical Engineering, Czech Technical University, Prague, in 2010. He is currently pursuing his PhD with the Institute of Information Theory and Automation Cooperating Institute of Czech Technical University, Prague. His research interests include medical imaging, image segmentation, and image forensics. Jan Flusser received his MSc degree in mathematical engineering from Czech Technical University, Prague, Czech Republic, in 1985, his PhD in computer science in 1990, his DrSc. degree in 2001, and became a full professor in 2004. Since 1985, he has been with the Institute of Information Theory and Automation, Czech Academy of Sciences (director of the Institute in 2007 to 2017). His research interests include moments and moment invariants, image registration, image fusion, multichannel blind deconvolution, and super-resolution imaging. Ilja Tachecí received his MD degree in general medicine from the Charles University (Faculty of Medicine in Hradec Kralove) in 1999 and his PhD in internal medicine in 2010. He is currently deputy head for Medical Education at 2nd Department of Internal Medicine, University Hospital in Hradec Kralove. His research interest covers invasive and experimental gastrointestinal endoscopy focused on the small bowel diseases. Lukáš Sulík received his MSc degree in applied informatics from the Faculty of Informatics and Management, University of Hradec Kralove, Czech Republic, in 2016. He is currently in postgraduate study at the Center of Basic and Applied Research of the same university. His interests are in image processing. Jan Bureš graduated from Charles University in 1979. He has been affiliated with the 2nd Department of Medicine, Charles University Faculty of Medicine and University Hospital in Hradec Kralove since 1979. He became an associate professor in 1995 and a professor of internal medicine in 2002. He was elected a fellow of the Czech Medical Academy (FCMA) in 2014. Currently, he is the head of the Academic Department of Internal Medicine. His clinical and experimental research is directed at digestive endoscopy, gastrointestinal microbiome, and inflammatory bowel disease. Ondřej Krejcar received his MSc degree in control and information systems from the Technical University of Ostrava, Ostrava, Czech Republic, in 2002, his PhD in technical cybernetics in 2008, and became an associate professor in 2011 in technical cybernetics at the same university. He is currently a vice-dean for science and research at the Faculty of Informatics and Management of the University of Hradec Kralove. His research interests include biomedicine, image segmentation and recognition, video processing, biometrics, technical cybernetics, and ubiquitous computing. |