1 Introduction

One of the primary goals of Wireless Capsule Endoscopy (WCE) is to detect the mucosal abnormalities such as blood, ulcer, polyp, and so on in the gastrointestinal tract. Only less than 5% of total 55,000 frames of a WCE video typically have abnormalities, so it is critical to developing a technique to discriminate abnormal findings from normal ones [1] automatically. We introduce "Bag-of-Visual-Words" method which has been successfully used in particular for image classification in non-medical domains. Initially, the training image patches are represented by color and texture features, and then the bag of words model is constructed by K-means clustering algorithm. Subsequently, the document is described as the histogram of the visual words which is the feature vector of the image. Finally, an SVM classifier is trained using these feature vectors to distinguish pictures with abnormal regions from ones without them [1]. Experimental results on our current dataset show that the proposed method achieves promising performances. Wireless Capsule Endoscopy is a recently established technology to help physicians identify gastrointestinal diseases and abnormalities in the human small intestine [2].

Case endoscopy is a system used to record inward pictures of the gastrointestinal tract for use in therapeutic analysis. The container is comparable fit as a fiddle to a standard pharmaceutical case, despite the fact that somewhat bigger, and contains a minor camera and a variety of LEDs fueled by a battery. After a patient swallows the case, it goes along the gastrointestinal tract taking various pictures every second which are transmitted remotely to a variety of beneficiaries associated with a versatile chronicle gadget conveyed by the patient. The essential utilization of case endoscopy is to look at regions of the small digestive system that can’t be seen by different sorts of endoscopy, for example, colonoscopy or esophagogastroduodenoscopy (EGD).

The detection of bleeding regions in the small intestine still an open and crucial issue. Moreover, a good detection classification approach of the blood-based abnormalities in the intestine is missing by the techniques used in this research area according to the gastroenterologists. Thus, this paper offers a novel approach to automatically detect and categorize blood-based abnormalities in the intestine by starting with their definitions. The measure of therapeutic imaging information produced has encountered exponential development, making a tremendous interest in information stockpiling. Cloud services symbolize a massive opportunity. Storing, archiving, partaking and retrieving images in the cloud may allow the industry to handle data more efficiently and cost-effectively while overwhelming many of the legal, regulatory and technological disputes that data needs a pose. Fig. 1 shows the overall storing of medical data.

Fig. 1
figure 1

Access, viewing, sharing and storing of Medical images and Data’s

2 Methods overview

Two common classification methods: SVM and KNN to evaluate their powers in differentiating standard images and blending images. SVM is a supervised machine learning method based on the foundation of statistical learning. The basic idea of the SVM is to find the optimal hyperplane that separates the points of different classes. KNN is a simple and intuitive method widely used in pattern recognition and data mining. This classifier decides on comparing a new testing data with the training data. A two-stage saliency extraction method is proposed to localize the bleeding areas in WCE images. Since these two-stage saliency maps highlight the bleeding regions and separate bleeding mucosa from the uninformative parts that could obtain the bleeding area candidates successfully.

3 Color histogram

A color histogram is a representation of the distribution of colors in an image. The frame shown in Fig. 2 is converted into a grayscale image. Each pixel value present in Fig. 3 characterizes only an amount of light, i.e., it carries only intensity information. The histogram of the image is plotted in Fig. 4. For digital images, a color histogram represents the number of pixels that have colors in each of a fixed list of color ranges that span the image’s color space, the set of all possible colors. Color Histogram is shown in Fig. 5

If the set of possible color values is sufficiently small, each of those colors may be placed on a range by itself; then the histogram is merely the count of pixels that have each possible color. Most often, space is divided into an appropriate number of ranges, often arranged as a regular grid, each containing many similar color values. The color histogram may also be represented and displayed as a smooth function defined over the color space that approximates the pixel counts. Like other kinds of histograms, the color histogram is a statistic that can be viewed as an approximation of an underlying continuous allocation of colors values. It focuses only on the proportion of the number of different types of colors, regardless of the spatial location of the colors. The values of a color histogram are from statistics. They show the statistical distribution of colors and the essential tone of an image.

A large-scale comparison of the naive Bayes classifier with state-of-the-art algorithms for decision tree induction, instance-based learning, and rule induction on standard benchmark datasets, and found it to be sometimes superior to the other learning schemes, even on datasets with substantial feature dependencies. The basic independent Bayes model has been modified in various ways in attempts to improve its performance [6]. Attempts to overcome the independence assumption are mainly based on adding extra edges to include some of the dependencies between the features. In this case, the network has the limitation that each feature can be related to only one other feature. The semi-naive Bayesian classifier is another important attempt to avoid the independence assumption.

Fig. 2
figure 2

Block diagram Bayesian framework

Fig. 3
figure 3

Example diagram for color histogram

Fig. 4
figure 4

Color Histogram

Fig. 5
figure 5

Segmentation

Fig. 6
figure 6

Bleeding area Localization

The resulting classifier is then used to assign class labels to the testing instances where the values of the predictor features are known, but the value of the class label is unknown. This paper describes various supervised machine learning classification techniques [7]. Of course, a single article cannot be a complete review of all supervised machine learning classification algorithms (also known induction classification algorithms), yet we hope that the references cited will cover the major theoretical issues, guiding the researcher in interesting research directions [3] and suggesting possible bias combinations that have yet to be explored.

The resulting classifier is then used to assign class labels to the testing instances where the values of the predictor features [8] are known, but the value of the class label is unknown. This paper describes various supervised machine learning classification techniques. A single article cannot be a complete review of all supervised machine learning classification algorithms [9] (also known induction classification algorithms), yet that the references cited will cover the major theoretical issues, guiding the researcher in interesting research directions [10] and suggesting possible bias combinations that have yet to be explored.

4 Salient region identification

Detection of salient image regions is useful for applications like image segmentation, adaptive compression [13], and region-based image retrieval. In this paper, we present a novel method to determine salient regions in images using low-level features of luminance and color. Figure 6 shows the segmented output and the bleeding area localized is shown in Fig. 7. The method is fast, easy to implement and generates high-quality saliency maps of the same size and resolution as the input image. We demonstrate the use of the algorithm [15] in the segmentation of semantically meaningful whole objects from digital images. In this case, it is a Naive Bayesian classifier. Although some of the features users have correlation thus ruining the assumption of feature independence, the performance of this classifier [11] is satisfactory. Classification is done per segment. Thus the output of the classifier holds saliency estimation [14] for each segment that is further used to reconstruct the corresponding saliency map using segmentation data. Learning was done using expectation-maximization (EM) learning method [5]. Various features like centroid, area, and eccentricity, etc., are extracted and fed to the classifier.

Fig. 7
figure 7

Clustered output

5 Naïve Bayes classifier

Naïve Bayes classification is a kind of simple probabilistic classification methods based on Bayes’ theorem [17] with the assumption of independence between features [3]. The model is trained on training dataset to make predictions by predict function. This article introduces two functions Naïve Bayes and train for the performance of Naïve Bayes classification. Bayes’ theorem can be used to make a prediction based on prior knowledge and current evidence. With accumulating evidence, the prediction is changed. In technical terms, the prediction is the posterior probability [18] that investigators are interested in. The prior knowledge is termed prior probability that reflects the most probable guess on the outcome without additional evidence.

A complex image registration issue arising while the dependencies between intensities of images to be registered are not spatially homogeneous. Such a situation is frequently encountered in medical imaging when pathology present in one of the images modifies locally intensity dependencies observed on normal tissues. Fig. 8 shows the clustered output from which features are extracted and then fed to the classifier. Usual image registration models, which are based on a single global intensity similarity criterion, fail to register such images, as they are blind to local deviations of intensity dependencies. Such a limitation is also encountered in contrast-enhanced images where there exist multiple pixel classes having different properties of contrast agent absorption. In this paper, we propose a new model in which the similarity criterion is adapted locally to images by classification of image intensity [16] dependencies. Defined in a Bayesian framework, the similarity criterion is a mixture of probability distributions describing dependencies on two classes (Table 1).

Table 1 Features extracted from RGB channels

The model also includes a class map which locates pixels of the two classes and weighs the two mixture components. The registration problem is formulated both as an energy minimization problem and as a maximum a posteriori estimation problem. It is solved using a gradient descent algorithm [18]. In the problem formulation and resolution, the image deformation and the class map are estimated simultaneously, leading to an original combination of registration and classification that we call image classifying registration. Whenever sufficient information about the class location is available in applications, the registration can also be performed on its own by fixing a given class map. Finally, we illustrate the interest of our model on two real applications from medical imaging: template-based segmentation of contrast-enhanced images and lesion detection in mammograms. We also conduct an evaluation of our model on simulated medical data and show its ability to take into account spatial variations of intensity dependencies while keeping good registration accuracy [4].

6 Simulation results

Input Frame

figure c

Gray Image

figure d

Histogram

figure e

7 Conclusion

An effective method to find the bleeding frame in an Endoscopic video is obtained. Our result shows that the proposed method has obtained more accurately. The CMYK color format helps in identification of the bleeding in a clear manner. The method is applied for numerous frames of few endoscopic videos and found effective.