Breast cancer has become a symbol of tremendous concern in the modern world, as it is one of the major causes of cancer mortality worldwide. In this regard, breast ultrasonography images are frequently utilized by doctors to diagnose breast cancer at an early stage. However, the complex artifacts and heavily noised breast ultrasonography images make diagnosis a great challenge. Furthermore, the ever-increasing number of patients being screened for breast cancer necessitates the use of automated end-to-end technology for highly accurate diagnosis at a low cost and in a short time. In this concern, to develop an end-to-end integrated pipeline for breast ultrasonography image classification, we conducted an exhaustive analysis of image preprocessing methods such as K Means++ and SLIC, as well as four transfer learning models such as VGG16, VGG19, DenseNet121, and ResNet50. With a Dice-coefficient score of 63.4 in the segmentation stage and accuracy and an F1-Score (Benign) of 73.72 percent and 78.92 percent in the classification stage, the combination of SLIC, UNET, and VGG16 outperformed all other integrated combinations. Finally, we have proposed an end to end integrated automated pipelining framework which includes preprocessing with SLIC to capture super-pixel features from the complex artifact of ultrasonography images, complementing semantic segmentation with modified U-Net, leading to breast tumor classification using a transfer learning approach with a pre-trained VGG16 and a densely connected neural network. The proposed automated pipeline can be effectively implemented to assist medical practitioners in making more accurate and timely diagnoses of breast cancer.
Artificial Intelligence in Medical Image Segmentation and Classification
Road sign recognition is considered to be one of the most fascinating and interesting field of research in intelligent vehicle and machine learning. Road signs are typically placed either by the roadside or above roads. They provide important information in order to make driving safer and easier. This paper proposes an algorithm that recognizes Bangla road sign with a better percentage. The algorithm starts with capture image from real video scene, text detection from images, character segmentation and recognition of characters through shape matrix. The constructed feature vectors for each individual Bangla road sign are learned into a neural network which later classifies new instance of Bangla road sign. The promising preliminary experimental results indicate a positive potential of our algorithm.
In computer vision, semantically accurate segmentation of an object is considered to be a critical problem. The different looking fragments of the same object impose the main challenge of producing a good segmentation. This leads to consider the high-level semantics of an image as well as the low-level visual features which require computationally intensive operations. This demands to optimize the computations as much as possible in order to reduce both computational and communication complexity. This paper proposes a framework which can be used to perform segmentation for a particular object by incorporating optimization in subsequent steps. The algorithm proposes an optimized K-means algorithm for image segmentation followed by balance calculations in multiple instance learning and topological relations with relative positions to identify OOI regions. Finally, a bayesian network is incorporated to contain the learned information about the model of the OOI. The preliminary experimental results suggest a significant drop in the complexity.
Semantically accurate segmentation of an object of interest (OOI) is a critical step in computer vision tasks. In order to bridge the gap between low-level visual features and high-level semantics, a more complete model of the OOI is needed. To this end, we revise the concept of directional spatial templates and introduce regional directional spatial templates as a means of including spatial relationships among OOI regions into the model. We present an object segmentation algorithm that learns a model which includes both visual and spatial information. Given a training set of images containing the OOI, each image is oversegmented into visually homogeneous regions. Next, Multiple Instance Learning identifies regions that are likely to be part of the OOI. For each pair of such regions and for each relationship, a regional template is formed. The computational cost of template generation is reduced by sampling the reference region with a pixel set that is descriptive of its shape. Experiments indicate that regional templates are an effective way of including spatial information into the model which in turn results in a very significant improvement in segmentation performance.
We present a Bayesian network-based tunable image segmentation algorithm that can be used to segment a particular object of interest (OOI). In tasks such as object recognition, semantically accurate segmentation of the OOI is a critical step. Due to the OOI consisting of different-looking fragments, traditional image segmentation algorithms that are based on the identification of homogeneous regions tend to oversegment. The algorithm presented in this paper uses Multiple Instance Learning to learn prototypical representations of each fragment of the OOI and a Bayesian network to learn the spatial relationships that exist among those fragments. The Bayesian network, as a probabilistic graphical model, in turn becomes evidence that is used for the process of semantically accurate segmentation of future instances of the OOI. The key contribution of this paper is the inclusion of domain-specific information in terms of spatial relationships as an input to a conventional Bayesian network structure learning algorithm. Preliminary results indicate that the proposed method improves segmentation performance.