In computer vision, semantically accurate segmentation of an object is considered to be a critical problem. The different looking fragments of the same object impose the main challenge of producing a good segmentation. This leads to consider the high-level semantics of an image as well as the low-level visual features which require computationally intensive operations. This demands to optimize the computations as much as possible in order to reduce both computational and communication complexity. This paper proposes a framework which can be used to perform segmentation for a particular object by incorporating optimization in subsequent steps. The algorithm proposes an optimized K-means algorithm for image segmentation followed by balance calculations in multiple instance learning and topological relations with relative positions to identify OOI regions. Finally, a bayesian network is incorporated to contain the learned information about the model of the OOI. The preliminary experimental results suggest a significant drop in the complexity.
Semantically accurate segmentation of an object of interest (OOI) is a critical step in computer vision tasks. In order to bridge the gap between low-level visual features and high-level semantics, a more complete model of the OOI is needed. To this end, we revise the concept of directional spatial templates and introduce regional directional spatial templates as a means of including spatial relationships among OOI regions into the model. We present an object segmentation algorithm that learns a model which includes both visual and spatial information. Given a training set of images containing the OOI, each image is oversegmented into visually homogeneous regions. Next, Multiple Instance Learning identifies regions that are likely to be part of the OOI. For each pair of such regions and for each relationship, a regional template is formed. The computational cost of template generation is reduced by sampling the reference region with a pixel set that is descriptive of its shape. Experiments indicate that regional templates are an effective way of including spatial information into the model which in turn results in a very significant improvement in segmentation performance.
Semantically accurate segmentation of a particular Object Of Interest (OOI) in an image is an important but challenging step in computer vision tasks. Our recently proposed object-specific segmentation algorithm learns a model of the OOI which includes information on both the visual appearance of and the spatial relationships among the OOI components. However, its performance heavily depends on the assumption that the visual appearance variability among OOI instances is low. We present an extension to our algorithm that relaxes this assumption by incorporating shape information into the OOI model. Experimental results and an ANOVA-based statistical test confirm that the incorporation of shape has a highly significant positive effect on segmentation performance.
We present a Bayesian network-based tunable image segmentation algorithm that can be used to segment a particular object of interest (OOI). In tasks such as object recognition, semantically accurate segmentation of the OOI is a critical step. Due to the OOI consisting of different-looking fragments, traditional image segmentation algorithms that are based on the identification of homogeneous regions tend to oversegment. The algorithm presented in this paper uses Multiple Instance Learning to learn prototypical representations of each fragment of the OOI and a Bayesian network to learn the spatial relationships that exist among those fragments. The Bayesian network, as a probabilistic graphical model, in turn becomes evidence that is used for the process of semantically accurate segmentation of future instances of the OOI. The key contribution of this paper is the inclusion of domain-specific information in terms of spatial relationships as an input to a conventional Bayesian network structure learning algorithm. Preliminary results indicate that the proposed method improves segmentation performance.