The search functionality is under construction.

Keyword Search Result

[Keyword] object segmentation(8hit)

1-8hit
  • Backbone Alignment and Cascade Tiny Object Detecting Techniques for Dolphin Detection and Classification

    Yih-Cherng LEE  Hung-Wei HSU  Jian-Jiun DING  Wen HOU  Lien-Shiang CHOU  Ronald Y. CHANG  

     
    PAPER-Image

      Pubricized:
    2020/09/29
      Vol:
    E104-A No:4
      Page(s):
    734-743

    Automatic tracking and classification are essential for studying the behaviors of wild animals. Owing to dynamic far-shooting photos, the occlusion problem, protective coloration, the background noise is irregular interference for designing a computerized algorithm for reducing human labeling resources. Moreover, wild dolphin images are hard-acquired by on-the-spot investigations, which takes a lot of waiting time and hardly sets the fixed camera to automatic monitoring dolphins on the ocean in several days. It is challenging tasks to detect well and classify a dolphin from polluted photos by a single famous deep learning method in a small dataset. Therefore, in this study, we propose a generic Cascade Small Object Detection (CSOD) algorithm for dolphin detection to handle small object problems and develop visualization to backbone based classification (V2BC) for removing noise, highlighting features of dolphin and classifying the name of dolphin. The architecture of CSOD consists of the P-net and the F-net. The P-net uses the crude Yolov3 detector to be a core network to predict all the regions of interest (ROIs) at lower resolution images. Then, the F-net, which is more robust, is applied to capture the ROIs from high-resolution photos to solve single detector problems. Moreover, a visualization to backbone based classification (V2BC) method focuses on extracting significant regions of occluded dolphin and design significant post-processing by referencing the backbone of dolphins to facilitate for classification. Compared to the state of the art methods, including faster-rcnn, yolov3 detection and Alexnet, the Vgg, and the Resnet classification. All experiments show that the proposed algorithm based on CSOD and V2BC has an excellent performance in dolphin detection and classification. Consequently, compared to the related works of classification, the accuracy of the proposed designation is over 14% higher. Moreover, our proposed CSOD detection system has 42% higher performance than that of the original Yolov3 architecture.

  • Recognition of Moving Object in High Dynamic Scene for Visual Prosthesis

    Fei GUO  Yuan YANG  Yang XIAO  Yong GAO  Ningmei YU  

     
    PAPER-Human-computer Interaction

      Pubricized:
    2019/04/17
      Vol:
    E102-D No:7
      Page(s):
    1321-1331

    Currently, visual perceptions generated by visual prosthesis are low resolution with unruly color and restricted grayscale. This severely restricts the ability of prosthetic implant to complete visual tasks in daily scenes. Some studies explore existing image processing techniques to improve the percepts of objects in prosthetic vision. However, most of them extract the moving objects and optimize the visual percepts in general dynamic scenes. The application of visual prosthesis in daily life scenes with high dynamic is greatly limited. Hence, in this study, a novel unsupervised moving object segmentation model is proposed to automatically extract the moving objects in high dynamic scene. In this model, foreground cues with spatiotemporal edge features and background cues with boundary-prior are exploited, the moving object proximity map are generated in dynamic scene according to the manifold ranking function. Moreover, the foreground and background cues are ranked simultaneously, and the moving objects are extracted by the two ranking maps integration. The evaluation experiment indicates that the proposed method can uniformly highlight the moving object and keep good boundaries in high dynamic scene with other methods. Based on this model, two optimization strategies are proposed to improve the perception of moving objects under simulated prosthetic vision. Experimental results demonstrate that the introduction of optimization strategies based on the moving object segmentation model can efficiently segment and enhance moving objects in high dynamic scene, and significantly improve the recognition performance of moving objects for the blind.

  • Semantic Motion Signature for Segmentation of High Speed Large Displacement Objects

    Yinhui ZHANG  Zifen HE  

     
    LETTER-Image Processing and Video Processing

      Pubricized:
    2016/10/05
      Vol:
    E100-D No:1
      Page(s):
    220-224

    This paper presents a novel method for unsupervised segmentation of objects with large displacements in high speed video sequences. Our general framework introduces a new foreground object predicting method that finds object hypotheses by encoding both spatial and temporal features via a semantic motion signature scheme. More specifically, temporal cues of object hypotheses are captured by the motion signature proposed in this paper, which is derived from sparse saliency representation imposed on magnitude of optical flow field. We integrate semantic scores derived from deep networks with location priors that allows us to directly estimate appearance potentials of foreground hypotheses. A unified MRF energy functional is proposed to simultaneously incorporate the information from the motion signature and semantic prediction features. The functional enforces both spatial and temporal consistency and impose appearance constancy and spatio-temporal smoothness constraints directly on the object hypotheses. It inherently handles the challenges of segmenting ambiguous objects with large displacements in high speed videos. Our experiments on video object segmentation benchmarks demonstrate the effectiveness of the proposed method for segmenting high speed objects despite the complicated scene dynamics and large displacements.

  • Nonlinear Regression of Saliency Guided Proposals for Unsupervised Segmentation of Dynamic Scenes

    Yinhui ZHANG  Mohamed ABDEL-MOTTALEB  Zifen HE  

     
    PAPER-Image Processing and Video Processing

      Pubricized:
    2015/11/06
      Vol:
    E99-D No:2
      Page(s):
    467-474

    This paper proposes an efficient video object segmentation approach that is tolerant to complex scene dynamics. Unlike existing approaches that rely on estimating object-like proposals on an intra-frame basis, the proposed approach employs temporally consistent foreground hypothesis using nonlinear regression of saliency guided proposals across a video sequence. For this purpose, we first generate salient foreground proposals at superpixel level by leveraging a saliency signature in the discrete cosine transform domain. We propose to use a random forest based nonlinear regression scheme to learn both appearance and shape features from salient foreground regions in all frames of a sequence. Availability of such features can help rank every foreground proposals of a sequence, and we show that the regions with high ranking scores are well correlated with semantic foreground objects in dynamic scenes. Subsequently, we utilize a Markov Random Field to integrate both appearance and motion coherence of the top-ranked object proposals. A temporal nonlinear regressor for generating salient object support regions significantly improves the segmentation performance compared to using only per-frame objectness cues. Extensive experiments on challenging real-world video sequences are performed to validate the feasibility and superiority of the proposed approach for addressing dynamic scene segmentation.

  • Foreground Segmentation via Dynamic Programming

    Bing LUO  Chao HUANG  Lei MA  Wei LI  Qingbo WU  

     
    LETTER-Image Recognition, Computer Vision

      Vol:
    E97-D No:10
      Page(s):
    2818-2822

    This paper proposes a novel method to segment the object of a specific class based on a rough detection window (such as Deformable Part Model (DPM) in this paper), which is robust to the positions of the bounding boxes. In our method, the DPM is first used to generate the root and part windows of the object. Then a set of object part candidates are generated by randomly sampling windows around the root window. Furthermore, an undirected graph (the minimum spanning tree) is constructed to describe the spatial relationships between the part windows. Finally, the object is segmented by grouping the part proposals on the undirected graph, which is formulated as an energy function minimization problem. A novel energy function consisting of the data term and the smoothness term is designed to characterize the combination of the part proposals, which is globally minimized by the dynamic programming on a tree. Our experimental results on challenging dataset demonstrate the effectiveness of the proposed method.

  • Semi-Automatic Video Object Segmentation Using LVQ with Color and Spatial Features

    Hariadi MOCHAMAD  Hui Chien LOY  Takafumi AOKI  

     
    PAPER-Image Processing and Multimedia Systems

      Vol:
    E88-D No:7
      Page(s):
    1553-1560

    This paper presents a semi-automatic algorithm for video object segmentation. Our algorithm assumes the use of multiple key video frames in which a semantic object of interest is defined in advance with human assistance. For video frames between every two key frames, the specified video object is tracked and segmented automatically using Learning Vector Quantization (LVQ). Each pixel of a video frame is represented by a 5-dimensional feature vector integrating spatial and color information. We introduce a parameter K to adjust the balance of spatial and color information. Experimental results demonstrate that the algorithm can segment the video object consistently with less than 2% average error when the object is moving at a moderate speed.

  • Scale-Space Processing of Point-Sampled Geometry for Efficient 3D Object Segmentation

    Hamid LAGA  Hiroki TAKAHASHI  Masayuki NAKAJIMA  

     
    PAPER

      Vol:
    E88-D No:5
      Page(s):
    963-970

    In this paper, we present a novel framework for analyzing and segmenting point-sampled 3D objects. Our algorithm computes a decomposition of a given point set surface into meaningful components, which are delimited by line features and deep concavities. Central to our method is the extension of the scale-space theory to the three-dimensional space to allow feature analysis and classification at different scales. Then, a new surface classifier is computed and used in an anisotropic diffusion process via partial differential equations (PDEs). The algorithm avoids the misclassifications due to fuzzy and incomplete line features. Our algorithm operates directly on points requiring no vertex connectivity information. We demonstrate and discuss its performance on a collection of point sampled 3D objects including CAD and natural models. Applications include 3D shape matching and retrieval, surface reconstruction and feature preserving simplification.

  • Context-Free Marker-Controlled Watershed Transform for Efficient Multi-Object Detection and Segmentation

    Kyung-Seok SEO  Chang-Joon PARK  Sang-Hyun CHO  Heung-Moon CHOI  

     
    PAPER

      Vol:
    E84-A No:6
      Page(s):
    1392-1400

    A high-speed context-free marker controlled and minima imposition-free watershed transform is proposed for efficient multi-object detection and segmentation from a complex background. The context-free markers are extracted from a complex backgrounded multi-object image using a noise tolerant attention operator. These make high speed marker-controlled watershed possible without over-segmentation and region merging. The proposed method presents a marker-constrained labeling that can speed up the segmentation of the marker-controlled watershed transform by eliminating the necessity of the minima imposition. Simulation results show that the proposed method can efficiently detect and segment multiple objects from a complex background while reducing the over-segmentation and computation time.