The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] scale invariant(6hit)

1-6hit
  • Estimation of Dense Displacement by Scale Invariant Polynomial Expansion of Heterogeneous Multi-View Images

    Kazuki SHIBATA  Mehrdad PANAHPOUR TEHERANI  Keita TAKAHASHI  Toshiaki FUJII  

     
    LETTER

      Pubricized:
    2017/06/14
      Vol:
    E100-D No:9
      Page(s):
    2048-2051

    Several applications for 3-D visualization require dense detection of correspondence for displacement estimation among heterogeneous multi-view images. Due to differences in resolution or sampling density and field of view in the images, estimation of dense displacement is not straight forward. Therefore, we propose a scale invariant polynomial expansion method that can estimate dense displacement between two heterogeneous views. Evaluation on heterogeneous images verifies accuracy of our approach.

  • Efficient Cloth Pattern Recognition Using Random Ferns

    Inseong HWANG  Seungwoo JEON  Beobkeun CHO  Yoonsik CHOE  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2014/10/31
      Vol:
    E98-D No:2
      Page(s):
    475-478

    This paper proposes a novel image classification scheme for cloth pattern recognition. The rotation and scale invariant delta-HOG (DHOG)-based descriptor and the entire recognition process using random ferns with this descriptor are proposed independent from pose and scale changes. These methods consider maximun orientation and various radii of a circular patch window for fast and efficient classification even when cloth patches are rotated and the scale is changed. It exhibits good performance in cloth pattern recognition experiments. It found a greater number of similar cloth patches than dense-SIFT in 20 tests out of a total of 36 query tests. In addition, the proposed method is much faster than dense-SIFT in both training and testing; its time consumption is decreased by 57.7% in training and 41.4% in testing. The proposed method, therefore, is expected to contribute to real-time cloth searching service applications that update vast numbers of cloth images posted on the Internet.

  • Combining LBP and SIFT in Sparse Coding for Categorizing Scene Images

    Shuang BAI  Jianjun HOU  Noboru OHNISHI  

     
    LETTER-Image Recognition, Computer Vision

      Vol:
    E97-D No:9
      Page(s):
    2563-2566

    Local descriptors, Local Binary Pattern (LBP) and Scale Invariant Feature Transform (SIFT) are widely used in various computer applications. They emphasize different aspects of image contents. In this letter, we propose to combine them in sparse coding for categorizing scene images. First, we regularly extract LBP and SIFT features from training images. Then, corresponding to each feature, a visual word codebook is constructed. The obtained LBP and SIFT codebooks are used to create a two-dimensional table, in which each entry corresponds to an LBP visual word and a SIFT visual word. Given an input image, LBP and SIFT features extracted from the same positions of this image are encoded together based on sparse coding. After that, spatial max pooling is adopted to determine the image representation. Obtained image representations are converted into one-dimensional features and classified by utilizing SVM classifiers. Finally, we conduct extensive experiments on datasets of Scene Categories 8 and MIT 67 Indoor Scene to evaluate the proposed method. Obtained results demonstrate that combining features in the proposed manner is effective for scene categorization.

  • SIFT-Based Non-blind Watermarking Robust to Non-linear Geometrical Distortions

    Toshihiko YAMASAKI  Kiyoharu AIZAWA  

     
    PAPER-Image Processing and Video Processing

      Vol:
    E96-D No:6
      Page(s):
    1368-1375

    This paper presents a non-blind watermarking technique that is robust to non-linear geometric distortion attacks. This is one of the most challenging problems for copyright protection of digital content because it is difficult to estimate the distortion parameters for the embedded blocks. In our proposed scheme, the location of the blocks are recorded by the translation parameters from multiple Scale Invariant Feature Transform (SIFT) feature points. This method is based on two assumptions: SIFT features are robust to non-linear geometric distortion and even such non-linear distortion can be regarded as “linear” distortion in local regions. We conducted experiments using 149,800 images (7 standard images and 100 images downloaded from Flickr, 10 different messages, 10 different embedding block patterns, and 14 attacks). The results show that the watermark detection performance is drastically improved, while the baseline method can achieve only chance level accuracy.

  • Position-Invariant Robust Features for Long-Term Recognition of Dynamic Outdoor Scenes

    Aram KAWEWONG  Sirinart TANGRUAMSUB  Osamu HASEGAWA  

     
    PAPER-Image Recognition, Computer Vision

      Vol:
    E93-D No:9
      Page(s):
    2587-2601

    A novel Position-Invariant Robust Feature, designated as PIRF, is presented to address the problem of highly dynamic scene recognition. The PIRF is obtained by identifying existing local features (i.e. SIFT) that have a wide baseline visibility within a place (one place contains more than one sequential images). These wide-baseline visible features are then represented as a single PIRF, which is computed as an average of all descriptors associated with the PIRF. Particularly, PIRFs are robust against highly dynamical changes in scene: a single PIRF can be matched correctly against many features from many dynamical images. This paper also describes an approach to using these features for scene recognition. Recognition proceeds by matching an individual PIRF to a set of features from test images, with subsequent majority voting to identify a place with the highest matched PIRF. The PIRF system is trained and tested on 2000+ outdoor omnidirectional images and on COLD datasets. Despite its simplicity, PIRF offers a markedly better rate of recognition for dynamic outdoor scenes (ca. 90%) than the use of other features. Additionally, a robot navigation system based on PIRF (PIRF-Nav) can outperform other incremental topological mapping methods in terms of time (70% less) and memory. The number of PIRFs can be reduced further to reduce the time while retaining high accuracy, which makes it suitable for long-term recognition and localization.

  • Classification of Rotated and Scaled Textured Images Using Invariants Based on Spectral Moments

    Yasuo YOSHIDA  Yue WU  

     
    PAPER

      Vol:
    E81-A No:8
      Page(s):
    1661-1666

    This paper describes a classification method for rotated and scaled textured images using invariant parameters based on spectral-moments. Although it is well known that rotation invariants can be derived from moments of grey-level images, the use is limited to binary images because of its computational unstableness. In order to overcome this drawback, we use power spectrum instead of the grey levels to compute moments and adjust the integral region of moment evaluation to the change of scale. Rotation and scale invariants are obtained as the ratios of the different rotation invariants on the basis of a spectral-moment property with respect to scale. The effectiveness of the approach is illustrated through experiments on natural textures from the Brodatz album. In addition, the stability of the invariants with respect to the change of scale is discussed theoretically and confirmed experimentally.