The search functionality is under construction.

Author Search Result

[Author] Shuang LIU(6hit)

1-6hit
  • Bilateral Convolutional Activations Encoded with Fisher Vectors for Scene Character Recognition

    Zhong ZHANG  Hong WANG  Shuang LIU  Tariq S. DURRANI  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2018/02/02
      Vol:
    E101-D No:5
      Page(s):
    1453-1456

    A rich and robust representation for scene characters plays a significant role in automatically understanding the text in images. In this letter, we focus on the issue of feature representation, and propose a novel encoding method named bilateral convolutional activations encoded with Fisher vectors (BCA-FV) for scene character recognition. Concretely, we first extract convolutional activation descriptors from convolutional maps and then build a bilateral convolutional activation map (BCAM) to capture the relationship between the convolutional activation response and the spatial structure information. Finally, in order to obtain the global feature representation, the BCAM is injected into FV to encode convolutional activation descriptors. Hence, the BCA-FV can effectively integrate the prominent features and spatial structure information for character representation. We verify our method on two widely used databases (ICDAR2003 and Chars74K), and the experimental results demonstrate that our method achieves better results than the state-of-the-art methods. In addition, we further validate the proposed BCA-FV on the “Pan+ChiPhoto” database for Chinese scene character recognition, and the experimental results show the good generalization ability of the proposed BCA-FV.

  • Consistent Sparse Representation for Abnormal Event Detection

    Zhong ZHANG  Shuang LIU  Zhiwei ZHANG  

     
    LETTER-Pattern Recognition

      Pubricized:
    2015/07/17
      Vol:
    E98-D No:10
      Page(s):
    1866-1870

    Sparsity-based methods have been recently applied to abnormal event detection and have achieved impressive results. However, most such methods suffer from the problem of dimensionality curse; furthermore, they also take no consideration of the relationship among coefficient vectors. In this paper, we propose a novel method called consistent sparse representation (CSR) to overcome the drawbacks. We first reconstruct each feature in the space spanned by the clustering centers of training features so as to reduce the dimensionality of features and preserve the neighboring structure. Then, the consistent regularization is added to the sparse representation model, which explicitly considers the relationship of coefficient vectors. Our method is verified on two challenging databases (UCSD Ped1 database and Subway batabase), and the experimental results demonstrate that our method obtains better results than previous methods in abnormal event detection.

  • Scene Character Recognition Using Coupled Spatial Learning

    Zhong ZHANG  Hong WANG  Shuang LIU  Liang ZHENG  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2017/04/17
      Vol:
    E100-D No:7
      Page(s):
    1546-1549

    Feature representation, as a key component of scene character recognition, has been widely studied and a number of effective methods have been proposed. In this letter, we propose the novel method named coupled spatial learning (CSL) for scene character representation. Different from the existing methods, the proposed CSL method simultaneously discover the spatial context in both the dictionary learning and coding stages. Concretely, we propose to build the spatial dictionary by preserving the corresponding positions of the codewords. Correspondingly, we introduce the spatial coding strategy which utilizes the spatiality regularization to consider the relationship among features in the Euclidean space. Based on the spatial dictionary and spatial coding, the spatial context can be effectively integrated in the visual representations. We verify our method on two widely used databases (ICDAR2003 and Chars74k), and the experimental results demonstrate that our method achieves competitive results compared with the state-of-the-art methods. In addition, we further validate the proposed CSL method on the Caltech-101 database for image classification task, and the experimental results show the good generalization ability of the proposed CSL.

  • Learning Discriminative Features for Ground-Based Cloud Classification via Mutual Information Maximization

    Shuang LIU  Zhong ZHANG  Baihua XIAO  Xiaozhong CAO  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2015/03/24
      Vol:
    E98-D No:7
      Page(s):
    1422-1425

    Texture feature descriptors such as local binary patterns (LBP) have proven effective for ground-based cloud classification. Traditionally, these texture feature descriptors are predefined in a handcrafted way. In this paper, we propose a novel method which automatically learns discriminative features from labeled samples for ground-based cloud classification. Our key idea is to learn these features through mutual information maximization which learns a transformation matrix for local difference vectors of LBP. The experimental results show that our learned features greatly improves the performance of ground-based cloud classification when compared to the other state-of-the-art methods.

  • Compact Sparse Coding for Ground-Based Cloud Classification

    Shuang LIU  Zhong ZHANG  Xiaozhong CAO  

     
    LETTER-Pattern Recognition

      Pubricized:
    2015/08/17
      Vol:
    E98-D No:11
      Page(s):
    2003-2007

    Although sparse coding has emerged as an extremely powerful tool for texture and image classification, it neglects the relationship of coding coefficients from the same class in the training stage, which may cause a decline in the classification performance. In this paper, we propose a novel coding strategy named compact sparse coding for ground-based cloud classification. We add a constraint on coding coefficients into the objective function of traditional sparse coding. In this way, coding coefficients from the same class can be forced to their mean vector, making them more compact and discriminative. Experiments demonstrate that our method achieves better performance than the state-of-the-art methods.

  • Contextual Max Pooling for Human Action Recognition

    Zhong ZHANG  Shuang LIU  Xing MEI  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2015/01/19
      Vol:
    E98-D No:4
      Page(s):
    989-993

    The bag-of-words model (BOW) has been extensively adopted by recent human action recognition methods. The pooling operation, which aggregates local descriptor encodings into a single representation, is a key determiner of the performance of the BOW-based methods. However, the spatio-temporal relationship among interest points has rarely been considered in the pooling step, which results in the imprecise representation of human actions. In this paper, we propose a novel pooling strategy named contextual max pooling (CMP) to overcome this limitation. We add a constraint term into the objective function under the framework of max pooling, which forces the weights of interest points to be consistent with their probabilities. In this way, CMP explicitly considers the spatio-temporal contextual relationships among interest points and inherits the positive properties of max pooling. Our method is verified on three challenging datasets (KTH, UCF Sports and UCF Films datasets), and the results demonstrate that our method achieves better results than the state-of-the-art methods in human action recognition.