The search functionality is under construction.

Author Search Result

[Author] Jiabao WANG(7hit)

1-7hit
  • Combining Color Features for Real-Time Correlation Tracking

    Yulong XU  Zhuang MIAO  Jiabao WANG  Yang LI  Hang LI  Yafei ZHANG  Weiguang XU  Zhisong PAN  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2016/10/04
      Vol:
    E100-D No:1
      Page(s):
    225-228

    Correlation filter-based approaches achieve competitive results in visual tracking, but the traditional correlation tracking methods failed in mining the color information of the videos. To address this issue, we propose a novel tracker combined with color features in a correlation filter framework, which extracts not only gray but also color information as the feature maps to compute the maximum response location via multi-channel correlation filters. In particular, we modify the label function of the conventional classifier to improve positioning accuracy and employ a discriminative correlation filter to handle scale variations. Experiments are performed on 35 challenging benchmark color sequences. And the results clearly show that our method outperforms state-of-the-art tracking approaches while operating in real-time.

  • Self-Channel Attention Weighted Part for Person Re-Identification

    Lin DU  Chang TIAN  Mingyong ZENG  Jiabao WANG  Shanshan JIAO  Qing SHEN  Wei BAI  Aihong LU  

     
    LETTER-Image

      Pubricized:
    2020/09/01
      Vol:
    E104-A No:3
      Page(s):
    665-670

    Part based models have been proved to be beneficial for person re-identification (Re-ID) in recent years. Existing models usually use fixed horizontal stripes or rely on human keypoints to get each part, which is not consistent with the human visual mechanism. In this paper, we propose a Self-Channel Attention Weighted Part model (SCAWP) for Re-ID. In SCAWP, we first learn a feature map from ResNet50 and use 1x1 convolution to reduce the dimension of this feature map, which could aggregate the channel information. Then, we learn the weight map of attention within each channel and multiply it with the feature map to get each part. Finally, each part is used for a special identification task to build the whole model. To verify the performance of SCAWP, we conduct experiment on three benchmark datasets, including CUHK03-NP, Market-1501 and DukeMTMC-ReID. SCAWP achieves rank-1/mAP accuracy of 70.4%/68.3%, 94.6%/86.4% and 87.6%/76.8% on three datasets respectively.

  • Dual Network Fusion for Person Re-Identification

    Lin DU  Chang TIAN  Mingyong ZENG  Jiabao WANG  Shanshan JIAO  Qing SHEN  Guodong WU  

     
    LETTER-Image

      Vol:
    E103-A No:3
      Page(s):
    643-648

    Feature learning based on deep network has been verified as beneficial for person re-identification (Re-ID) in recent years. However, most researches use a single network as the baseline, without considering the fusion of different deep features. By analyzing the attention maps of different networks, we find that the information learned by different networks can complement each other. Therefore, a novel Dual Network Fusion (DNF) framework is proposed. DNF is designed with a trunk branch and two auxiliary branches. In the trunk branch, deep features are cascaded directly along the channel direction. One of the auxiliary branch is channel attention branch, which is used to allocate weight for different deep features. Another one is multi-loss training branch. To verify the performance of DNF, we test it on three benchmark datasets, including CUHK03NP, Market-1501 and DukeMTMC-reID. The results show that the effect of using DNF is significantly better than a single network and is comparable to most state-of-the-art methods.

  • Inequality-Constrained RPCA for Shadow Removal and Foreground Detection

    Hang LI  Yafei ZHANG  Jiabao WANG  Yulong XU  Yang LI  Zhisong PAN  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2015/03/02
      Vol:
    E98-D No:6
      Page(s):
    1256-1259

    State-of-the-art background subtraction and foreground detection methods still face a variety of challenges, including illumination changes, camouflage, dynamic backgrounds, shadows, intermittent object motion. Detection of foreground elements via the robust principal component analysis (RPCA) method and its extensions based on low-rank and sparse structures have been conducted to achieve good performance in many scenes of the datasets, such as Changedetection.net (CDnet); however, the conventional RPCA method does not handle shadows well. To address this issue, we propose an approach that considers observed video data as the sum of three parts, namely a row-rank background, sparse moving objects and moving shadows. Next, we cast inequality constraints on the basic RPCA model and use an alternating direction method of multipliers framework combined with Rockafeller multipliers to derive a closed-form solution of the shadow matrix sub-problem. Our experiments have demonstrated that our method works effectively on challenging datasets that contain shadows.

  • Deep Discriminative Supervised Hashing via Siamese Network

    Yang LI  Zhuang MIAO  Jiabao WANG  Yafei ZHANG  Hang LI  

     
    LETTER-Artificial Intelligence, Data Mining

      Pubricized:
    2017/09/12
      Vol:
    E100-D No:12
      Page(s):
    3036-3040

    The latest deep hashing methods perform hash codes learning and image feature learning simultaneously by using pairwise or triplet labels. However, generating all possible pairwise or triplet labels from the training dataset can quickly become intractable, where the majority of those samples may produce small costs, resulting in slow convergence. In this letter, we propose a novel deep discriminative supervised hashing method, called DDSH, which directly learns hash codes based on a new combined loss function. Compared to previous methods, our method can take full advantages of the annotated data in terms of pairwise similarity and image identities. Extensive experiments on standard benchmarks demonstrate that our method preserves the instance-level similarity and outperforms state-of-the-art deep hashing methods in the image retrieval application. Remarkably, our 16-bits binary representation can surpass the performance of existing 48-bits binary representation, which demonstrates that our method can effectively improve the speed and precision of large scale image retrieval systems.

  • Deep Correlation Tracking with Backtracking

    Yulong XU  Yang LI  Jiabao WANG  Zhuang MIAO  Hang LI  Yafei ZHANG  Gang TAO  

     
    LETTER-Vision

      Vol:
    E100-A No:7
      Page(s):
    1601-1605

    Feature extractor is an important component of a tracker and the convolutional neural networks (CNNs) have demonstrated excellent performance in visual tracking. However, the CNN features cannot perform well under conditions of low illumination. To address this issue, we propose a novel deep correlation tracker with backtracking, which consists of target translation, backtracking and scale estimation. We employ four correlation filters, one with a histogram of oriented gradient (HOG) descriptor and the other three with the CNN features to estimate the translation. In particular, we propose a backtracking algorithm to reconfirm the translation location. Comprehensive experiments are performed on a large-scale challenging benchmark dataset. And the results show that the proposed algorithm outperforms state-of-the-art methods in accuracy and robustness.

  • Feature Adaptive Correlation Tracking

    Yulong XU  Yang LI  Jiabao WANG  Zhuang MIAO  Hang LI  Yafei ZHANG  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2016/11/28
      Vol:
    E100-D No:3
      Page(s):
    594-597

    Feature extractor plays an important role in visual tracking, but most state-of-the-art methods employ the same feature representation in all scenes. Taking into account the diverseness, a tracker should choose different features according to the videos. In this work, we propose a novel feature adaptive correlation tracker, which decomposes the tracking task into translation and scale estimation. According to the luminance of the target, our approach automatically selects either hierarchical convolutional features or histogram of oriented gradient features in translation for varied scenarios. Furthermore, we employ a discriminative correlation filter to handle scale variations. Extensive experiments are performed on a large-scale benchmark challenging dataset. And the results show that the proposed algorithm outperforms state-of-the-art trackers in accuracy and robustness.