The search functionality is under construction.

Keyword Search Result

[Keyword] contextual information(3hit)

1-3hit
  • Deep Learning-Based Fault Localization with Contextual Information

    Zhuo ZHANG  Yan LEI  Qingping TAN  Xiaoguang MAO  Ping ZENG  Xi CHANG  

     
    LETTER-Software Engineering

      Pubricized:
    2017/09/08
      Vol:
    E100-D No:12
      Page(s):
    3027-3031

    Fault localization is essential for solving the issue of software faults. Aiming at improving fault localization, this paper proposes a deep learning-based fault localization with contextual information. Specifically, our approach uses deep neural network to construct a suspiciousness evaluation model to evaluate the suspiciousness of a statement being faulty, and then leverages dynamic backward slicing to extract contextual information. The empirical results show that our approach significantly outperforms the state-of-the-art technique Dstar.

  • Enriching Contextual Information for Fault Localization

    Zhuo ZHANG  Xiaoguang MAO  Yan LEI  Peng ZHANG  

     
    LETTER-Software Engineering

      Vol:
    E97-D No:6
      Page(s):
    1652-1655

    Existing fault localization approaches usually do not provide a context for developers to understand the problem. Thus, this paper proposes a novel approach using the dynamic backward slicing technique to enrich contexts for existing approaches. Our empirical results show that our approach significantly outperforms five state-of-the-art fault localization techniques.

  • Incorporating Contextual Information into Bag-of-Visual-Words Framework for Effective Object Categorization

    Shuang BAI  Tetsuya MATSUMOTO  Yoshinori TAKEUCHI  Hiroaki KUDO  Noboru OHNISHI  

     
    PAPER-Image Recognition, Computer Vision

      Vol:
    E95-D No:12
      Page(s):
    3060-3068

    Bag of visual words is a promising approach to object categorization. However, in this framework, ambiguity exists in patch encoding by visual words, due to information loss caused by vector quantization. In this paper, we propose to incorporate patch-level contextual information into bag of visual words for reducing the ambiguity mentioned above. To achieve this goal, we construct a hierarchical codebook in which visual words in the upper hierarchy contain contextual information of visual words in the lower hierarchy. In the proposed method, from each sample point we extract patches of different scales, all of which are described by the SIFT descriptor. Then, we build the hierarchical codebook in which visual words created from coarse scale patches are put in the upper hierarchy, while visual words created from fine scale patches are put in the lower hierarchy. At the same time, by employing the corresponding relationship among these extracted patches, visual words in different hierarchies are associated with each other. After that, we design a method to assign patch pairs, whose patches are extracted from the same sample point, to the constructed codebook. Furthermore, to utilize image information effectively, we implement the proposed method based on two sets of features which are extracted through different sampling strategies and fuse them using a probabilistic approach. Finally, we evaluate the proposed method on dataset Caltech 101 and dataset Caltech 256. Experimental results demonstrate the effectiveness of the proposed method.