The search functionality is under construction.
The search functionality is under construction.

Author Search Result

[Author] ShengYing YANG(3hit)

1-3hit
  • An Improved Online Multiclass Classification Algorithm Based on Confidence-Weighted

    Ji HU  Chenggang YAN  Jiyong ZHANG  Dongliang PENG  Chengwei REN  Shengying YANG  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2021/03/15
      Vol:
    E104-D No:6
      Page(s):
    840-849

    Online learning is a method which updates the model gradually and can modify and strengthen the previous model, so that the updated model can adapt to the new data without having to relearn all the data. However, the accuracy of the current online multiclass learning algorithm still has room for improvement, and the ability to produce sparse models is often not strong. In this paper, we propose a new Multiclass Truncated Gradient Confidence-Weighted online learning algorithm (MTGCW), which combine the Truncated Gradient algorithm and the Confidence-weighted algorithm to achieve higher learning performance. The experimental results demonstrate that the accuracy of MTGCW algorithm is always better than the original CW algorithm and other baseline methods. Based on these results, we applied our algorithm for phishing website recognition and image classification, and unexpectedly obtained encouraging experimental results. Thus, we have reasons to believe that our classification algorithm is clever at handling unstructured data which can promote the cognitive ability of computers to a certain extent.

  • SDChannelNets: Extremely Small and Efficient Convolutional Neural Networks

    JianNan ZHANG  JiJun ZHOU  JianFeng WU  ShengYing YANG  

     
    LETTER-Biocybernetics, Neurocomputing

      Pubricized:
    2019/09/10
      Vol:
    E102-D No:12
      Page(s):
    2646-2650

    Convolutional neural networks (CNNS) have a strong ability to understand and judge images. However, the enormous parameters and computation of CNNS have limited its application in resource-limited devices. In this letter, we used the idea of parameter sharing and dense connection to compress the parameters in the convolution kernel channel direction, thus greatly reducing the number of model parameters. On this basis, we designed Shared and Dense Channel-wise Convolutional Networks (SDChannelNets), mainly composed of Depth-wise Separable SD-Channel-wise Convolution layer. The advantage of SDChannelNets is that the number of model parameters is greatly reduced without or with little loss of accuracy. We also introduced a hyperparameter that can effectively balance the number of parameters and the accuracy of a model. We evaluated the model proposed by us through two popular image recognition tasks (CIFAR-10 and CIFAR-100). The results showed that SDChannelNets had similar accuracy to other CNNs, but the number of parameters was greatly reduced.

  • Vector Quantization of High-Dimensional Speech Spectra Using Deep Neural Network

    JianFeng WU  HuiBin QIN  YongZhu HUA  LiHuan SHAO  Ji HU  ShengYing YANG  

     
    LETTER-Artificial Intelligence, Data Mining

      Pubricized:
    2019/07/02
      Vol:
    E102-D No:10
      Page(s):
    2047-2050

    This paper proposes a deep neural network (DNN) based framework to address the problem of vector quantization (VQ) for high-dimensional data. The main challenge of applying DNN to VQ is how to reduce the binary coding error of the auto-encoder when the distribution of the coding units is far from binary. To address this problem, three fine-tuning methods have been adopted: 1) adding Gaussian noise to the input of the coding layer, 2) forcing the output of the coding layer to be binary, 3) adding a non-binary penalty term to the loss function. These fine-tuning methods have been extensively evaluated on quantizing speech magnitude spectra. The results demonstrated that each of the methods is useful for improving the coding performance. When implemented for quantizing 968-dimensional speech spectra using only 18-bit, the DNN-based VQ framework achieved an averaged PESQ of about 2.09, which is far beyond the capability of conventional VQ methods.