The search functionality is under construction.

Keyword Search Result

[Keyword] neural architecture search(2hit)

1-2hit
  • PSTNet: Crowd Flow Prediction by Pyramidal Spatio-Temporal Network

    Enze YANG  Shuoyan LIU  Yuxin LIU  Kai FANG  

     
    LETTER-Biocybernetics, Neurocomputing

      Pubricized:
    2021/04/12
      Vol:
    E104-D No:10
      Page(s):
    1780-1783

    Crowd flow prediction in high density urban scenes is involved in a wide range of intelligent transportation and smart city applications, and it has become a significant topic in urban computing. In this letter, a CNN-based framework called Pyramidal Spatio-Temporal Network (PSTNet) for crowd flow prediction is proposed. Spatial encoding is employed for spatial representation of external factors, while prior pyramid enhances feature dependence of spatial scale distances and temporal spans, after that, post pyramid is proposed to fuse the heterogeneous spatio-temporal features of multiple scales. Experimental results based on TaxiBJ and MobileBJ demonstrate that proposed PSTNet outperforms the state-of-the-art methods.

  • Neural Architecture Search for Convolutional Neural Networks with Attention

    Kohei NAKAI  Takashi MATSUBARA  Kuniaki UEHARA  

     
    PAPER-Image Recognition, Computer Vision

      Pubricized:
    2020/10/26
      Vol:
    E104-D No:2
      Page(s):
    312-321

    The recent development of neural architecture search (NAS) has enabled us to automatically discover architectures of neural networks with high performance within a few days. Convolutional neural networks extract fruitful features by repeatedly applying standard operations (convolutions and poolings). However, these operations also extract useless or even disturbing features. Attention mechanisms enable neural networks to discard information of no interest, having achieved the state-of-the-art performance. While a variety of attentions for CNNs have been proposed, current NAS methods have paid a little attention to them. In this study, we propose a novel NAS method that searches attentions as well as operations. We examined several patterns to arrange attentions and operations, and found that attentions work better when they have their own search space and follow operations. We demonstrate the superior performance of our method in experiments on CIFAR-10, CIFAR-100, and ImageNet datasets. The found architecture achieved lower classification error rates and required fewer parameters compared to those found by current NAS methods.