The search functionality is under construction.

Author Search Result

[Author] Rui YANG(2hit)

1-2hit
  • Improving Noised Gradient Penalty with Synchronized Activation Function for Generative Adversarial Networks

    Rui YANG  Raphael SHU  Hideki NAKAYAMA  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2022/05/27
      Vol:
    E105-D No:9
      Page(s):
    1537-1545

    Generative Adversarial Networks (GANs) are one of the most successful learning principles of generative models and were wildly applied to many generation tasks. In the beginning, the gradient penalty (GP) was applied to enforce the discriminator in GANs to satisfy Lipschitz continuity in Wasserstein GAN. Although the vanilla version of the gradient penalty was further modified for different purposes, seeking a better equilibrium and higher generation quality in adversarial learning remains challenging. Recently, DRAGAN was proposed to achieve the local linearity in a surrounding data manifold by applying the noised gradient penalty to promote the local convexity in model optimization. However, we show that their approach will impose a burden on satisfying Lipschitz continuity for the discriminator. Such conflict between Lipschitz continuity and local linearity in DRAGAN will result in poor equilibrium, and thus the generation quality is far from ideal. To this end, we propose a novel approach to benefit both local linearity and Lipschitz continuity for reaching a better equilibrium without conflict. In detail, we apply our synchronized activation function in the discriminator to receive a particular form of noised gradient penalty for achieving local linearity without losing the property of Lipschitz continuity in the discriminator. Experimental results show that our method can reach the superior quality of images and outperforms WGAN-GP, DiracGAN, and DRAGAN in terms of Inception Score and Fréchet Inception Distance on real-world datasets.

  • Triple Loss Based Framework for Generalized Zero-Shot Learning

    Yaying SHEN  Qun LI  Ding XU  Ziyi ZHANG  Rui YANG  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2021/12/27
      Vol:
    E105-D No:4
      Page(s):
    832-835

    A triple loss based framework for generalized zero-shot learning is presented in this letter. The approach learns a shared latent space for image features and attributes by using aligned variational autoencoders and variants of triplet loss. Then we train a classifier in the latent space. The experimental results demonstrate that the proposed framework achieves great improvement.