The search functionality is under construction.
The search functionality is under construction.

Author Search Result

[Author] Jinyoung KIM(3hit)

1-3hit
  • A Robust Recursive Least Square Algorithm against Impulsive Noise

    Seong-Joon BAEK  Jinyoung KIM  Dae-Jin KIM  Dong-Soo HAR  Kiseon KIM  

     
    LETTER-Digital Signal Processing

      Vol:
    E87-A No:9
      Page(s):
    2463-2465

    In this paper, we propose a robust adaptive algorithm for impulsive noise suppression. The perturbation of the input signal as well as the perturbation of the estimation error are restricted by M-estimation. The threshold used in M-estimation is obtained from the proposed adaptive variance estimation. Simulations show that the proposed algorithm is less vulnerable to the impulsive noise than the conventional algorithm.

  • An Efficient Lip-Reading Method Robust to Illumination Variations

    Jinyoung KIM  Joohun LEE  Katsuhiko SHIRAI  

     
    LETTER-Speech and Hearing

      Vol:
    E85-A No:9
      Page(s):
    2164-2168

    In this paper, for real-time automatic image transform based lip-reading under illumination variations, an efficient (smaller feature data size) and robust (better recognition under different lighting conditions) method is proposed. Image transform based approach obtains a compressed representation of image pixel values of speaker's mouth and is reported to show superior lip-reading performance. However, this approach inevitably produces large feature vectors relevant to lip information to require much computation time for lip-reading even when principal component analysis (PCA) is applied. To reduce the necessary dimension of feature vectors, the proposed method folded the lip image based on its symmetry in a frame image. This method also compensates the unbalanced illumination between the left and the right lip areas. Additionally, to filter out the inter-frame time-domain spectral distortion of each pixel contaminated by illumination noise, our method adapted the hi-pass filtering on the variations of pixel values between consecutive frames. In the experimental results performed on database recorded at various lighting conditions, the proposed lip-folding or/and inter-frame filtering reduced much the necessary number of feature data, principal components in this work, and showed superior recognition rate compared to the conventional method.

  • Development of a Lip-Sync Algorithm Based on an Audio-Visual Corpus

    Jinyoung KIM  Joohun LEE  Katsuhiko SHIRAI  

     
    LETTER-Databases

      Vol:
    E86-D No:2
      Page(s):
    334-339

    In this paper, we propose a corpus-based lip-sync algorithm for natural face animation. For this purpose, we constructed a Korean audio-visual (AV) corpus. Based on this AV corpus, we propose a concatenation method of AV units, which is similar to a corpus-based text-to-speech system. For our AV corpus, lip-related parameters were extracted from every video-recorded facial shot which of speaker reads the given texts selected from newspapers. The spoken utterances were labeled with HTK and such prosodic information as duration, pitch and intensity was extracted as lip-sync parameters. Based on the constructed AV corpus, basic synthetic units are set by CVC-syllable units. For the best concatenation performance, based on the phonetic environment distance and the prosodic distance, the best path is estimated by a general Viterbi search algorithm. From the computer simulation results, we found that the information concerned with not only duration but also pitch and intensity is useful to enhance the lip-sync performance. And the reconstructed lip parameters have almost equal values to those of the original parameters.