1-1hit |
Yang YANG Zejian YUAN Nanning ZHENG Yuehu LIU Lei YANG Yoshifumi NISHIO
This paper introduces an interactive expression editing system that allows users to design facial expressions easily. Currently, popular example-based methods construct face models based on the examples of target face. The shortcoming of these methods is that they cannot create expressions for novel faces: target faces not previously recorded in the database. We propose a solution to overcome this limitation. We present an interactive facial-geometric-feature animation system for generating expressions of novel faces. Our system is easy to use. By click-dragging control points on the target face, on the computer screen display, unique expressions are generated automatically. To guarantee natural animation results, our animation model employs prior knowledge based on various individuals' expressions. One model prior is learned from motion vector fields to guarantee effective facial motions. Another, different, model prior is learned from facial shape space to ensure the result has a real facial shape. Interactive animation problem is formulated in a maximum a posterior (MAP) framework to search for optimal results by combining the priors with user-defined constraints. We give an extension of the Motion Propagation (MP) algorithm to infer facial motions for novel target faces from a subset of the control points. Experimental results on different facial animations demonstrate the effectiveness of the proposed method. Moreover, one application of our system is exhibited in this paper, where users create expressions for facial sketches interactively.