1-3hit |
Our research group has been working on attractiveness prediction, reasoning, and even enhancement for multimedia content, which we call “attractiveness computing.” Attractiveness includes impressiveness, instagrammability, memorability, clickability, and so on. Analyzing such attractiveness was usually done by experienced professionals but we have experimentally revealed that artificial intelligence (AI) based on big multimedia data can imitate or reproduce professionals' skills in some cases. In this paper, we introduce some of the representative works and possible real-life applications of our attractiveness computing for image media.
The last decade has witnessed an explosion of interest in research on human emotion modeling for generating intelligent virtual agents. This paper proposes a novel personality model based on the Revised NEO Personality Inventory (NEO PI-R). Compared to the popular Big-Five-Personality Factors (Big5) model, our proposed model is more capable than Big5 on describing a variety of personalities. Combining with emotion models it helps to produce more reasonable emotional reactions to external stimuli. A novel Resistant formulation is also proposed to effectively simulate the complicated negative emotions. Emotional reactions towards multiple stimuli are also effectively simulated with the proposed personality model.
The manner of a person's eye movement conveys much about nonverbal information and emotional intent beyond speech. This paper describes work on expressing emotion through eye behaviors in virtual agents based on the parameters selected from the AU-Coded facial expression database and real-time eye movement data (pupil size, blink rate and saccade). A rule-based approach to generate primary (joyful, sad, angry, afraid, disgusted and surprise) and intermediate emotions (emotions that can be represented as the mixture of two primary emotions) utilized the MPEG4 FAPs (facial animation parameters) is introduced. Meanwhile, based on our research, a scripting tool, named EEMML (Emotional Eye Movement Markup Language) that enables authors to describe and generate emotional eye movement of virtual agents, is proposed.