The search functionality is under construction.

Keyword Search Result

[Keyword] affective computing(3hit)

1-3hit
  • Attractiveness Computing in Image Media

    Toshihiko YAMASAKI  

     
    INVITED PAPER-Vision

      Pubricized:
    2023/06/16
      Vol:
    E106-A No:9
      Page(s):
    1196-1201

    Our research group has been working on attractiveness prediction, reasoning, and even enhancement for multimedia content, which we call “attractiveness computing.” Attractiveness includes impressiveness, instagrammability, memorability, clickability, and so on. Analyzing such attractiveness was usually done by experienced professionals but we have experimentally revealed that artificial intelligence (AI) based on big multimedia data can imitate or reproduce professionals' skills in some cases. In this paper, we introduce some of the representative works and possible real-life applications of our attractiveness computing for image media.

  • A Personality Model Based on NEO PI-R for Emotion Simulation

    Yi ZHANG  Ling LI  

     
    PAPER-Affective Computing

      Vol:
    E97-D No:8
      Page(s):
    2000-2007

    The last decade has witnessed an explosion of interest in research on human emotion modeling for generating intelligent virtual agents. This paper proposes a novel personality model based on the Revised NEO Personality Inventory (NEO PI-R). Compared to the popular Big-Five-Personality Factors (Big5) model, our proposed model is more capable than Big5 on describing a variety of personalities. Combining with emotion models it helps to produce more reasonable emotional reactions to external stimuli. A novel Resistant formulation is also proposed to effectively simulate the complicated negative emotions. Emotional reactions towards multiple stimuli are also effectively simulated with the proposed personality model.

  • Generating and Describing Affective Eye Behaviors

    Xia MAO  Zheng LI  

     
    PAPER-Kansei Information Processing, Affective Information Processing

      Vol:
    E93-D No:5
      Page(s):
    1282-1290

    The manner of a person's eye movement conveys much about nonverbal information and emotional intent beyond speech. This paper describes work on expressing emotion through eye behaviors in virtual agents based on the parameters selected from the AU-Coded facial expression database and real-time eye movement data (pupil size, blink rate and saccade). A rule-based approach to generate primary (joyful, sad, angry, afraid, disgusted and surprise) and intermediate emotions (emotions that can be represented as the mixture of two primary emotions) utilized the MPEG4 FAPs (facial animation parameters) is introduced. Meanwhile, based on our research, a scripting tool, named EEMML (Emotional Eye Movement Markup Language) that enables authors to describe and generate emotional eye movement of virtual agents, is proposed.