The search functionality is under construction.

IEICE TRANSACTIONS on Information

Multimodal Affect Recognition Using Boltzmann Zippers

Kun LU, Xin ZHANG

  • Full Text Views

    0

  • Cite this

Summary :

This letter presents a novel approach for automatic multimodal affect recognition. The audio and visual channels provide complementary information for human affective states recognition, and we utilize Boltzmann zippers as model-level fusion to learn intrinsic correlations between the different modalities. We extract effective audio and visual feature streams with different time scales and feed them to two component Boltzmann chains respectively. Hidden units of the two chains are interconnected to form a Boltzmann zipper which can effectively avoid local energy minima during training. Second-order methods are applied to Boltzmann zippers to speed up learning and pruning process. Experimental results on audio-visual emotion data recorded by ourselves in Wizard of Oz scenarios and collected from the SEMAINE naturalistic database both demonstrate our approach is robust and outperforms the state-of-the-art methods.

Publication
IEICE TRANSACTIONS on Information Vol.E96-D No.11 pp.2496-2499
Publication Date
2013/11/01
Publicized
Online ISSN
1745-1361
DOI
10.1587/transinf.E96.D.2496
Type of Manuscript
LETTER
Category
Image Recognition, Computer Vision

Authors

Kun LU
  BIT
Xin ZHANG
  BIT

Keyword