The search functionality is under construction.
The search functionality is under construction.

Efficient Action Spotting Using Saliency Feature Weighting

Yuzhi SHI, Takayoshi YAMASHITA, Tsubasa HIRAKAWA, Hironobu FUJIYOSHI, Mitsuru NAKAZAWA, Yeongnam CHAE, Björn STENGER

  • Full Text Views

    0

  • Cite this

Summary :

Action spotting is a key component in high-level video understanding. The large number of similar frames poses a challenge for recognizing actions in videos. In this paper we use frame saliency to represent the importance of frames for guiding the model to focus on keyframes. We propose the frame saliency weighting module to improve frame saliency and video representation at the same time. Our proposed model contains two encoders, for pre-action and post-action time windows, to encode video context. We validate our design choices and the generality of proposed method in extensive experiments. On the public SoccerNet-v2 dataset, the method achieves an average mAP of 57.3%, improving over the state of the art. Using embedding features obtained from multiple feature extractors, the average mAP further increases to 75%. We show that reducing the model size by over 90% does not significantly impact performance. Additionally, we use ablation studies to prove the effective of saliency weighting module. Further, we show that our frame saliency weighting strategy is applicable to existing methods on more general action datasets, such as SoccerNet-v1, ActivityNet v1.3, and UCF101.

Publication
IEICE TRANSACTIONS on Information Vol.E107-D No.1 pp.105-114
Publication Date
2024/01/01
Publicized
2023/10/17
Online ISSN
1745-1361
DOI
10.1587/transinf.2022EDP7210
Type of Manuscript
PAPER
Category
Image Processing and Video Processing

Authors

Yuzhi SHI
  Chubu University
Takayoshi YAMASHITA
  Chubu University
Tsubasa HIRAKAWA
  Chubu University
Hironobu FUJIYOSHI
  Chubu University
Mitsuru NAKAZAWA
  Rakuten Group, Inc.
Yeongnam CHAE
  Rakuten Group, Inc.
Björn STENGER
  Rakuten Group, Inc.

Keyword