Multi-skeleton structures graph convolutional network for action quality assessment in long videos

Qing Lei, Huiying Li, Hongbo Zhang, Jixiang Du, Shangce Gao*

*この論文の責任著者

研究成果: ジャーナルへの寄稿学術論文査読

5 被引用数 (Scopus)

抄録

In most existing action quality assessment (AQA) methods, how to score simple actions in short-term sport videos has been widely explored. Recently, a few studies have attempted to solve the AQA problem of long-duration activity by extracting dynamic or static information directly from RGB video. However, these methods may ignore specific postures defined by dynamic changes in human body joints, which makes the results inaccurate and unexplainable. In this work, we propose a novel graph convolution network based on multiple skeleton structure modelling to address the problem of effective pose feature learning to improve the performance of AQA in complex activity. Specifically, three kinds of skeleton structures, including the joints’ self-connection, the intra-part connection, and the inter-part connection, are defined to model the motion patterns of joints and body parts. Moreover, a temporal attention learning module is designed to extract temporal relations between skeleton subsequences. We evaluate the proposed method on two benchmark datasets, the MIT-skate dataset and the Rhythmic Gymnastics dataset. Extensive experiments are conducted to verify the effectiveness of the proposed method. The experimental results show that our method achieves state-of-the-art performance.

本文言語英語
ページ(範囲)21692-21705
ページ数14
ジャーナルApplied Intelligence
53
19
DOI
出版ステータス出版済み - 2023/10

ASJC Scopus 主題領域

  • 人工知能

フィンガープリント

「Multi-skeleton structures graph convolutional network for action quality assessment in long videos」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル