Multi-Stream GCN and CNN for Skeleton-Based Action Recognition

Kenzhebek Taniyev, Tomiris Zhaksylyk, Nguyen Anh Tu

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

This paper presents two novel approaches for improving skeleton-based action recognition using Graph Convolutional Networks (GCN) and Convolutional Neural Networks (CNN). In the first approach, we combine GCN and CNN streams that process position and velocity features to improve classification accuracy. In the second approach, we use GCN as an embedding layer for support network CNN to extract features from skeleton data, which significantly improves recognition accuracy. Our experiments on the JHMDB dataset demonstrate that our approaches outperform state-of-the-art methods while using significantly fewer parameters. Additionally, we extended our evaluation to the Kinetics-400 dataset, where our methods showed comparable results with considerably lower model complexity. Our work contributes to the development of more efficient and robust action recognition models.

Original languageEnglish
Title of host publication2024 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2024 - Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages1718-1722
Number of pages5
ISBN (Electronic)9781665410205
DOIs
Publication statusPublished - 2024
Event2024 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2024 - Kuching, Malaysia
Duration: Oct 6 2024Oct 10 2024

Publication series

NameConference Proceedings - IEEE International Conference on Systems, Man and Cybernetics
ISSN (Print)1062-922X

Conference

Conference2024 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2024
Country/TerritoryMalaysia
CityKuching
Period10/6/2410/10/24

ASJC Scopus subject areas

  • Electrical and Electronic Engineering
  • Control and Systems Engineering
  • Human-Computer Interaction

Fingerprint

Dive into the research topics of 'Multi-Stream GCN and CNN for Skeleton-Based Action Recognition'. Together they form a unique fingerprint.

Cite this