Skip to content

hsyntemiz/VA-ES-CNNs-for-Skeleton-based-Human-Action-Recognition

 
 

Repository files navigation

View-Adaptive-E(2)-Equivariant-Steerable-CNNs-for-Skeleton-based-Human-Action-Recognition

Introduction

Regardless of input type, humans in an action can be seen from different viewpoints. Therefore, action recognition methods should be robust to viewpoint variation.To relieve the adverse effects of view variations, Zhang et al introduce a novel view adaptation scheme. In this repository, we investigate the performance of Equivariant-Steerable-CNNs in the view adaptation module.

Our Report

This repository forked from the codes of the following paper:

View Adaptive Neural Networks for High Performance Skeleton-based Human Action Recognition. TPAMI, 2019.

Flowchart

image

Figure 1: Pipeline of the end-to-end view adaptive neural network.

Prerequisites

The code is built with following libraries:

Data Preparation

We use the NTU60 RGB+D dataset as an example for description. We need to first dowload the NTU-RGB+D dataset

  • Extract the dataset to ./data/ntu/nturgb+d_skeletons/
  • Process the data
 cd ./data/ntu
 # Get skeleton of each performer
 python get_raw_skes_data.py
 # Remove the bad skeleton 
 python get_raw_denoised_data.py
 # Transform the skeleton to the center of the first frame
 python seq_transformation.py

Training

# For CNN-based model with view adaptation module
python  va-cnn.py --model VA --aug 1 --train 1 

# For steer-CNN-based model with view adaptation module
python  va-cnn.py --model VA --aug 1 --train 1 --steer

# For CNN-based model without view adaptation module
python  va-cnn.py --model baseline --aug 1 --train 1




## Testing

```bash
# For CNN-based model with view adaptation module
python  va-cnn.py --model VA --aug 1 --train 0

# For steer-CNN-based model with view adaptation module
python  va-cnn.py --model VA --aug 1 --train 0 --steer

# For CNN-based model without view adaptation module
python  va-cnn.py --model baseline --aug 1 --train 0



## Reference
If you find the papers and repo useful, you can cite the paper: 

View Adaptive Neural Networks for High Performance Skeleton-based Human Action Recognition. TPAMI, 2019.
Microsoft Open Source Code of Conduct: https://opensource.microsoft.com/codeofconduct

About

View Adaptive Neural Networks for High Performance Skeleton-based Human Action Recognition

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%