Skip to content

monchhichizzq/Real-time-human-action-recognition

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

23 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Real-time-human-action-recognition by 2D input pose

With the rapid development of deep learning algorithms and high-level calculation devices, the intelligent system can be in charge of more complicated and challengeable missions for human activity recognition with higher accuracy than before. Industrial partners have already showed their great interests in the implementation of AI in real-life applications, such as autonomous driving, human-robot collaboration and healthcare monitoring. In order to fit the growing industrial needs, more efficient and lower-latency gesture recognition methods should be developed to solve the real-time problems in those intelligent systems.

This master project is based on the idea of input 2D skeletal body joints for an LSTM RNN Network to complete the human gesture recognition in real time. Our assumption is:

1 Deep Learning algorithms (CNN, RNN) could help us to solve the real-time problems in gesture recognition without a drop in accuracy compared to the traditional machine learning methods (random forest, SVM, Logistic Regression, ...)

2 Using a series of 2D poses, rather than 3D poses or a raw 2D images, can produce an accurate estimation of the behavior of a person

3 Combining the pre-trained network 'openpose' (https://github.com/CMU-Perceptual-Computing-Lab/openpose's) and self-trained RNN could make the gesture recognition in real-time with high accuracy and low latency, compared to the existing posture recognition methods, or other offline gesture recognition methods. However, we also expect this method would exceed the state-of-the-art performance (like 3D CNN, TSN) in real-time and accuracy

Run the inference

  1. Download the latest 'Openpose' Release here: https://github.com/CMU-Perceptual-Computing-Lab/openpose/releases
    • GPU accelerating release could reach $\sim$ 20FPS on the video processing (strongly recommended - openpose-1.6.0-binaries-win64-gpu-python-flir-3d_recommended.zip)
    • Enable model Body-25 during the test, which gives the best pose estimation accuracy and higest processing speed
  2. Download KTH human actions dataset :https://www.csc.kth.se/cvap/actions/
  3. Enter test and Put your test videos in the folder 'video_samples'
  4. Put the pretrained mmodel in the folder 'models'
  5. Set the gpu for real-time test
    Nvida GPU (GTX 2070)\ os.environ["CUDA_VISIBLE_DEVICES"]= "i", i=0, 1, 2...
    One gpu would be enough to run the test, you could modify the parameters (self.gpu_lstm, self.gpu_openpose, self.gpu0_free) to assign the GPU memory for different inference tasks
  6. Run testBodymotion.py (** Load the model and warm it up)
  7. Record your own video and test it \ Remember to record your whole body and follow the standarlized actions in the video dataset for an accurate prediction

Demo

image image

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages