Giter VIP home page Giter VIP logo

ha0tang / handgesturerecognition Goto Github PK

View Code? Open in Web Editor NEW
97.0 8.0 26.0 13.7 MB

[Neurocomputing 2019] Fast and Robust Dynamic Hand Gesture Recognition via Key Frames Extraction and Feature Fusion

Home Page: http://disi.unitn.it/~hao.tang/project/HandGestureRecognition.html

License: Other

Shell 0.02% MATLAB 19.06% M 0.05% C 5.66% Makefile 0.30% C++ 33.40% Python 2.10% HTML 29.56% Java 3.62% M4 2.26% TeX 3.78% CSS 0.17% Objective-C 0.02%
hand-gesture-recognition keyframe feature-fusion fast robust efficiency accuracy hand-gestures journal neurocomputing

handgesturerecognition's Introduction

License CC BY-NC-SA 4.0 Packagist Last Commit Maintenance Contributing Ask Me Anything ! GitHub issues

Key Frames Extraction and Feature Fusion for Dynamic Hand Gesture Recognition

Key Frames Extraction Framework

Key Frames Extraction Framework

Feature Fusion Module

Feature Fusion Module

Fast and Robust Dynamic Hand Gesture Recognition via Key Frames Extraction and Feature Fusion.
Hao Tang1, Hong Liu2, Wei Xiao3 and Nicu Sebe1.
1University of Trento, Italy, 2Peking University, China, 3Lingxi Artificial Intelligence Co., Ltd, China.
In Neurocomputing 2019.
The repository offers the official implementation of our paper in MATLAB.

Copyright (C) 2019 University of Trento, Italy.

All rights reserved. Licensed under the CC BY-NC-SA 4.0 (Attribution-NonCommercial-ShareAlike 4.0 International)

The code is released for academic research use only. For commercial use, please contact [email protected].

Installation

Clone this repo.

git clone https://github.com/Ha0Tang/HandGestureRecognition
cd HandGestureRecognition/

This code requires MATLAB. Please install it.

Dataset Preparation

For Cambridge Hand Gesture or Northwestern Hand Gesture, the datasets must be downloaded beforehand. Please download them on the respective webpages. Please cite their papers if you use the data.

Preparing Cambridge Hand Gesture Dataset. The dataset can be downloaded here. You can also download this dataset use the following script:

bash ./datasets/download_handgesture_dataset.sh Cambridge_Hand_Gesture

Preparing Northwestern Hand Gesture Dataset. The dataset is proposed in this paper. You can download this dataset use the following script:

bash ./datasets/download_handgesture_dataset.sh Northwestern_Hand_Gesture

Preparing HandGesture Dataset. This dataset consists of 132 video sequences of 640 by 360 resolution, each of which recorded from a different subject (7 males and 4 females) with 12 different gestures (“0”-“9”, “NO” and “OK”). Download this dataset use the following script:

bash ./datasets/download_handgesture_dataset.sh HandGesture

Preparing Action3D Dataset. This dataset consists of 1620 image sequences of 6 hand gesture classes (box, high wave, horizontal wave, curl, circle and hand up), which are defined by 2 different hands (right and left hand) and 5 situations (sit, stand, with a pillow, with a laptop and with a person). Each class contains 270 image sequences (5 different situations × 2 different hands × 3 times × 9 subjects). Each sequence was recorded in front of a fixed camera having roughly isolated gestures in space and time. All video sequences were uniformly resized into 320 × 240 in our method. Download this dataset use the following script:

bash ./datasets/download_handgesture_dataset.sh Action3D

Training New Models

New models can be trained with the following commands.

  1. Prepare your own dataset like in this folder.

  2. Extract key frame:

matlab -nodesktop -nosplash -r "key_frames_extraction"

Key frames will be extrated in the folder ./datasets/sample_keyframe.

  1. Go this folder for further processes.

Related Works

Citation

If you use this code for your research, please cite our papers.

@article{tang2019fast,
  title={Fast and Robust Dynamic Hand Gesture Recognition via Key Frames Extraction and Feature Fusion},
  author={Tang, Hao and Liu, Hong and Xiao, Wei and Sebe, Nicu},
  journal={Neurocomputing},
  volume={331},
  pages={424--433},
  year={2019},
  publisher={Elsevier}
}

Acknowledgments

This work is partially supported by National Natural Science Foundation of China (NSFC, U1613209), Shen- zhen Key Laboratory for Intelligent Multimedia and Virtual Reality (ZDSYS201703031405467), Scientific Research Project of Shenzhen City (JCYJ20170306164738129).

Contributions

If you have any questions/comments/bug reports, feel free to open a github issue or pull a request or e-mail to the author Hao Tang ([email protected]).

handgesturerecognition's People

Contributors

ha0tang avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

handgesturerecognition's Issues

训练

你好,我下载完Cambrige Hand Gesture 数据集后是不带标签的,请问可以给一个详细的数据集制作吗?谢谢

数据集

你好!请问可以将数据集以百度云链接分享吗?谢谢!

Questions about step4 and step5

Hello, I would like to ask you these questions:

  1. The step4_coding.m file is the sample code given by the vl-feat package. How do you set the parameters dimension, numFeatures, and numClusters in your paper? Can you give an example of the code used in the paper?
    I think numClusters is the number of gestures categories, and numDataToBeEncoded is the Dictionary Size, which is the abscissa of Figure 3 in your paper. Is this understanding correct?

  2. In the step5_classification_Cambridge.m file, are the data saved in the parameter
    maindir = 'F:\Myprojects\matlabProjects\featureExtraction\surf_feature\Cambridge_color_9_9entropy_4096\'
    visual words that have been coded?
    Assuming that the experiment has 30 gesture videos, first extract the key frames of 30 videos, then extract SIFT features (for example) from the key frames (step2), then perform word vectors clustering on the extracted features (step3), and then encode the word vectors. After that set the number of categories and dictionary size (step4), and finally perform classification prediction (step5); is it stored under the maindir folder in step5: The encoded word vectors extracted from these 30 videos, that is, there are 30 mat files respectively loaded Corresponding coded characteristic vocabulary.
    Is the following content stored in the maindir folder in the step5.m file: the encoded word vectors extracted from these 30 videos, that is, there are 30 mat files each containing the corresponding encoded words.

Thank you!

about ALGORITHM 2

Hello, I read your paper carefully.
Would you mind to leave some details about the algorithm 2?
Because I notice that you just upload the code about the key frame extraction not including the hand gesture recognition.
Thank you so much if you could help me.

Project page 404

Not Found

The requested URL /~hao.tang/project/HandGestureRecognition.html was not found on this server.

Apache/2.4.6 (CentOS) OpenSSL/1.0.2k-fips PHP/5.4.16 Server at [disi.unitn.it](mailto:[email protected]) Port 80

Help to run

Hi sir
I am a engineering student,
I want to this project, can you please guide me to complete this..
This is my mail id: [email protected]

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.