All Projects → Svito-zar → speech-driven-hand-gesture-generation-demo

Svito-zar / speech-driven-hand-gesture-generation-demo

Licence: Apache-2.0 license
This repository contains the gesture generation model from the paper "Moving Fast and Slow" (https://www.tandfonline.com/doi/full/10.1080/10447318.2021.1883883) trained on the English dataset

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to speech-driven-hand-gesture-generation-demo

Speech driven gesture generation with autoencoder
This is the official implementation for IVA '19 paper "Analyzing Input and Output Representations for Speech-Driven Gesture Generation".
Stars: ✭ 76 (+300%)
Mutual labels:  gesture-controller, gesture-generation
embla-carousel-wheel-gestures
wheel interactions for Embla Carousel
Stars: ✭ 30 (+57.89%)
Mutual labels:  gestures
Gainput
Cross-platform C++ input library supporting gamepads, keyboard, mouse, touch
Stars: ✭ 636 (+3247.37%)
Mutual labels:  gestures
Mobile First Animation
Gesture-driven animation on the mobile web (React Conf 2019)
Stars: ✭ 1,998 (+10415.79%)
Mutual labels:  gestures
React Spring Lightbox
📷 A flexible image gallery lightbox with native-feeling touch gestures and buttery smooth animations, built with react-spring.
Stars: ✭ 76 (+300%)
Mutual labels:  gestures
Gebaar Libinput
Gebaar, A Super Simple WM Independent Touchpad Gesture Daemon for libinput
Stars: ✭ 249 (+1210.53%)
Mutual labels:  gestures
Advance
Physics-based animations for iOS, tvOS, and macOS.
Stars: ✭ 4,445 (+23294.74%)
Mutual labels:  gestures
React Native Easy Gestures
React Native Gestures. Support: Drag, Scale and Rotate a Component.
Stars: ✭ 153 (+705.26%)
Mutual labels:  gestures
Sign Language Interpreter Using Deep Learning
A sign language interpreter using live video feed from the camera.
Stars: ✭ 157 (+726.32%)
Mutual labels:  gestures
mix-stage
Official Repository for the paper Style Transfer for Co-Speech Gesture Animation: A Multi-Speaker Conditional-Mixture Approach published in ECCV 2020 (https://arxiv.org/abs/2007.12553)
Stars: ✭ 22 (+15.79%)
Mutual labels:  gesture-generation
Touche
The desktop application to configure Touchégg
Stars: ✭ 101 (+431.58%)
Mutual labels:  gestures
Reanimated 2 Animations Challenges
React Native animations challenges using the new Reanimated2 API.
Stars: ✭ 136 (+615.79%)
Mutual labels:  gestures
use-pan-and-zoom
👆+🔎 React hook for panning and zooming a container
Stars: ✭ 57 (+200%)
Mutual labels:  gestures
Swipe Listener
Zero-dependency, minimal swipe-gesture listener for the web.
Stars: ✭ 68 (+257.89%)
Mutual labels:  gestures
gestures
A library for normalized events and gesture for desktop and mobile.
Stars: ✭ 31 (+63.16%)
Mutual labels:  gestures
React Native Ratings
Tap and Swipe Ratings component for React Native.
Stars: ✭ 571 (+2905.26%)
Mutual labels:  gestures
Decomposed
CATransform3D manipulation made easy.
Stars: ✭ 184 (+868.42%)
Mutual labels:  gestures
BottomSheet
Example project showing how to implement a bottom sheet in Swift 5 using container/child view controllers, UIPanGestureRecognizer, and constraint animations.
Stars: ✭ 31 (+63.16%)
Mutual labels:  gestures
use-gesture
👇Bread n butter utility for component-tied mouse/touch gestures in React and Vanilla Javascript.
Stars: ✭ 6,624 (+34763.16%)
Mutual labels:  gestures
multi-touch
Multi-Touch XAML Behaviors implementing Multi-Touch Manipulation (Gestures) and Inertia.
Stars: ✭ 23 (+21.05%)
Mutual labels:  gestures

Speech-driven Hand Gesture Generation Demo

This repository can be used to reproduce our results of applying our model to the English dataset.

If you want to learn more about the model - this video is a good start.

Example of generate motion can be seen in the demo video.

Requirements

  • python 3
  • ffmpeg (to visualize the results)

Install dependencies

pip install --upgrade pip
pip install -r requirements.txt

Usage

./generate.sh  data/audio*.wav

Where in place of audio*.wav you can use any file from the folder data, which are chunks of the test sequences. Alternatively, you can download more audios for testing from the Trinity Speech-Gesture dataset. (The recordings 'NaturalTalking_01.wav' and 'NaturalTalking_02.wav' were not used in training and were left them for testing)

Training on your own data

For training on your own data we refer you to the original repository with the official implementation of the paper.

Citation

Here is the citation of our paper in bib format:

@article{kucherenko2021moving,
author = {Taras Kucherenko and Dai Hasegawa and Naoshi Kaneko and Gustav Eje Henter and Hedvig Kjellström},
title = {Moving Fast and Slow: Analysis of Representations and Post-Processing in Speech-Driven Automatic Gesture Generation},
journal = {International Journal of Human–Computer Interaction},
volume = {37},
number = {14},
pages = {1300-1316},
year  = {2021},
publisher = {Taylor & Francis},
doi = {10.1080/10447318.2021.1883883},
URL = {https://doi.org/10.1080/10447318.2021.1883883},
eprint = {https://doi.org/10.1080/10447318.2021.1883883}
}

If you are going to use Trinity Speech-Gesture dataset, please don't forget to cite them as described in their website

Contact

If you encounter any problems/bugs/issues please contact me on Github or by emailing me at [email protected] for any bug reports/questions/suggestions. I prefer questions and bug reports on Github as that provides visibility to others who might be encountering same issues or who have the same questions.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].