All Projects → hysts → pytorch_mpiigaze_demo

hysts / pytorch_mpiigaze_demo

Licence: MIT license
Gaze estimation using MPIIGaze and MPIIFaceGaze

Programming Languages

python
139335 projects - #7 most used programming language
Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to pytorch mpiigaze demo

gaze-estimation-with-laser-sparking
Deep learning based gaze estimation demo with a fun feature :-)
Stars: ✭ 32 (-83.67%)
Mutual labels:  gaze, gaze-estimation
attention-target-detection
[CVPR2020] "Detecting Attended Visual Targets in Video"
Stars: ✭ 105 (-46.43%)
Mutual labels:  gaze, gaze-estimation
awesome-gaze
A curated list of awesome gaze estimation frameworks, datasets and other awesomeness.
Stars: ✭ 60 (-69.39%)
Mutual labels:  gaze, gaze-estimation
GazeToolkit
Toolkit for processing eye movement data, fixation filtering, smoothing, etc.
Stars: ✭ 21 (-89.29%)
Mutual labels:  gaze
edusense
EduSense: Practical Classroom Sensing at Scale
Stars: ✭ 44 (-77.55%)
Mutual labels:  gaze
flame
Original Pytorch Implementation of FLAME: Facial Landmark Heatmap Activated Multimodal Gaze Estimation
Stars: ✭ 13 (-93.37%)
Mutual labels:  gaze-estimation
Laser-Eye
Gaze Estimation via Deep Neural Networks
Stars: ✭ 79 (-59.69%)
Mutual labels:  gaze-estimation
GazeML-keras
A keras port of swook/GazeML for pupil, iris and eye-lid detection
Stars: ✭ 44 (-77.55%)
Mutual labels:  gaze-estimation
GPCycleGAN
Datasets, code and models from our driver gaze estimation works over the last few years.
Stars: ✭ 35 (-82.14%)
Mutual labels:  gaze-estimation

A demo program of gaze estimation models (MPIIGaze, MPIIFaceGaze, ETH-XGaze)

PyPI version Downloads Open In Colab MIT License GitHub stars

With this program, you can run gaze estimation on images and videos. By default, the video from a webcam will be used.

ETH-XGaze video01 result ETH-XGaze video02 result ETH-XGaze video03 result

MPIIGaze video00 result MPIIFaceGaze video00 result

MPIIGaze image00 result

To train a model for MPIIGaze and MPIIFaceGaze, use this repository. You can also use this repo to train a model with ETH-XGaze dataset.

Quick start

This program is tested only on Ubuntu.

Installation

pip install ptgaze

Run demo

ptgaze --mode eth-xgaze

Usage

usage: ptgaze [-h] [--config CONFIG] [--mode {mpiigaze,mpiifacegaze,eth-xgaze}]
              [--face-detector {dlib,face_alignment_dlib,face_alignment_sfd,mediapipe}]
              [--device {cpu,cuda}] [--image IMAGE] [--video VIDEO] [--camera CAMERA]
              [--output-dir OUTPUT_DIR] [--ext {avi,mp4}] [--no-screen] [--debug]

optional arguments:
  -h, --help            show this help message and exit
  --config CONFIG       Config file. When using a config file, all the other commandline arguments
                        are ignored. See
                        https://github.com/hysts/pytorch_mpiigaze_demo/ptgaze/data/configs/eth-
                        xgaze.yaml
  --mode {mpiigaze,mpiifacegaze,eth-xgaze}
                        With 'mpiigaze', MPIIGaze model will be used. With 'mpiifacegaze',
                        MPIIFaceGaze model will be used. With 'eth-xgaze', ETH-XGaze model will be
                        used.
  --face-detector {dlib,face_alignment_dlib,face_alignment_sfd,mediapipe}
                        The method used to detect faces and find face landmarks (default:
                        'mediapipe')
  --device {cpu,cuda}   Device used for model inference.
  --image IMAGE         Path to an input image file.
  --video VIDEO         Path to an input video file.
  --camera CAMERA       Camera calibration file. See https://github.com/hysts/pytorch_mpiigaze_demo/
                        ptgaze/data/calib/sample_params.yaml
  --output-dir OUTPUT_DIR, -o OUTPUT_DIR
                        If specified, the overlaid video will be saved to this directory.
  --ext {avi,mp4}, -e {avi,mp4}
                        Output video file extension.
  --no-screen           If specified, the video is not displayed on screen, and saved to the output
                        directory.
  --debug

While processing an image or video, press the following keys on the window to show or hide intermediate results:

  • l: landmarks
  • h: head pose
  • t: projected points of 3D face model
  • b: face bounding box

References

  • Zhang, Xucong, Seonwook Park, Thabo Beeler, Derek Bradley, Siyu Tang, and Otmar Hilliges. "ETH-XGaze: A Large Scale Dataset for Gaze Estimation under Extreme Head Pose and Gaze Variation." In European Conference on Computer Vision (ECCV), 2020. arXiv:2007.15837, Project Page, GitHub
  • Zhang, Xucong, Yusuke Sugano, Mario Fritz, and Andreas Bulling. "Appearance-based Gaze Estimation in the Wild." Proc. of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015. arXiv:1504.02863, Project Page
  • Zhang, Xucong, Yusuke Sugano, Mario Fritz, and Andreas Bulling. "It's Written All Over Your Face: Full-Face Appearance-Based Gaze Estimation." Proc. of the IEEE Conference on Computer Vision and Pattern Recognition Workshops(CVPRW), 2017. arXiv:1611.08860, Project Page
  • Zhang, Xucong, Yusuke Sugano, Mario Fritz, and Andreas Bulling. "MPIIGaze: Real-World Dataset and Deep Appearance-Based Gaze Estimation." IEEE transactions on pattern analysis and machine intelligence 41 (2017). arXiv:1711.09017
  • Zhang, Xucong, Yusuke Sugano, and Andreas Bulling. "Evaluation of Appearance-Based Methods and Implications for Gaze-Based Applications." Proc. ACM SIGCHI Conference on Human Factors in Computing Systems (CHI), 2019. arXiv, code
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].