Project Icon

speech-driven-animation

Generate Facial Animations Through Speech Synthesis Models

Product DescriptionThis library provides tools for generating speech-driven facial animations using an end-to-end synthesis model. Developed by Konstantinos Vougioukas, Honglie Chen, and Pingchuan Ma, this library features models trained on datasets such as GRID, TCD-TIMIT, and CREMA-D. The VideoAnimator class supports input through both audio files and numpy arrays. Model files are available on GoogleDrive. Audio and video encoders included can be used for feature extraction in classification tasks, making it a valuable resource for researchers.
Project Details