Project Icon

DiffSHEG

Diffusion-Based Real-Time Speech-Driven 3D Expression and Gesture Generation

Product DescriptionThis project utilizes a diffusion-based method to generate realistic 3D expressions and gestures in real-time from speech input. It is optimized for Ubuntu and uses dependencies like PyTorch, with models trained on datasets like BEAT and SHOW. Ideal for AI and 3D animation fields, it includes guidance for custom audio input inference and uses tools like Blender for visualization, enhanced by contributions from top researchers.
Project Details