Starred repositories
The "classic" version of the Forge WebUI
Powerful & Easy-to-Use Video Face Swapping and Editing Software
Official implementation of “GaussianTalker: Real-Time High-Fidelity Talking Head Synthesis with Audio-Driven 3D Gaussian Splatting” by Kyusun Cho, Joungbin Lee, Heeji Yoon, Yeobin Hong, Jaehoon Ko,…
[ACM MM 2025] Ditto: Motion-Space Diffusion for Controllable Realtime Talking Head Synthesis
Project Page for paper “Exploring Timeline Control for Facial Motion Generation” (CVPR 2025)
HumanAIGC / Wan2.2
Forked from Wan-Video/Wan2.2Wan: Open and Advanced Large-Scale Video Generative Models
[NeurIPS 2025] Let Them Talk: Audio-Driven Multi-Person Conversational Video Generation
Have a natural, spoken conversation with AI!
MoDA: Multi-modal Diffusion Architecture for Talking Head Generation
The official code of our ICCV2023 work: Implicit Identity Representation Conditioned Memory Compensation Network for Talking Head video Generation
Freeform Body Motion Generation from Speech
The authors' implementation of the "Neural Head Reenactment with Latent Pose Descriptors" (CVPR 2020) paper.
PyTorch implementation for NED (CVPR 2022). It can be used to manipulate the facial emotions of actors in videos based on emotion labels or reference styles.
The pytorch implementation of our WACV23 paper "Cross-identity Video Motion Retargeting with Joint Transformation and Synthesis".
lipsync is a simple and updated Python library for lip synchronization, based on Wav2Lip. It synchronizes lips in videos and images based on provided audio, supports CPU/CUDA, and uses caching for …
Code for ACCV 2020 "Speech2Video Synthesis with 3D Skeleton Regularization and Expressive Body Poses"
Long-Inference, High Quality Synthetic Speaker (AI avatar/ AI presenter)
[CVPR-2025] The official code of HunyuanPortrait: Implicit Condition Control for Enhanced Portrait Animation
DiffPoseTalk: Speech-Driven Stylistic 3D Facial Animation and Head Pose Generation via Diffusion Models
[CVPR2023] OTAvatar: One-shot Talking Face Avatar with Controllable Tri-plane Rendering.
[NeurIPS 2025] OmniTalker: Real-Time Text-Driven Talking Head Generation with In-Context Audio-Visual Style Replication
ICCV 2025 ACTalker: an end-to-end video diffusion framework for talking head synthesis that supports both single and multi-signal control (e.g., audio, expression).
A Survey on Deepfake Generation and Detection
CVPR2023 talking face implementation for Identity-Preserving Talking Face Generation With Landmark and Appearance Priors
Official code for CVPR2022 paper: Depth-Aware Generative Adversarial Network for Talking Head Video Generation