OmniHuman v1.5
Generates video from image and audio input with correlated emotions and movements using ByteDance's OmniHuman v1.5 model. Produces realistic talking avatars with natural lip-sync, facial expressions, and body movements synchronized to audio input. Advanced emotional understanding enables facial expressions and body language that match the emotional tone of the audio. Creates highly realistic talking-head videos suitable for presentations, explainers, and interactive applications.