we're building a foundational model to create, modify, and understand humans in video.
we're starting by learning specific behaviors and building specialized generative computer vision models.
the first api we released is a zero-shot state-of-the-art lipsync model our customers use it to animate anyone in any video to say (or sing) anything you want in any language.