Wan-2.1
Provided by Fal — Learn More
Wan-2.1 is an advanced and powerful visual generation model developed by Tongyi Lab. It can generate videos based on text, images and other control signals. Wan-2.1 excels at generating realistic videos featuring extensive body movements, complex rotations, dynamic scene transitions, and fluid camera motions, and can accurately simulate real-world physics and realistic object interactions, while offering movie-like visuals with rich textures and a variety of stylized effects. It can also create text and dynamic text effects in videos directly from text prompts.
Vidu Reference
Vidu Reference is a video generation model that enables seamless interaction between multiple subjects, including characters, props, objects, and environments in the same scene. This model is ideal for creating videos with complex scenes and multiple characters, where characters interact naturally within the same scene. Vidu supports feature fusion, allowing elements from different subjects—such as the front of Character A and the back of Character B—to merge seamlessly into a new character or object.
Wan-2.2 Video
Wan-2.2 is a leading-edge and highly capable image and video generation model developed by Tongyi Lab of Alibaba Group. Wan-2.2 achieves professional cinematic narratives through a deep command of shot language, offering fine-grained control over lighting, color, and composition for versatile styles with delicate detail. It effortlessly recreates all kinds of complex motion, with enhanced fluidity and control, with better understanding and execution of prompts for complex scenes and multi-object generation. Wan-2.2 can generalize across multiple dimensions such as motions, semantics, and aesthetics. In addition to text-to-video and image-to-video, Wan-2.2 also supports video-to-video generation, with the ability to perform a wide range of edits on an input video such as adding, removing, and transforming objects.