Veo 3.1 Fast
Veo 3.1 Fast is Google’s upgraded video-generation model, released in paid preview within the Gemini API alongside Veo 3.1. It enables developers to create cinematic, high-quality videos from text prompts or reference images at a much faster processing speed. The model introduces native audio generation with natural dialogue, ambient sound, and synchronized effects for lifelike storytelling. Veo 3.1 Fast also supports advanced controls such as “Ingredients to Video,” allowing up to three reference images, “Scene Extension” for longer sequences, and “First and Last Frame” transitions for seamless shot continuity. Built for efficiency and realism, it delivers improved image-to-video quality and character consistency across multiple scenes. With direct integration into Google AI Studio and Vertex AI, Veo 3.1 Fast empowers developers to bring creative video concepts to life in record time.
Learn more
Act-Two
Act-Two enables animation of any character by transferring movements, expressions, and speech from a driving performance video onto a static image or reference video of your character. By selecting the Gen‑4 Video model and then the Act‑Two icon in Runway’s web interface, you supply two inputs; a performance video of an actor enacting your desired scene and a character input (either a single image or a video clip), and optionally enable gesture control to map hand and body movements onto character images. Act‑Two automatically adds environmental and camera motion to still images, supports a range of angles, non‑human subjects, and artistic styles, and retains original scene dynamics when using character videos (though with facial rather than full‑body gesture mapping). Users can adjust facial expressiveness on a sliding scale to balance natural motion with character consistency, preview results in real time, and generate high‑resolution clips up to 30 seconds long.
Learn more
Ray3
Ray3 is an advanced video generation model by Luma Labs, built to help creators tell richer visual stories with pro-level fidelity. It introduces native 16-bit High Dynamic Range (HDR) video generations, enabling more vibrant color, deeper contrasts, and overall pro studio pipelines. The model incorporates sophisticated physics and improved consistency (motion, anatomy, lighting, reflections), supports visual controls, and has a draft mode that lets you explore ideas quickly before up-rendering selected pieces into high-fidelity 4K HDR output. Ray3 can interpret prompts with nuance, reason about intent, self-evaluate early drafts, and adjust to satisfy the articulation of scene and motion more accurately. Other features include support for keyframes, loop and extend functions, upscaling, and export of frames for seamless integration into professional workflows.
Learn more
Seedance
Seedance 1.0 API is officially live, giving creators and developers direct access to the world’s most advanced generative video model. Ranked #1 globally on the Artificial Analysis benchmark, Seedance delivers unmatched performance in both text-to-video and image-to-video generation. It supports multi-shot storytelling, allowing characters, styles, and scenes to remain consistent across transitions. Users can expect smooth motion, precise prompt adherence, and diverse stylistic rendering across photorealistic, cinematic, and creative outputs. The API provides a generous free trial with 2 million tokens and affordable pay-as-you-go pricing from just $1.8 per million tokens. With scalability and high concurrency support, Seedance enables studios, marketers, and enterprises to generate 5–10 second cinematic-quality videos in seconds.
Learn more