Wan AI: Wan 2.7 Text-to-Video

wan-ai/wan2.7-t2v

Access Wan 2.7 Text-to-Video AI video generation using Puter.js API.

Get Started

Model Card

Wan 2.7 Text-to-Video is a diffusion-based video generation model from Alibaba, designed to produce cinematic video clips directly from text prompts.

It generates native 720p and 1080p video with durations from 2 to 15 seconds, supporting flexible aspect ratios including 16:9, 9:16, and 1:1. A standout feature is optional audio input, which synchronizes character motion and lip movement to a provided audio track during generation.

The model responds well to detailed, structured prompts and supports multi-shot narrative control through prompt language alone. It's part of a broader four-model suite that includes image-to-video, reference-to-video, and video editing capabilities.

Best suited for marketing content, social media clips, film pre-visualization, and any production pipeline that needs programmatic access to high-quality video generation.

Max Duration N/A

seconds

Frame Rate N/A

fps

Aspect Ratio N/A

supported

Release Date Apr 3, 2026

 

Code Example

Use Wan 2.7 Text-to-Video in your app with the Puter.js AI API — no API keys or setup required.

// npm install @heyputer/puter.js
import { puter } from '@heyputer/puter.js';

puter.ai.txt2vid("A cat playing with a ball of yarn", {
    model: "wan-ai/wan2.7-t2v"
}).then(video => {
    document.body.appendChild(video);
});
<html>
<body>
    <script src="https://js.puter.com/v2/"></script>
    <script>
        puter.ai.txt2vid("A cat playing with a ball of yarn", {
            model: "wan-ai/wan2.7-t2v"
        }).then(video => {
            document.body.appendChild(video);
        });
    </script>
</body>
</html>

More AI Models From Wan AI

Find other Wan AI models

Image

Wan2.6 Image

Wan 2.6 Image is a 20-billion-parameter diffusion-based image generation and transformation model developed by Alibaba Cloud. Built on a Multimodal Diffusion Transformer (MMDiT) architecture, it supports text-to-image generation, image-to-image editing, and multi-reference style transfer. The model accepts up to three reference images per request, allowing developers to explicitly control style, subject, and composition by referencing inputs in the prompt (e.g., "image 1" for style, "image 2" for background). It generates outputs up to 2048×2048 pixels across a wide range of aspect ratios. Wan 2.6 Image is particularly strong at localized content generation, with sophisticated understanding of Asian cultural contexts and aesthetics. It's well suited for e-commerce product visualization, brand asset creation, marketing materials, and automated content pipelines where controllability and visual consistency matter more than pure artistic exploration.

Video

Wan 2.2 Image-to-Video 14B

Wan 2.2 I2V A14B is an open-source image-to-video generation model that transforms static images into 5-second videos at 480P or 720P resolution. It uses a Mixture-of-Experts (MoE) architecture with dual 14B-parameter experts to achieve stable video synthesis with reduced unrealistic camera movements and enhanced support for diverse stylized scenes.

Video

Wan 2.2 Text-to-Video 14B

Wan 2.2 T2V A14B is an open-source text-to-video generation model that creates 5-second videos at 480P or 720P resolution from text prompts. Built with a Mixture-of-Experts (MoE) architecture featuring specialized high-noise and low-noise experts, it delivers cinematic-quality output with granular control over lighting, composition, and motion.

Frequently Asked Questions

How do I use Wan 2.7 Text-to-Video?

You can access Wan 2.7 Text-to-Video by Wan AI through Puter.js AI API. Include the library in your web app or Node.js project and start making calls with just a few lines of JavaScript — no backend and no configuration required.

Is Wan 2.7 Text-to-Video free?

Yes, it is free if you're using it through Puter.js. With the User-Pays Model, you can add Wan 2.7 Text-to-Video to your app at no cost — your users pay for their own AI usage directly, making it completely free for you as a developer.

Who created Wan 2.7 Text-to-Video?

Wan 2.7 Text-to-Video was created by Wan AI and released on Apr 3, 2026.

Does it work with React / Vue / Vanilla JS / Node / etc.?

Yes — the Wan 2.7 Text-to-Video API works with any JavaScript framework, Node.js, or plain HTML through Puter.js. Just include the library and start building. See the documentation for more details.

Get started with Puter.js

Add Wan 2.7 Text-to-Video video generation to your app without worrying about API keys or setup.

Read the Docs View Tutorials