# Generate video

The core video-generation flow. Pick a video mode from the visualizer's mode dropdown, pick a model, write or adjust the prompt, click **Generate Video** (or **Capture Animation** in Video From Animation). The result lands in the gallery as a video file with playback controls.

## Watch

{% embed url="<https://www.youtube.com/watch?v=gC8yn3Zyedo>" %}

![Visualize mode in video mode, Kling 2.5 Turbo model selected, prompt panel and Generate Video button on the right, an authored soccer-stadium scene in the viewport](/files/BbJB8S4Gt3Yzvv5rVJBv)

## What it does

Reads the 3D scene plus the active shot's camera, assembles the prompt, picks the chosen video model, and generates a clip whose duration matches the shot's timeline. Animated subjects and animated cameras are honored if the shot has them; static shots produce subtle camera motion baked in by the model.

The video models on hand: Veo 3.1, Veo 3.1 Fast (Google); Kling 2.0 Master, 2.1 Standard, 2.1 Pro, 2.1 Master, 2.5 Turbo, 2.6 Pro, 3 Standard, 3 Pro (Kuaishou); Luma Ray 3, Luma Ray 3.14 (Luma); Seedance 2.0 (ByteDance); Runway Gen 4.5 (Runway). See [Models](/visualize/ai-models.md) for full guidance.

## The four visualizer modes

The mode dropdown at the top of the visualizer panel switches the panel between four jobs. Three of the four produce video.

![Visualizer mode dropdown open with four options: Image, Video (Start Frame), Video (Start + End Frames), Video (From Animation)](/files/TcBTHfAwbzJyr0ype74P)

* **Image.** Still image generation from the scene plus prompt. Covered in [Generate image](/visualize/generate-image.md).
* **Video (Start Frame).** Video from one starting image plus prompt. The model invents the motion. The default video mode.
* **Video (Start + End Frames).** Video from two anchor images. The model interpolates the motion between them. Covered in [First and last frame](/visualize/first-and-last-frame.md).
* **Video (From Animation).** Video driven by the keyframes you set in Compose. The model follows your authored motion rather than inventing its own, or Direct Render captures the scene with no AI. Covered in [Video (From Animation)](/visualize/video-from-animation.md).

The model picker is scoped to the active mode. Pick the mode that matches the job, then the dropdown shows you the models that support it.

## How to use it

1. **Set the mode.** Click the mode dropdown and pick **Video (Start Frame)** for the default flow, or one of the other two video modes for the alternative jobs.
2. **Pick a video model.** Different strengths per model; see the table below for a starting heuristic.
3. **Populate the inputs the mode requires.** Start Frame for Video (Start Frame); Start and End Frame for Video (Start + End Frames); nothing extra for Video (From Animation) because the scene drives it.
4. **Set the three video options:** Duration (e.g. 4 seconds), Orientation (Landscape or Portrait), and Audio (No audio, or Audio enabled). The available options depend on the chosen model.
5. **Read the auto-prompt.** Same Scene / Style / Lighting tabs as image generation. Confirm the prompt reads right before committing. For audio-capable models, include the sounds and dialog you want in the prompt (see below).
6. **Click Generate Video** (or **Capture Animation** in Video From Animation). Generation takes longer than image – often several minutes for complex shots. The job runs in the background; you can keep working in other parts of the project. A toast notifies when the result lands.

![Visualize mode in video mode with Veo 3.1 selected, Start Frame and End Frame slots visible, plus three dropdowns - 4 seconds (Duration), Landscape (Orientation), No audio (Audio) - above the Generate Video button](/files/MfQ2sHOHUdj3Bqjo9xzd)

## Choosing a video model

Quick heuristics. See [Models](/visualize/ai-models.md) for the full table.

| If the shot is...                                     | Try                                                                    |
| ----------------------------------------------------- | ---------------------------------------------------------------------- |
| Action-heavy with fast camera moves                   | Kling 2.6 Pro or Kling 3 Pro                                           |
| A long, slow, atmospheric move                        | Veo 3.1                                                                |
| Faster generation at lower fidelity                   | Veo 3.1 Fast or Kling 2.5 Turbo                                        |
| Honoring detailed 3D structure (like animated meshes) | Luma Ray 3                                                             |
| First-and-last frame interpolation                    | Veo 3.1, Kling 2.6 Pro or later, or Seedance 2.0                       |
| Quick experimentation across many variants            | Kling 2.5 Turbo                                                        |
| Brand or product hero with strict consistency         | Match the image model that worked, then pick its closest video sibling |

## Authoring audio with the prompt

A subset of the video models generate audio along with the video. Veo 3.1 supports it natively. Kling 2.6 Pro and later expose an optional audio track. Other models are video-only; the Audio toggle on the panel is disabled when an audio-incapable model is selected.

When you enable audio on a supporting model, the prompt drives both the video and the audio. Describe the sounds you want in the same prompt:

* **Dialog.** Write the lines a character speaks. Models that generate dialog will lip-sync to it within their tolerance.
* **Sound effects.** Name the SFX you want and where they happen in the scene, prose-style: "the squeak of shopping carts, low conversation, a distant intercom announcement."
* **Music or ambience.** Describe the mood or genre: "low cinematic strings under the dialog", "wind moving through dry grass."

There's no separate audio prompt field. Whatever you write in the main prompt is what the model uses for both.

![Video model dropdown with audio-capable models grouped by provider, including Veo 3.1 with audio support and Kling 2.6 Pro with optional generated audio](/files/WeJavIrV00blPFCpa105)

Cost note: audio adds materially to the per-clip cost on the models that support it. Run audio-off iterations to lock the visual, then add audio on the final pass.

## Animation feeds the model

```mermaid
flowchart LR
    A[Compose-mode shot animation] --> B[Camera moves over time]
    A --> C[Object animates over time]
    B --> D[Visualizer]
    C --> D
    D --> E[Video model]
    E --> F[Rendered video that honors the motion]
```

If your shot has authored camera animation (keyframes) or animated subjects, the visualizer feeds the per-frame state to the video model. The result tracks the motion you authored. Static shots produce subtle motion baked in by the model itself; if you want specific camera action, author it in Compose mode first.

When motion fidelity matters more than the model's interpretation, use [Video (From Animation)](/visualize/video-from-animation.md). For wireframe preview with no AI in the loop, use [Direct Render](/visualize/direct-render.md).

{% hint style="info" %}
Video at 2K with audio is the most expensive single action in the product. Run 1K previews first to confirm composition; commit to 2K only when the shot is locked.
{% endhint %}

## When the video is wrong

* **The motion isn't what you authored.** The video model is reinterpreting your animation. Try a different model: Luma Ray 3 honors authored 3D motion most strictly; some Kling variants take more liberties.
* **The subject drifts across frames.** Image reference on the hero object helps, but video models honor references less tightly than image models. Multi-view references and consistent object names help most.
* **The render is too long or too short.** Adjust the shot's duration in [Shot details](/compose/shot-details.md). The video duration matches the shot.

## Limits and known issues

* **Video-to-video isn't supported.** A rendered video can't be fed back into the visualizer as a motion reference.
* **Audio doesn't sync to specific moments.** The audio track is generated holistically; you can't specify "audio cue at frame 30". For precise sync, generate without audio and add the audio in your editing tool.
* **Orientation options depend on the model.** Most video models default to landscape; Veo exposes landscape and portrait. Other models may offer only one. The Orientation dropdown reflects what the active model supports.

## Related

* [Generate image](/visualize/generate-image.md)
* [First and last frame](/visualize/first-and-last-frame.md) – the Video (Start + End Frames) mode in detail.
* [Video (From Animation)](/visualize/video-from-animation.md) – use your authored keyframes to drive the motion.
* [Direct Render](/visualize/direct-render.md) – render the scene as wireframe video with no AI.
* [Models](/visualize/ai-models.md)
* [Animation](/compose/animation.md)
* [Resolution and cost](/visualize/resolution-and-cost.md)
* [Managing outputs](/visualize/managing-outputs.md)


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://help.intangible.ai/visualize/generate-video.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
