Wan 2.2 First Last Frame Video

Published Categorized as Tutorial Tagged , , No Comments on Wan 2.2 First Last Frame Video

This Wan 2.2 image-to-video workflow lets you fix the first and last frames and generates a video connecting the two (FLF2V). See the example below.

Input images:

Output video:

I will provide instructions for the following two workflows:

  1. Wan 2.2 First Last Frame to Video (FLF2V) – Fast LoRA version
  2. Wan 2.2 First Last Frame to Video (FLF2V) – High quality version

Software needed

We will use ComfyUI, a free AI image and video generator. You can use it on Windows, Mac, or Google Colab

Think Diffusion provides an online ComfyUI service. They offer an extra 20% credit to our readers.

Read the ComfyUI beginner’s guide if you are new to ComfyUI. See the Quick Start Guide if you are new to AI images and videos.

Take the ComfyUI course to learn how to use ComfyUI step by step.

ComfyUI Colab Notebook

If you use my ComfyUI Colab notebook, you don’t need to download the model as instructed below. Select the Wan_2_2 model before running the notebook.

Wan 2.2 First Last Frame workflow (Fast 4-step LoRA)

This workflow takes two images as input and produces a video with the images as the first and last frames.

This workflow takes only 3 minutes on an RTX 4090 GPU card, thanks to the Wan 2.2 Lightning LoRA.

Step 0: Update ComfyUI

Before loading the workflow, make sure your ComfyUI is up to date. The easiest way to do this is to use ComfyUI Manager.

Click the Manager button on the top toolbar.

Select Update ComfyUI.

Restart ComfyUI.

Step 1: Load the workflow

Download the workflow below.

Step 2: Download models

After loading the workflow JSON file, ComfyUI should prompt you to download the missing model files.

Here are the models you need to download:

  1. Download wan2.2_i2v_high_noise_14B_fp8_scaled.safetensors and put it in ComfyUI > models > diffusion_models.
  2. Download wan2.2_i2v_low_noise_14B_fp8_scaled.safetensors and put it in ComfyUI > models > diffusion_models.
  3. Download umt5_xxl_fp8_e4m3fn_scaled.safetensors and put it in ComfyUI > models > text_encoders.
  4. Download wan_2.1_vae.safetensors and put it in ComfyUI > models > vae.
  5. Download wan2.2_i2v_lightx2v_4steps_lora_v1_low_noise.safetensors and put it in ComfyUI > models > loras.
  6. Download wan2.2_i2v_lightx2v_4steps_lora_v1_high_noise.safetensors and put it in ComfyUI > models > loras.

Step 3: Upload the start and end frame images

Upload the images you wish to use as the first and last frames.

You can use the test image below.

First Frame
Last Frame

Step 4: Revise the prompt

Describe your video in the prompt. Use keywords to direct the camera.

Step 5: Set video size and length

Adjust the width and height according to the aspect ratio of your input images.

The length is the number of frames of the video.

Step 6: Generate the video

Click the Run button to run the workflow.

Wan 2.2 First Last Frame workflow (High quality)

This workflow is functionally the same as the last one: It takes two images as input and produces a video with the images as the first and last frames.

Except that it doesn’t speed up the workflow using the Lightning LoRA. It takes longer to run, but the quality is higher (at least theoretically).

This workflow takes 26 minutes on an RTX 4090 GPU card.

Step 0: Update ComfyUI

Before loading the workflow, make sure your ComfyUI is up to date. The easiest way to do this is to use ComfyUI Manager.

Click the Manager button on the top toolbar.

Select Update ComfyUI.

Restart ComfyUI.

Step 1: Load the workflow

Download the workflow below.

Step 2: Download models

Skip this step if you have installed models in the first workflow.

After loading the workflow JSON file, ComfyUI should prompt you to download the missing model files.

Here are the models you need to download:

  1. Download wan2.2_i2v_high_noise_14B_fp8_scaled.safetensors and put it in ComfyUI > models > diffusion_models.
  2. Download wan2.2_i2v_low_noise_14B_fp8_scaled.safetensors and put it in ComfyUI > models > diffusion_models.
  3. Download umt5_xxl_fp8_e4m3fn_scaled.safetensors and put it in ComfyUI > models > text_encoders.
  4. Download wan_2.1_vae.safetensors and put it in ComfyUI > models > vae.

Step 3: Upload the start and end frame images

Upload the images you wish to use as the first and last frames.

You can use the test image below.

First Frame
Last Frame

Step 4: Revise the prompt

Describe your video in the prompt. Use keywords to direct the camera.

Step 5: Set video size and length

Adjust the width and height according to the aspect ratio of your input images.

The length is the number of frames of the video.

Step 6: Generate the video

Click the Run button to run the workflow.

Useful links

Andrew

By Andrew

Andrew is an experienced software engineer with a specialization in Machine Learning and Artificial Intelligence. He is passionate about programming, art, and education. He has a doctorate degree in engineering.

Leave a comment

Your email address will not be published. Required fields are marked *