Mechanical insect video (ComfyUI)

Published Categorized as Workflow Tagged , , , , No Comments on Mechanical insect video (ComfyUI)

This workflow generates beautiful videos of mechanical insects from text prompts. You can run it locally or using a ComfyUI service. It uses Flux AI to generate a high-quality image, followed by Wan 2.1 Video for animation with Teacache speed up.

You must be a member of this site to download the following ComfyUI workflow.

Software

We will use ComfyUI, an alternative to AUTOMATIC1111. You can use it on Windows, Mac, or Google Colab. If you prefer using a ComfyUI service, Think Diffusion offers our readers an extra 20% credit.

Read the ComfyUI beginner’s guide if you are new to ComfyUI. See the Quick Start Guide if you are new to AI images and videos.

Take the ComfyUI course to learn how to use ComfyUI step by step.

How does this workflow work?

The Flux text to image block generates a static image using the Flux.1 Dev model.

The Wan 2.1 block converts the static image to a video using the same prompt.

Step-by-step guide

Step 1: Load workflow

Download the ComfyUI JSON workflow below.

Become a member of this site to see this content

Already a member? Log in here.

Drag and drop the JSON file to ComfyUI.

Step 2: Install missing nodes

If you see nodes with red borders, you don’t have the custom nodes required for this workflow. You should have ComfyUI Manager installed before performing this step.

Click Manager > Install Missing Custom Nodes.

Install the nodes that are missing.

Restart ComfyUI.

Refresh the ComfyUI page.

Step 3: Download models

Flux

Download the flux1.Dev model flux1-dev-fp8.safetensors. Put it in ComfyUI > models > checkpoints.

Wan 2.1

Download the diffusion model wan2.1_i2v_720p_14B_fp8_e4m3fn.safetensors and put it in ComfyUI > models > diffusion_models.

Download the text encoder model umt5_xxl_fp8_e4m3fn_scaled.safetensors and put it in ComfyUI > models > text_encoders.

Download the CLIP vision model clip_vision_h.safetensors and put it in ComfyUI > models > clip_vision.

Download the Wan VAE model wan_2.1_vae.safetensors and put it in ComfyUI > models > vae.

Google Colab

If you use my ComfyUI Colab notebook, you don’t need to download the model. Select

  • the Flux1_dev model
  • WAN_2_1 video models
  • WAN_2_1 custom nodes
  • VideoHelperSuite custom nodes

Step 4: Revise the prompt

Revise the prompt above the groups.

Step 5: Run the workflow

Click the Run button to run the workflow.

2-step workflow

You can use the Fast Groups Muter node to control the video generation process more. You can first focus on the initial image generation. Once you are happy with it, go ahead and turn on the video generation.

Generate image only

Turn off the Wan 2.1 group in the Fast Groups Muter.

Click the Run button to run the workflow.

You should have run the Flux text-to-image group and seen the preview image.

Revise the prompt and seed value until you are happy with the result.

Enable the video generation

Now, turn the Wan 2.1 group back on.

Click the Run button to run the workflow.

You now also get the video.

Change the seed value of the KSampler in the Wan 2.1 group to generate a new video.

Andrew

By Andrew

Andrew is an experienced software engineer with a specialization in Machine Learning and Artificial Intelligence. He is passionate about programming, art, and education. He has a doctorate degree in engineering.

Leave a comment

Your email address will not be published. Required fields are marked *