Clone Your Voice Using AI (ComfyUI)

Have you ever wondered how those deepfakes of celebrities like Mr. Beast were able to clone their voices? Well, those people use voice cloners like the F5-TTS model in order to do this, and in this article, you’ll learn that it’s actually not that hard! How it works: Table of ContentsSoftwareStep-by-step guideStep 0: Update ComfyUIStep…

How to run Wan VACE video-to-video in ComfyUI

WAN 2.1 VACE (Video All-in-One Creation and Editing) is a video generation and editing AI model that you can run locally on your computer. It unifies text-to-video, reference-to-video (reference-guided generation), video-to-video (pose and depth control), inpainting, and outpainting under a single framework. VACE supports the following core functions: This tutorial covers the Wan Vace Video-to-Video (V2V)…

Wan VACE ComfyUI reference-to-video tutorial

WAN 2.1 VACE (Video All-in-One Creation and Editing) is a video generation and editing model developed by the Alibaba team. It unifies text-to-video, reference-to-video (reference-guided generation), video-to-video (pose and depth control), inpainting, and outpainting under a single framework. VACE supports the following core functions: You can use the WAN VACE model in ComfyUI with the…

How to run LTX Video 13B on ComfyUI (image-to-video)

LTX Video is a popular local AI model known for its generation speed and low VRAM usage. The LTXV-13B model has 13 billion parameters, a 6-fold increase over the previous 2B model. This translates to better details, prompt adherence, and more coherent videos. In this tutorial, I will show you how to install and run…

Flux-Wan 2.1 four-clip movie (ComfyUI)

This workflow generates four video clips and combines them into a single video. To improve the quality and control of each clip, the initial frame is generated with the Flux AI image model, followed by Wan 2.1 Video with Teacache speed up. You can run it locally or using a ComfyUI service. You must be…

Stylize photos with ChatGPT

Do you know you can use ChatGPT to stylize photos? This free, straightforward method yields impressive results. In this tutorial, I will show you how to convert an image into different styles like these: Table of ContentsHow does it workAlternativesStep-by-step guideStep 1: Upload a photoStep 2: Enter a promptStep 3: Start image conversionOther style promptsStudio…

How to generate OmniHuman-1 lip sync video

Lip sync is notoriously tricky to get right with AI because we naturally talk with body movement. OmniHuman-1 is a human video generation model that can generate lip sync videos from a single image and an audio clip. The motion is highly realistic and matches the voice. OmniHuman-1 is currently only available through an online…

How to create FramePack videos on Google Colab

FramePack is a video generation method that allows you to create long AI videos with limited VRAM. If you don’t have a decent Nvidia GPU card, you can use FramePack on the Google Colab online service. It’s a cost-effective option, costing only around $0.20 per hour to use. Table of ContentsWhat is FramePack?What is Google…