-
-
Notifications
You must be signed in to change notification settings - Fork 361
LTX 2 and Z Image Base Full Tutorial Audio to Video Lip Sync ComfyUI SwarmUI Windows Cloud
Full tutorial link > https://www.youtube.com/watch?v=SkXrYezeEDc
LTX 2 is the newest state of the art (SOTA) Open Source video generation model and tutorial will show you how to use it with very best and most performant way in ComfyUI and also in SwarmUI. Moreover, Z Image Base model published and I will show how to use Z Image Base with most amazing preset and workflow as well. Furthermore, this tutorial will show you how to install, update, setup, download ComfyUI and SwarmUI and models and presets and workflows both on Windows and on RunPod, Massed Compute and SimplePod. Linux users can use Massed Compute scripts and installers directly. This is a masterpiece entire lecture level complete tutorial. This video will kickstart your AI journey 100x. Both local Windows and Cloud.
📂 Resources & Links:
🤖 ComfyUI Installer and Presets Zip File With CUDA 13: [ https://www.patreon.com/posts/ComfyUI-Installers-105023709 ]
💻 SwarmUI Installer and Presets Zip File: [ https://www.patreon.com/posts/SwarmUI-Install-Presets-114517862 ]
🚀 Model Downloader Zip File: [ https://www.patreon.com/posts/Model-Downloader-114517862 ]
🚆 SECourses Musubi Trainer (Model Quantize and Train App): [ https://www.patreon.com/posts/SECourses-Musubi-Trainer-137551634 ]
🛠️ Image Comparison Slider Tool: [ https://www.patreon.com/posts/image-video-comparison-slider-app-133935178 ]
👋 SECourses Discord Channel for 7/24 Support: [ https://bit.ly/SECoursesDiscord ]
ℹ️ SimplePod Register : https://simplepod.ai/ref?user=secourses
ℹ️ SimplePod Template : https://dash.simplepod.ai/account/explore/100/ref-secourses/
ℹ️ RunPod Register : https://get.runpod.io/955rkuppqv4h
ℹ️ RunPod Template : https://get.runpod.io/SECourses_CU13
ℹ️ Massed Compute Register : https://bit.ly/SECoursesMassedCompute
⏱️ TIMESTAMPS
00:00:00 Intro: ComfyUI + SwarmUI presets, Z-Image, model downloader, cloud installs
00:00:28 Free prompt enhancement with Google AI Studio (prepared prompt file)
00:01:26 Demo: 45s audio-driven lip-sync image→video (LTX 2)
00:02:25 Quick demos: ComfyUI image→video, text→video, Z-Image base
00:03:43 Quick demos: SwarmUI LTX 2 image→video + Z-Image base
00:04:46 Install/update presets zip (v78+): extract & overwrite everything
00:05:11 Upgrade to CUDA 13 safely: delete venv then run ComfyUI update/install
00:06:00 Windows prereqs + the always-updated setup guide referenced in description
00:06:42 Install required node bundles: (1) SwarmUI extra nodes + (100) LTX audio
00:07:29 VRAM-optimized launcher: no-VRAM / cache / smart-memory / precision choices
00:09:36 Share one model library: configure extra_model_paths.yaml (no duplicates)
00:10:25 Model Downloader overview: set base path + one-click bundles for SwarmUI/ComfyUI
00:11:43 Download LTX 2 core bundle: multi-connection download + merge + hash verify
00:12:12 Low-VRAM path: GGUF distilled models vs recommended FP8-scaled defaults
00:14:16 URL Downloader: Civitai/HF links, folder targeting, and optional API keys
00:14:54 ComfyUI preset pack tour: LTX2 (I2V/T2V/audio) + Z-Image (base/2× upscale)
00:15:16 ComfyUI audio lip-sync preset: image + resolution + audio setup
00:16:58 Frames & prompting: 24fps math, run/stop frame count, lyrics/subtitles prompts
00:17:59 Quality/perf knobs: CRF, VRAM monitoring, and low-VRAM args recap
00:20:22 Review result + move to LTX 2 image→video preset workflow
00:21:50 Prompt enhancer workflow: drag prompt file into AI Studio (optionally add image)
00:24:11 Z-Image troubleshooting: disable Sage Attention and restart
00:24:32 Z-Image base + 2× upscale preset: when to use it and what to expect
00:26:47 Outputs & reproducibility: where renders save + drag PNG to reload metadata
00:27:57 Update SwarmUI via zip: get latest presets + utilities
00:28:27 SwarmUI setup: ComfyUI backend, passing args, and pointing to model folders
00:30:21 SwarmUI image→video: direct apply + init image workflow
00:31:17 Fix model load/caching issues: add --use-cache-none when needed
00:32:51 SwarmUI text→video + upscale: duration/frames, half-res then upscale rules
00:33:52 SwarmUI outputs: output_local/raw + metadata saved with generations
00:35:05 SwarmUI Z-Image: base vs 2× upscale comparison + speed notes
00:36:58 Image comparison slider tool: quick before/after inspection
00:37:51 RunPod start: template choice, CUDA/driver constraints, optional storage volume
00:43:30 RunPod Jupyter: upload ComfyUI zip, extract, install bundles (1+100)
00:46:28 RunPod: run Model Downloader, start ComfyUI, connect via exposed port
00:52:05 RunPod: switch to SwarmUI, map folders (case-sensitive), import presets
00:55:43 RunPod: download outputs as archive + stop vs terminate cost control
00:57:55 SimplePod: cheaper/faster alternative + persistent volume setup highlights
01:03:52 Massed Compute: deploy with coupon + connect via ThinLinc (shared folder setup)
01:08:17 Massed Compute: install bundles + download models (disk speed advantage)
01:10:48 Massed Compute: start ComfyUI, connect from PC, run lip-sync preset
01:14:26 Massed Compute: hook SwarmUI to ComfyUI backend + text→video demo
01:17:02 Wrap-up: recap providers + next steps
#comfyui #comfyuitutorial #runpod
-
00:00:00 Greetings everyone. Today I am going to show you LTX 2 workflows, presets in ComfyUI. I
-
00:00:07 have prepared all the configurations, one click to install and use. Moreover, I will
-
00:00:12 show Z-Image base model in ComfyUI. Moreover, I will show you how to use the LTX 2 in our famous
-
00:00:21 SwarmUI with presets, Z-Image base in SwarmUI with presets. Furthermore, I will show you how
-
00:00:28 to improve your prompts in Google AI Studio for free with our prepared commands file. Moreover,
-
00:00:36 downloading these preset models will be one click with our interface updated model
-
00:00:42 downloader. This is independent from both ComfyUI and SwarmUI. It has both Z-Image and LTX 2 models.
-
00:00:51 I also have prepared Z-Image Quant FP8 scaled model for you and also working on NVFP4 models
-
00:00:59 and NVFP4 generator. This application will have everything SECourses Musubi Trainer. It will
-
00:01:06 have all the options to quantize your models. Also, I will make fresh installation on RunPod,
-
00:01:13 Massed Compute, and SimplePod so that we will refresh our memory to how to install and use
-
00:01:20 ComfyUI and SwarmUI on cloud platforms in addition to your local Windows computer.
-
00:01:26 I am going to show you an amazing workflow in ComfyUI which you can provide your audio,
-
00:01:32 image, prompt and generate amazing audio driven
-
00:01:35 video. It has lip sync and it has animation. Let me demonstrate you.
-
00:01:40 SECourses. Let's go. Late night screen glow. Ideas on fire. Turning what if into something higher.
-
00:01:48 From zero to flow. We learn it fast. Build it ship it make the future last. New tools new
-
00:01:53 rules we break the doubt. Take that prompt watch the AI route. If you want the shortcuts done the
-
00:01:57 right way. Hit play let's level up today. Hands on the keyboard mind on the dream. Watch the
-
00:02:02 model run like a laser beam. No fear no fluff just clarity. One more step to your next victory. SECourses
-
00:02:08 AI on the rise. Lighting up the dark with electric skies. From prompts to pixels
-
00:02:13 code to sound. We make it real we run this ground. That's courses learn build repeat. Turn big ideas
-
00:02:19 into something elite. If you want the best place to start. That's courses straight to the heart.
-
00:02:25 As you have seen it was amazing video. It was 45 seconds. After like 30 seconds 35 seconds
-
00:02:32 it starts to degrade but this is what LTX 2 can do. This will be literally one click to
-
00:02:39 install and use. I will show that. So here another example in ComfyUI. This is image
-
00:02:43 to video workflow. It is also one click to use and I will show in SwarmUI as well. Let's see.
-
00:02:50 This is the tutorial that you have been waiting for so long.
-
00:02:53 I mean amazing. You see it did a very nice emotion at the end. The LTX video
-
00:02:59 can generate up to 20 seconds very good. It can generate up to 1 minute
-
00:03:03 but after 20 seconds it starts to degrade. Let's see text to video.
-
00:03:13 You took a wrong turn partner.
-
00:03:18 By the way these were quick demos that I did. They are not perfect. They are not
-
00:03:22 cherry picked. And this is Z-Image workflow. This is an amazing quality. So many people
-
00:03:29 is not using accurately and as you see this is what kind of quality Z-Image base can generate
-
00:03:35 in ComfyUI. All is preset I will show. Now let me show you the examples on the SwarmUI. This
-
00:03:43 is image to video in SwarmUI. It is so easy to use with our preset. I will show how to use.
-
00:03:49 This is the tutorial that you have been waiting for so long.
-
00:03:58 As you have seen it is amazing. These are quick demos so you can
-
00:04:02 generate even much better videos. So here another example in SwarmUI. This is video.
-
00:04:12 Faith falls! Witness me!
-
00:04:17 With video upscalers they would become much better or with more tries or improving the text prompt.
-
00:04:24 These are quick demos. And we have the Z-Image base preset in SwarmUI as well. It is working
-
00:04:30 amazing. Amazing quality, amazing realism and it is following the prompts very well very good.
-
00:04:36 So you should begin with upgrading your ComfyUI if you haven't yet or download the latest zip
-
00:04:41 file to obtain the all the presets that we have. So the link will be in the description
-
00:04:46 of the video. Download the latest zip file version 78. It can be bigger version when
-
00:04:51 you are watching this tutorial. Then copy the downloaded zip file. Go to your ComfyUI
-
00:04:56 installation or into a fresh folder. Both works. Then you need to extract archive and overwrite all
-
00:05:04 the files. This is really important. Overwrite all. Then if you didn't upgrade to CUDA 13 yet,
-
00:05:11 enter inside ComfyUI, delete your virtual environment folder like this. This is proper
-
00:05:17 way of upgrading CUDA 13 if you haven't yet. Since my ComfyUI is open it is preventing me
-
00:05:23 to delete the virtual environment folder. So let me delete all of them. Try again. Deleting
-
00:05:29 virtual environment will not cause you to lose any data. Then install or update ComfyUI run.
-
00:05:36 This is not mandatory every time. This is mandatory when you are upgrading from CUDA
-
00:05:40 12.9. So I am using Python 3.10. You need to have this installed in your computer. And this is the
-
00:05:47 version that I recommend. This is same for fresh installation and upgrade. We use the same file.
-
00:05:54 If you are first time into generative AI you really should follow Windows requirements
-
00:06:00 tutorial. When you scroll down a little bit you will see this. We have a up to date video
-
00:06:06 here so watch it and we have an up to date post which you will follow while you are
-
00:06:12 watching this tutorial video. It has all the links instructions. It is fully up to date.
-
00:06:17 Last time updated is 8 January 2026. I am using exactly as in here so this is
-
00:06:26 identical to my setup so therefore when you watch this tutorial all of the AI
-
00:06:31 applications will work on your PC perfectly fine on your computer perfectly fine.
-
00:06:36 So the ComfyUI installation and update has been completed. To use the newest presets you need to
-
00:06:42 use the Windows custom nodes bundles installer. Double click it. This is a unified bundles
-
00:06:49 installer that I have made for you. So you need to install option 1 SwarmUI extra nodes to use
-
00:06:55 the presets. Plus you need to install bundle 100 LTX audio to video bundle. So when you install
-
00:07:02 these two you will be ready. So 1 and comma and 100 and hit enter. It will ask you to proceed.
-
00:07:09 You can hit Y or enter and it will install the necessary bundles and update them. So this is for
-
00:07:16 both update and installation. This covers both of them. You see it is updating my bundles as
-
00:07:22 well and it is all getting updated. Okay all the nodes have been installed or updated properly.
-
00:07:29 Then I also have added windows_run_vram_optimized.bat
-
00:07:34 file. When you run this it will ask you to start your installation with specific configuration.
-
00:07:40 So if you are getting out of VRAM options you can do this. You can start with no VRAM
-
00:07:46 mode. When you do this you can even generate 1 minute videos on LTX 2. This will use very very
-
00:07:52 minimal amount of VRAM memory. Actually let me demonstrate you. So you see currently I am using
-
00:07:57 2 megabytes of VRAM. So I will choose the option 5. And if you are also low on RAM memory you can
-
00:08:04 start with no cache. So it is option 3. You can also disable smart memory. So let's disable as
-
00:08:12 well. And it will be GPU default or you can also choose VAE on CPU. I will make it GPU.
-
00:08:19 I am going to use Sage Attention. You shouldn't use Sage Attention for Z-Image base model at the
-
00:08:24 moment but for all other models you should use it. Currently Z-Image base model is not
-
00:08:29 working with Sage Attention. And the precision I am going to just make it auto. And for this
-
00:08:34 option just hit yes. Additional optimizations I am just going to leave it auto. Disable pinned
-
00:08:41 memory no. And it is all done. So you see it shows you the configuration options. You can
-
00:08:48 do this if you don't want to pick these every time. Copy this windows run GPU and paste. So
-
00:08:55 you see I have copied version. Right click edit in Notepad++. So you can append these options to here
-
00:09:04 like this. You see or like this. So this way you can customize your starting bat file and use it
-
00:09:13 to start directly with all these optimizations you want. This is also how you remove the sage
-
00:09:21 attention from run GPU here. So you can also edit this file and remove the Sage Attention
-
00:09:30 from here to start without Sage Attention. This is how you customize your starting with ComfyUI.
-
00:09:36 Moreover to use the SwarmUI downloaded models you need to do this. Copy extra model paths into your
-
00:09:43 ComfyUI folder. Then right click and edit it. Then select the models base path. This is where
-
00:09:51 I have downloaded all of my models so therefore when I give this it sees all the models inside
-
00:09:57 my SwarmUI installation folder or wherever I want. You can add multiple base paths to here.
-
00:10:04 Okay so the ComfyUI started with the minimal amount of VRAM memory usage. So let me show you
-
00:10:11 45 seconds audio to video generation. I have the preset here. How you are going to use them? You
-
00:10:18 need to first download the models. To download the models we are going to use the model downloader
-
00:10:25 which is included in SwarmUI model downloader zip file. This is also how we are going to
-
00:10:29 install SwarmUI updated and get the presets. So let's download this. The link will be in
-
00:10:34 the description of the video. So move the SwarmUI model downloader into your SwarmUI installation
-
00:10:40 or into any drive make a fresh installation. Right click. Then you need to extract here and
-
00:10:46 overwrite all the files. This is super important. Overwrite all the files so that it will update the
-
00:10:53 utilities and everything. Then double click Windows start download models application
-
00:10:57 file. Independent from SwarmUI. You don't need to install SwarmUI for this. This supports ComfyUI
-
00:11:05 as well. So this is our new interface. This is where you give the base model path. So you see
-
00:11:11 currently it is seeing automatically my SwarmUI models. You can change this into your ComfyUI
-
00:11:17 models as well. So enter inside here. Go to the models and copy the path. Copy paste it
-
00:11:24 here. You can also remember settings. And ComfyUI folder structure. Then you need to download the
-
00:11:30 LTX 2 video core bundle or Z-Image core bundle if you want to use them. But I am using SwarmUI
-
00:11:38 mainly therefore I am downloading models there. So refresh and it will return back to original
-
00:11:43 since I didn't remember settings. Then I need to click download all models of the LTX 2 core
-
00:11:51 bundle. Then it will start downloading with 16 connections. You can also drag and drop this
-
00:11:56 like this to expand or make it small or click here. And it will start the downloads. It will
-
00:12:01 do hash verification. It will download all the models that I have prepared for you. You can
-
00:12:06 also see their names here. If you have a very low VRAM GPU alternatively go to the video generation
-
00:12:12 models. Go to the LTX 2 video models and you can download the LTX 2 distilled GGUF model.
-
00:12:22 Which is here. So you can download GGUF Q4 model if you are on very low RAM and VRAM. This is not
-
00:12:30 recommended. Normally I use FP8 scaled. This is also what the bundle downloads. And additionally
-
00:12:35 you can also download FP4 mixed text encoder but by default I am using this one. This is mixed FP8
-
00:12:43 scaled. The name is like this because SwarmUI uses this automatically and my presets are also
-
00:12:49 set for this so don't worry the presets will be one click. So follow the download process
-
00:12:54 here. It will download all the original models as you are seeing very fast. 100 megabytes per
-
00:13:01 second on my computer. Then it will verify them then you will be ready to start using them.
-
00:13:06 Moreover since I am using extra model paths.yaml file inside ComfyUI when I start my ComfyUI you
-
00:13:12 see it adds all the models that I have given it. So it sees all the models that I have in SwarmUI
-
00:13:20 in ComfyUI. Therefore I am avoiding all the model duplication. And I am using central model folder.
-
00:13:28 You can also modify this file according to your needs. Add more paths or more mapping.
-
00:13:34 It works. SwarmUI also has this feature. You can give another base model path and it see
-
00:13:40 all of them. So you can have all of your models inside ComfyUI and also use in SwarmUI as well.
-
00:13:46 Since model downloader downloads with 16 connections it will merge all the downloaded
-
00:13:52 pieces into single file at the end. So you will see like this it is merging then it will verify
-
00:13:59 the hash. So now it is verifying its hash value. This is also working on cloud and Linux machines.
-
00:14:05 This way we are ensuring that we have accurately downloaded the models. Therefore we will never
-
00:14:10 have any corrupted model at all. Moreover this downloader application supports custom model
-
00:14:16 downloads as well. So let's hide this. Go to the URL downloader. In here you just need to give the
-
00:14:23 model path. It can be CivitAI. It can be Hugging Face. And you can select the folder where it will
-
00:14:29 download so that you can download your custom models this way on cloud machines or on your
-
00:14:35 Windows computer very fast. You can also provide your API keys to download even faster. This is
-
00:14:42 very very useful especially on RunPod on Massed Compute to download fast into the accurate folder.
-
00:14:48 Okay so all the models have been downloaded. Let's return back to our ComfyUI. Let's refresh.
-
00:14:54 Then inside ComfyUI extracted zip file you will see there are presets. So the newest presets are
-
00:15:02 LTX 2 image to video, text to video, audio lip sync and Z-Image base and Z-Image base with 2x
-
00:15:09 upscale. For Z-Image base don't forget you need to download same way Z-Image core bundle. Okay first
-
00:15:16 I will show you LTX 2 audio to lip sync image to video. Drag and drop it. So it is loaded. You
-
00:15:24 shouldn't have any red or warnings here if you have followed the steps that I have just shown
-
00:15:29 you. This is extremely high quality and fully optimized to work on low VRAM GPUs and systems
-
00:15:36 as well. So first of all you need to pick your image. Let's choose file to upload. I am going
-
00:15:43 to use this anime image. Then you need to set your width and height accurately. This is 9
-
00:15:51 to 16 therefore I am going to use 720 to 1280. If you are not sure how to set your width and
-
00:16:00 height accurately I recommend you to use SwarmUI. Unfortunately this preset doesn't exist in SwarmUI
-
00:16:08 because it doesn't support it. It uses a lot of custom nodes. Then quick tools reset params
-
00:16:14 to default. I am going to use my LTX 2 image to video. Then you can see that it will show you the
-
00:16:21 resolution like this and init image. So choose the same file in here. Resolution use closest
-
00:16:27 aspect ratio. You see this image is 720 to 1280 pixel for this model. Okay it is set. Then you
-
00:16:35 need to set your audio file which will be driving audio. I am going to use this 45 seconds audio.
-
00:16:44 SECourses. Let's go.
-
00:16:48 Okay this way you can verify your audio is working. So you see there is audio length in
-
00:16:53 seconds. So first set your audio length. You can also set audio start index here.
-
00:16:58 Then when you click this run and stop immediately you will see that it is
-
00:17:04 showing you how many frames you need here. So I need 1081 frames for this video. After that
-
00:17:11 you need to write your prompt. When you are using audio this is how you should prompt it.
-
00:17:19 Describe your audio in details plus write the full spoken or singing like this. You see this is the
-
00:17:30 lyrics of my audio. If it was a speaking it would be the subtitles transcription of the speaking.
-
00:17:38 So this is the way of it. Then I have after few tries I have come up with this prompt. You can
-
00:17:44 pause the video and read it. And like this. So this is my prompt like this and run. Then all
-
00:17:53 you need to do is just wait. You don't need to change anything else. You can also change your
-
00:17:59 final video quality from here. You see CRF this is the video quality. If you make this very low
-
00:18:05 it will be very big generated MP4 file. Let's also monitor our VRAM usage and the speed. Currently we
-
00:18:14 are running it on no VRAM. Therefore it will use minimal absolutely minimal amount of VRAM. So we
-
00:18:22 can see both of the screens like this. It will be therefore slow but it will work on every GPU.
-
00:18:30 Okay once the generation starts. So even if you have a 6 gigabyte GPU 4 gigabyte GPU you can use
-
00:18:37 all these presets with the way I have shown you in the beginning of the video how to start your
-
00:18:44 ComfyUI. This applies to SwarmUI as well. In the SwarmUI you set the same arguments into
-
00:18:51 your ComfyUI backend. And it is also using my GPU very decent. I mean look at that. It
-
00:18:57 is using like 500 watts. This is really really good. It is using my GPU very well. And it is
-
00:19:02 only using 5000 megabytes VRAM right now. So this way you can generate very long videos on very low
-
00:19:10 VRAM GPUs. And you can use the same arguments in your backend in your SwarmUI. It will work
-
00:19:16 exactly same. Remember when we started the ComfyUI it has shown what parameters what
-
00:19:24 arguments have been used like no VRAM cache none disable smart memory. Use Sage Attention disable
-
00:19:31 asynchronous offload. This way I have minimized my VRAM usage. So it is only using 5 gigabytes
-
00:19:39 of VRAM for 1081 frames at HD resolution. And it is generating it right now. This workflow first
-
00:19:49 do 8 steps for initial video generation then it does 3 steps of the upscale. The upscale will
-
00:19:56 be of course slower. And during the upscale it uses more VRAM as expected since it is
-
00:20:03 now bigger resolution. So you can reduce your length if you get out of VRAM error or on your
-
00:20:09 computer it may do even further optimization since I have VRAM maybe it didn't do I am not
-
00:20:15 sure. And the time is still very decent. I mean very acceptable time even at these settings.
-
00:20:24 Let's see it.
-
00:20:27 SECourses. Let's go. Late night screen glow. Ideas on fire. Turning what if into something higher.
-
00:20:34 From zero to flow. We learn it fast. Build it ship it make the future last. New tools new
-
00:20:39 rules we break the doubt. Take that prompt watch the AI route. If you want the shortcuts done the
-
00:20:43 right way. Hit play let's level up today. Hands on the keyboard mind on the dream. Watch the
-
00:20:49 model run like a laser beam. No fear no fluff just clarity. One more step to your next victory. SECourses
-
00:20:54 AI on the rise. Lighting up the dark with electric skies. From prompts to pixels
-
00:20:59 code to sound. We make it real we run this ground. That's courses learn build repeat. Turn big ideas
-
00:21:06 into something elite. If you want the best place to start. That's courses straight to the heart.
-
00:21:11 Purely amazing. This time there weren't any noise. There weren't any text. There weren't
-
00:21:16 any issues. This was even better than my initial demo that I have shown you. So what about using
-
00:21:22 the presets? They will be surely faster and they will be using lesser VRAM since we will be doing
-
00:21:29 lesser amount of frames. So let's begin with the image to video. Drag and drop it. You can
-
00:21:35 also open. So this is much more straightforward. Choose and upload your image. Same way like this.
-
00:21:43 Set your video height and width. I am going to use the same like this. Type your prompt. And
-
00:21:50 how you can make your prompt better? Now let me show you. In the SwarmUI zip file you will
-
00:21:56 see that there is LTX 2 enhance prompt feed for LLMs. When you open it you will see its content
-
00:22:04 like this. So to use this I am going to use Google AI Studio. Then drag and drop the file like this.
-
00:22:11 Enter your prompt and if you are also using an image add your image as well. This will help
-
00:22:18 it to also modify prompt according to the image. And run. If you are not using image just text to
-
00:22:25 video just type like this and it will enhance it automatically for you. The logic is same. Okay it
-
00:22:32 did enhance it. By the way using Google AI Studio is free. Therefore just copy paste it and run.
-
00:22:39 You can also change the duration from frame count. This is 24 FPS model therefore when
-
00:22:45 you divide this 24 it makes 5 second. So 5 multiplied with 24 plus 1 equal to our frame
-
00:22:54 count. If I want it 10 seconds then I need to make this 241 frames. So let's see the VRAM usage. It
-
00:23:03 will use very minimal amount of VRAM since we are only generating 121 frames. Okay let's see
-
00:23:11 how much VRAM it will use. It is also using very minimal amount of RAM memory too. During
-
00:23:16 the upscale it will use more GPU obviously. And the speed is also amazing even though I
-
00:23:22 am running it with no VRAM command argument. Okay with the upscale it is using 2.7. It is
-
00:23:32 using 3 gigabytes of VRAM during the upscale. This is amazing. I mean you will be able to
-
00:23:37 generate 5 second HD videos with only 3 gigabytes of VRAM memory. Can you believe that? Okay at the
-
00:23:45 end it used some more for the VAE tiling but it is fine. And we have the video. So let's see.
-
00:23:53 Hello everyone welcome.
-
00:23:55 It did some animation maybe changing the seed will help or I need to change the prompt but it
-
00:24:02 is working great so play with it to see. Then I will restart my ComfyUI with proper starting
-
00:24:11 and I will show Z-Image but for Z-Image I need to remove the Sage Attention because currently
-
00:24:18 Z-Image base is not working. The text to video is also same I am not going to show it. So let's
-
00:24:24 run the application with full optimizations with full speed and without Sage Attention. Okay the
-
00:24:32 application started. For Z-Image we are going to use the Z-Image with 2x upscale. I recommend this.
-
00:24:39 This is working really amazing. Amazing quality. Then type your prompt as you wish you know this
-
00:24:46 is just regular prompting. Let's use this prompt to see what we are going to get. Okay run. I am
-
00:24:52 also working on Z-Image NVFP4 model a properly made NVFP4 model so it will be 2x faster with
-
00:25:01 almost same quality. Our SECourses Musubi Trainer application is going to have all these convert to
-
00:25:09 quant options FP8 int8 and FP4 so that you will be also able to convert any model you
-
00:25:16 want into FP8 scaled quantized FP8 it will be amazing quality like original BF16 or NVFP4. It
-
00:25:24 will be 100 percentage faster on Blackwell GPUs. All of them. 5060 5070 5080 whatever 50 or newer
-
00:25:34 versions you get in future. So this is going to generate 4 megapixel resolution image. This has
-
00:25:41 the best upscale. If you don't want to upscale you just need to use the Z-Image base preset.
-
00:25:48 Okay let's see our generated image. By default I am doing 40 steps but you can reduce the step
-
00:25:55 count to 20 as well. With LTX 2 the step count is not changing. It is 8 and in the
-
00:26:02 upscale it is 3. But with the Z-Image base or with majority of the other models you can
-
00:26:08 change the steps count according to your need the speed. Currently it is using 18 gigabytes
-
00:26:14 of VRAM but with optimizations you can run this as low as 6 gigabyte GPUs. Since we are doing 4
-
00:26:21 megapixel image it is taking some time but we will get an amazing quality image. By the way
-
00:26:26 I don't know how will this video prompt work but you know the drill you can change the prompt as
-
00:26:32 you wish. Okay it generated image let's open in the new tab. So open image like this. Yeah for
-
00:26:41 this prompt this is a very decent output. All the generated files will be saved inside your
-
00:26:47 ComfyUI into your output folder. You see they are all saved. Even the videos are saved. To
-
00:26:54 remember the videos you need to do this. You see there is PNG file drag and drop it and it will
-
00:26:59 load everything from that PNG with the sets. So with the videos it saves a PNG file to save the
-
00:27:06 metadata. And the images are also saved like this with the ComfyUI extension. If you want to save
-
00:27:12 images with another prefix you can change the file name prefix from here. Or with the Z-Image
-
00:27:18 you can change the save prefix from here. You see ComfyUI you can change it to anything you want.
-
00:27:24 So this is all for ComfyUI for today. All other presets are working exactly same logic. You can
-
00:27:30 use all of our amazing presets that is coming with our zip file. You see all of them is working
-
00:27:37 tested working perfect. You just need to download the bundles. Like Z-Image, Flux 2, Wan, Qwen. We
-
00:27:45 have bundles for every major model out there. And for SwarmUI I am going to use the same ComfyUI
-
00:27:52 installation the backend. So let's just close our ComfyUI. So as usual download the latest SwarmUI
-
00:27:57 model downloader zip file. Overwrite your previous files. Then you should update your SwarmUI Windows
-
00:28:04 update SwarmUI. This will install necessary newer stuff and also update your SwarmUI to the latest
-
00:28:12 version. I have developed an extension for SwarmUI to support LTX 2 image to video upscale. Normally
-
00:28:20 SwarmUI is not supporting it but we are now supporting it. You can see that my extension is
-
00:28:27 compiled like this when you start it. Then go to the server backends and enter your backend as in
-
00:28:34 the same as other videos. You see ComfyUI backend. How you give the ComfyUI backend it is so easy. Go
-
00:28:41 to your ComfyUI installation ComfyUI folder copy it. Edit the here. Actually let me add a new one
-
00:28:50 to show you. So let's refresh. And ComfyUI self starting and edit it. Copy paste main.py. Extra
-
00:28:59 arguments. This is where you define your ComfyUI starting arguments like --use-sage-attention like
-
00:29:06 --cache-none like --no-vram. This way you can customize and optimize your ComfyUI backend.
-
00:29:14 Normally I just use Sage Attention. The only case that I don't use at the moment is Z-Image
-
00:29:20 base model because it degrades the quality significantly. It is broken at the moment. So
-
00:29:25 wait for your backend to start and it is started. Quick tools reset params to default. You need to
-
00:29:31 get the latest presets. So either import from here or my recommended way is Windows preset
-
00:29:38 delete import. So run it and hit yes and it will delete all of your previous presets and refresh
-
00:29:44 them with the new one. So refresh sort by name. All the newer presets are arrived like Z-Image
-
00:29:50 base or LTX 2 text to video LTX 2 upscale. When you use upscale you need to reduce your resolution
-
00:29:57 to half. So let's make a demo of image to video. Quick tools reset params to default. Click here
-
00:30:02 and direct apply. Since I have downloaded the LTX 2 video core bundle all the models
-
00:30:09 are automatically downloaded and selected. I have spent huge time to prepare these presets.
-
00:30:14 Then to use the image to video you need to choose your init image like this one. Then you need to
-
00:30:21 go to the image to video tab and set your video frame count. It is 121 frames right now. So it
-
00:30:28 will be 5 second video. Then enter your prompt same as in ComfyUI and you can generate as many
-
00:30:36 as videos you want. Let's generate one. So this will generate an image to video. You can always
-
00:30:41 follow the status from server logs debug menu to see what is happening. Let's also write another
-
00:30:48 prompt with an enhancement. So I am just going to use the Google AI Studio Gemini 3 Pro preview and
-
00:30:56 from SwarmUI we have LTX 2 enhance type my prompt. So this is my prompt. Let's see the enhancement.
-
00:31:05 Okay you see it had an error. Sometimes you may get this because of how the ComfyUI working just
-
00:31:11 hit generate again. This is happening because when you first time run it for some reason it
-
00:31:17 fails to properly load model. To prevent it you need to add --use-cache-none. So when you first
-
00:31:25 time click generate if you get error just wait and click generate again. This happens with the
-
00:31:32 LTX 2 video they didn't fix it yet. I am expecting fix hopefully soon. Don't worry just hit generate
-
00:31:39 again. You can also add the optimizations like disable smart memory or like cache none or low
-
00:31:49 VRAM. So one of them will work for your case. Once the first generation is completed it will
-
00:31:56 keep working. Okay now it is generating image to video. It should be also pretty fast. Yes. Okay
-
00:32:02 you see the image is changed like this because of my prompt so I need to modify
-
00:32:09 my prompt. For example like this. A video of a talking man saying hello everyone. So if your
-
00:32:17 image changes dramatically like this because of your prompt. You need to make it in a way
-
00:32:24 that it will not change the initial image. Let's see the new prompt impact. You see the
-
00:32:30 first video was done in 55 seconds. Second one is done in 35 seconds. Let's see it.
-
00:32:37 Hello everyone welcome.
-
00:32:39 Yes perfect quality. Perfectly matching the initial image. There is no changes so it is
-
00:32:44 all about your prompt. Prompt enhance may break image to video however it works really good at
-
00:32:51 text to video. Now let's see the text to video. So quick tools reset params to default and we
-
00:32:57 have LTX 2 text to video direct apply. Let's see our prompt like this. So this is enhanced prompt.
-
00:33:05 Change the resolution or aspect ratio. You can also change the duration from text to video. The
-
00:33:12 image to video is here image to video you change the frame count there. For text to video it is
-
00:33:18 here. So let's make this 241 frames and that's it generate. You can also upscale with text to
-
00:33:27 video how for text to video upscale you need to custom and half resolution like 640 to 360 and
-
00:33:36 go to presets and just apply in additionally this LTX 2 upscale like direct apply and let's generate
-
00:33:44 another one. The first one will be original HD the second one will use the upscaler model of
-
00:33:52 the LTX 2. All the generations of the SwarmUI will be saved inside SwarmUI output local raw
-
00:34:02 and they are categorized by the days that you generated them. You see all my images videos
-
00:34:08 are all saved here with their metadata as well. Okay the first video generated let's watch it.
-
00:34:15 Faithfulls! Witness me!
-
00:34:24 Yeah for this prompt this is what it has generated. Now we have the upscaled version
-
00:34:29 let's watch it as well. So this way you can tune your prompt try different prompts
-
00:34:44 and see what you are getting you can also compare. This was a 10 second video and it
-
00:34:49 was only generated in 40 seconds. 40 seconds for 10 second video. Amazing. I mean this is
-
00:34:56 241 frames. This would take forever on Wan 2.2 model. It is only taking 40 seconds on LTX 2.
-
00:35:05 So what about Z-Image? For Z-Image I am going to go to backends and edit my backend and remove the
-
00:35:12 Sage Attention. Currently this is mandatory. Probably it will be fixed soon. Then let's
-
00:35:18 open another tab reset params to default and in the preset our base preset is this one. So
-
00:35:25 let's use the same prompt as in this case I don't know what it will generate. Then I am also going
-
00:35:33 to apply the upscale. Then with the same seed let's also use the upscale. So for upscale we
-
00:35:40 are going to use image upscale direct apply and generate another one. The first one will
-
00:35:45 be 1024 to 1024 the second image will be 2048 to 2048 with our image upscale preset. This image
-
00:35:54 upscale preset works on every image generation. Okay the first image is done it is here it is
-
00:36:01 like this. Let's see its upscaled version too. So this is what it has generated as a base. I
-
00:36:08 am doing 40 steps for highest quality but you can reduce this to 20 steps as well. There will
-
00:36:15 be minimal impact on quality and it will be twice faster. As I also said I am also working on NVFP4
-
00:36:24 version of the Z-Image base model therefore it will be 100 percentage faster and almost
-
00:36:30 same quality. Moreover we have FP8 scaled of the Z-Image base model so you can download
-
00:36:37 it with our downloader and also use it if you are low on VRAM like 6 gigabyte GPUs. However
-
00:36:44 this model is not very big. The FP8 scaled is only 5.6 gigabyte the BF16 is like only
-
00:36:51 12 gigabytes. Okay the upscaled image is here. So this is the base image this is the upscaled
-
00:36:58 image. Let's compare them with our comparison application video comparison slider. The link
-
00:37:04 will be also description of the video to install this. Start image comparison. Let's put both the
-
00:37:11 files and full screen. So on the left we see the original normal image and on the right
-
00:37:17 we see the upscaled version. I mean look at the quality difference. The upscaled version
-
00:37:22 is much better. Many times better. It has added amazing amount of details quality so I recommend
-
00:37:30 you to use upscaled preset of the Z-Image base model to get really good images. The
-
00:37:38 realism of the Z-Image is already amazing so you can type a realistic prompt to get
-
00:37:44 realistic image like in this case. So it is all about your prompting to get whatever you want.
-
00:37:51 Now I will show you how to use ComfyUI and SwarmUI on RunPod. It is exactly actually same as before.
-
00:37:58 So once you downloaded the zip file you will see that we have RunPod SimplePod ComfyUI
-
00:38:03 instructions. RunPod and SimplePod are working exactly same way. You just need to follow this
-
00:38:09 file. You see we have RunPod section. So please use this link to register I appreciate that. Once
-
00:38:16 you registered login if it doesn't auto login you. Then go to billing add some credits. Then
-
00:38:25 return back to our instructions file always use the template mentioned here. So currently
-
00:38:32 we need to use this template for CUDA 13. This is based on official template don't
-
00:38:37 worry it just have some tweaks for you. Now this is super important. If you are going to
-
00:38:44 use your permanent storage select your permanent storage from here. So everything you do will be
-
00:38:49 saved 100 percentage. If you are not going to use permanent storage so just don't select anything
-
00:38:55 here. What you need to select here is additional filters and CUDA version 12.9. Unfortunately they
-
00:39:04 still didn't add CUDA 13 here. They need to add and upgrade their Nvidia drivers but they still
-
00:39:11 didn't do. Therefore we are selecting 12.9. This should bring us machines that has Nvidia drivers
-
00:39:19 that can run CUDA 13. Then select NVME and 100 gigabytes. So when we apply these two presets
-
00:39:27 unfortunately there aren't many machines we can use yet with CUDA 13 ComfyUI. So these are the
-
00:39:34 machines available. I am going to use RTX Pro 6000. This is one of the best as you know. Then
-
00:39:39 you see it has selected our template. I am going to contact to RunPod team to upgrade their drivers
-
00:39:45 so that CUDA 13 will arrive and they will upgrade their Nvidia drivers so we will be able to use
-
00:39:52 more machines. And in the template you can edit and change your volume disk if you are using your
-
00:40:00 permanent network storage. It will use whatever the space you have there but this will make a new
-
00:40:07 storage. So making permanent storage is so easy go to storage tab here click new network volume
-
00:40:13 select the server according to your server you will be limited to that region only so
-
00:40:19 this will limit your GPU availability. You can also click from here to see which regions have
-
00:40:25 more GPUs of this for example this one. Type your storage name 600 gigabytes type like 600
-
00:40:32 gigabyte and create network volume. Then you will be able to use it but I am not going to
-
00:40:37 use right now so I am going to delete it to not waste any money. Okay let's delete it.
-
00:40:43 So then deploy on demand. Now you need to wait a little to start the machine. It will take a
-
00:40:51 while like 5 minutes depending on the region on the server maybe 2 minutes. Unfortunately RunPod
-
00:40:58 machines are slow. Therefore Massed Compute is much faster. Now we are also supporting
-
00:41:04 SimplePod so easy to use same way actually. It is also faster and more GPUs available there.
-
00:41:10 I started 3 identical machines to see which one is starting fastest. Once I verify the
-
00:41:17 starting fastest one I will delete the other ones. Unfortunately with RunPod this is like
-
00:41:23 this you don't have many options you can also pick another GPU to make it start faster for
-
00:41:31 example like H200. This is a beast machine. It will work even faster than RTX Pro 6000.
-
00:41:38 So if you don't want to wait too long you can pick your machine accordingly or you can use
-
00:41:43 permanent storage to avoid this. So all of them is failing to start right now. Okay this one started.
-
00:41:51 Yeah this is probably H200 let's see yes. You see the H200 has a much better connection and
-
00:41:58 disk speed but I just emailed the RunPod team I think they will upgrade their Nvidia drivers
-
00:42:05 therefore we will have more machines available. Of course always you can use SimplePod and Massed
-
00:42:13 Compute. They need to upgrade their Nvidia driver so that we can use any machine available. Yeah I
-
00:42:19 will continue with H200 for this tutorial because it is already starting so I will
-
00:42:24 disable other machines stop not waste any time any money start. This is a shame the size of
-
00:42:32 the template is 9 gigabytes therefore it should be instant but you see the RunPod is very very
-
00:42:38 slow nowadays. They are degrading in quality unfortunately. I hope they become better. So
-
00:42:44 we are giving them feedback you should also give them feedback email them contact their support
-
00:42:50 tell them to upgrade their Nvidia driver add more GPUs. Okay this should start fairly fast.
-
00:42:56 You can see the download process. It doesn't matter which GPU you select my installers and
-
00:43:02 template supports all of them. What matters is the Nvidia driver which we are going to see in
-
00:43:08 telemetry when it starts it will show the Nvidia driver here. It should be 575 plus.
-
00:43:15 I just emailed them to upgrade all into 580 plus so that we can use any machine hopefully they do.
-
00:43:23 Okay the machine started when I go to connect you see it shows connect from 3001 this is ComfyUI
-
00:43:30 port we didn't start it yet and JupyterLab. So click the JupyterLab interface. You see the driver
-
00:43:36 is when we go to telemetry we see the driver is 575. You can also try CUDA 12.8 but it may not
-
00:43:44 work. This is guaranteed to work with all cases 575 driver. Hopefully they will upgrade and we
-
00:43:52 will have it. Okay so the JupyterLab interface started I will just drag and drop my ComfyUI
-
00:43:58 zip file into workspace wait for upload to be completed. So you see it is uploading right now.
-
00:44:04 Then I will extract archive refresh. Yes all done. Now open the RunPod SimplePod ComfyUI instructions
-
00:44:13 copy this command open a new terminal paste and hit enter. Then it will ask you which options to
-
00:44:20 install same as Windows 1 and 100 we are going to install these two bundles and hit enter then
-
00:44:26 hit enter. Then it will install the ComfyUI fully properly for us. Now you need to download models.
-
00:44:33 For downloading models we are going to use SwarmUI model downloader so it is here just drag and drop
-
00:44:38 it here refresh. It is still uploading wait for upload to be completed. This installation
-
00:44:45 is extremely optimized and fast but it depends on your machine as always. By the way you should
-
00:44:52 also verify your GPU open a new terminal pip install nvitop then type nvitop. And verify
-
00:45:01 that GPU is there and it is empty. Sometimes GPUs are being stuck or frozen or something.
-
00:45:08 Okay the SwarmUI installer also downloaded. So I will also install SwarmUI exactly same as
-
00:45:16 in Windows tutorial part. Okay so for SwarmUI installation we have RunPod SimplePod SwarmUI
-
00:45:24 install instructions. Just copy this entire command open a new terminal and paste it.
-
00:45:32 This will install SwarmUI very fast. Once you see the SwarmUI folder here you can start downloading
-
00:45:38 models or you can not install SwarmUI and just download models into ComfyUI as well.
-
00:45:44 This machine is a little bit slow at installation unfortunately. Okay SwarmUI folder appeared so we
-
00:45:50 can begin downloading models. RunPod SimplePod model download instructions copy this open a new
-
00:45:57 terminal paste it. Meanwhile both SwarmUI and ComfyUI are getting installed. You can follow
-
00:46:04 each of them. Unfortunately this machine is very slow. So for SwarmUI installation you
-
00:46:09 see it has generated Cloudflared link agree customize okay next just yourself none this
-
00:46:17 is important because we are always going to use our custom ComfyUI installation next I
-
00:46:22 am not going to download anything next yes I am sure to install and SwarmUI installed. So
-
00:46:28 our model downloader started let's open it from this Gradio link. It will automatically see my
-
00:46:34 SwarmUI installation and its accurate folder. So you see it sees the models folder here.
-
00:46:41 If you don't want to install SwarmUI and use it go to ComfyUI right click models copy path.
-
00:46:48 Then put a backslash and type like this. So it starts with backslash then enable ComfyUI
-
00:46:55 folder structure and download the models. So let's download LTX 2 for ComfyUI installation. I will
-
00:47:04 also use same folder for SwarmUI. Okay let's wait for ComfyUI installation to be completed
-
00:47:11 meanwhile the models are getting downloaded we can see them here it is downloading them.
-
00:47:17 Okay so our ComfyUI installation has been completed. Now let's start our ComfyUI. To do
-
00:47:24 that return back to your RunPod SimplePod ComfyUI instructions txt file and in the below you will
-
00:47:31 see the run command. This is important to use it. Open a new terminal and copy paste and hit enter.
-
00:47:39 You see we are starting with use Sage Attention. You can remove this add here other additional
-
00:47:45 arguments. This is how we start the ComfyUI instance. Initial starting may be a little bit
-
00:47:51 slow but you should see the appropriate messages in here. We can see that how much VRAM and RAM
-
00:47:58 memory our machine has. We have 143 gigabytes of VRAM and 1.8 terabytes of RAM memory yes this is
-
00:48:07 an massive RAM memory. So it is using CUDA 13 torch 2.9.1 as I said my installers are
-
00:48:15 properly made for every GPU out there so you can use my installer with every GPU on cloud
-
00:48:22 or on Windows. So the ComfyUI has been started. When you see this it means it has been started.
-
00:48:28 How to connect? Return back to your MyPod instance and you will see that there is port 3001. When you
-
00:48:35 open it now you will see your ComfyUI interface which is running on RunPod and the rest is exactly
-
00:48:44 same as using on your Windows computer. So my ComfyUI is started. All of the models also have
-
00:48:51 been downloaded. Now I am ready to use presets. So let's drag and drop our LTX 2 audio lip sync image
-
00:48:57 to video preset. It is loaded. You see there is no red no error messages it is perfect. So
-
00:49:06 choose file to upload. I will do the same as on the Windows part. So let's see how much it will
-
00:49:13 take. This is 45 seconds. Let's also copy paste our prompt. I also need to set my duration the
-
00:49:21 rest is exactly same as on the Windows tutorial part. You can also calculate 45 seconds equal to
-
00:49:29 45 multiplied 24 plus 1 1081 frames. And we can hit run here. Now you can follow the CMD window
-
00:49:41 on here. Let's also see the nvitop. pip install nvitop. nvitop. So we can follow the VRAM usage
-
00:49:51 from here. The initial loading of the model may be slow but we will see. You will also get these
-
00:49:58 messages you can ignore them it is still working perfect. This is ComfyUI mistake actually on some
-
00:50:05 models it shows these as unexpected or some other messages but it is working perfect right now. This
-
00:50:11 is also ComfyUI mistake which they have to fix but it is not fixed yet. By the way on this machine we
-
00:50:18 can also use BF16 model. Our unified downloader is downloading the FP8. How you can download it
-
00:50:24 on this machine? In the video generation models you can go to LTX 2 video models and you can
-
00:50:30 download the BF16 version directly. Because this machine has a huge amount of VRAM memory.
-
00:50:37 So the initial loading of the model is taking forever on RunPod disks because they are slow.
-
00:50:43 Okay model loaded and generation started. You see it is almost instant even for 45 seconds.
-
00:50:52 The first 8 steps took only 13 seconds. And now it is upscaling. It is using 48 gigabytes of
-
00:50:59 VRAM memory. So if you have RTX Pro 6000 on your Windows machine you can run this model very fast.
-
00:51:06 I mean shame on Nvidia. They are still selling us RTX 5090 from only 32 gigabytes of VRAM memory.
-
00:51:15 Okay this was also done very fast so it is like 30 seconds to generate 45 seconds video. We are
-
00:51:22 almost done. It is doing the now VAE decoding. It is using 57 gigabytes of VRAM memory right
-
00:51:29 now. Yeah it is already okay it is now combining video after that it will be completed. So it takes
-
00:51:36 like 1 minute to generate 45 seconds video on H200 and it is already done yes. Let me show.
-
00:51:58 Okay it is same as in the Windows tutorial part it is working perfect no issues directly working.
-
00:52:05 So how you can also use the SwarmUI in here? Since everything is ready all I need to do is
-
00:52:12 I need to turn off the ComfyUI. So that SwarmUI can use the entire GPU. To turn off the ComfyUI
-
00:52:21 normally you wouldn't need this if you don't run the ComfyUI. When you return back to the
-
00:52:26 prompts you will see that it has this command. So copy and open a new terminal if you want to
-
00:52:33 terminate your ComfyUI. This will terminate the ComfyUI properly and you will see that
-
00:52:39 it is using zero amount of VRAM. Now my SwarmUI already had started. I am going to go to server
-
00:52:47 backends ComfyUI self starting. I need to copy the path of the ComfyUI main.py so right click
-
00:52:56 copy path. Go to your SwarmUI put a backslash like this and --use-sage-attention and save. Now the
-
00:53:07 SwarmUI will be ready. Normally we won't have any models here because we didn't download them into
-
00:53:14 here we downloaded them into ComfyUI folder. So I can go to server configuration and I can go to
-
00:53:21 models folder here copy path put a backslash like this save after that we may be needed to restart
-
00:53:30 but let's see. Okay let's refresh. Yes the model arrived but it doesn't see the other model because
-
00:53:38 I need to change this into checkpoints. This is where it is downloaded in the ComfyUI. Yes. Now
-
00:53:46 I am ready to use everything. The VAE is here let's see. Okay it doesn't show but let's see
-
00:53:54 the Loras. Okay Lora is also different it must be something like this. By the way you don't
-
00:54:00 need this if you just download them into your SwarmUI folder. So let's see the Loras arrived.
-
00:54:07 Okay server configuration checkpoints Loras VAE. Oh I know why it is not working because it is not
-
00:54:15 Loras but Lora in ComfyUI. Yes like this. And the VAE folder is we need to check inside models it
-
00:54:24 is yeah accurate VAE. So you can look for the folder names. The Lora is Lora here let's see.
-
00:54:32 Okay this is empty. The Lora empty. Loras empty. And it is Loras. So it is case sensitive on Linux
-
00:54:42 machine therefore it is this. Yes. Save. And yes now I can see the Loras. Now I can import
-
00:54:51 the presets. SwarmUI model downloader the preset 45 import and it is ready. So now I can generate
-
00:55:02 anything I want. Let's generate text to video. So it is let's sort by the name. You can also
-
00:55:09 search from here like LTX LTX 2 text to video direct apply. So models are selected. The Loras
-
00:55:18 are also selected. All I need to do is just type my prompt. A man talking to camera and speaking
-
00:55:27 this is amazing. This is a just simple prompt you know the drill from the Windows tutorial part. You
-
00:55:34 can follow the status from logs debug menu. Now the SwarmUI will work perfectly fine on RunPod.
-
00:55:43 So how you can download the generations? Go to your output folder. So in here I can download
-
00:55:50 entire output with right click and download as an archive. It will start to download like
-
00:55:56 this. So you see it is downloaded the entire output. Since our downloader also downloads
-
00:56:02 the clip models text encoder models VAE models you won't see any of them downloaded from here.
-
00:56:08 We do everything fully automatically for you with the fastest and most accurate way both
-
00:56:15 in Windows and also in Linux. My Massed Compute instructions Massed Compute install files are
-
00:56:22 all for Linux so if you are a Linux user you can use them right away. It is still loading
-
00:56:27 model for SwarmUI to generate. Okay so the generation has been completed let's see it.
-
00:56:33 This is amazing.
-
00:56:35 Yeah this was our simple prompt and we got it. So how to download SwarmUI generations? Go to
-
00:56:41 SwarmUI go to output local. You can download entire raw folder download as an archive and
-
00:56:48 it will start download and save it and you have that. So how to terminate your machine
-
00:56:53 from stop pod. This will reduce your cost per hour but it will not make it zero. To
-
00:57:00 make it zero just terminate pod and everything is deleted. If we were using storage we would
-
00:57:06 still have it in here. But we are not using permanent storage so all is gone.
-
00:57:11 And one more thing before we proceed the RunPod developer team had replied my emails and they
-
00:57:18 are working on improving the Nvidia driver of the pods. You can see their answer here.
-
00:57:27 So hopefully in future all of the machines will work with our ComfyUI CUDA 13 installation. So
-
00:57:35 if your machine doesn't work currently you can select the CUDA 12.9 probably
-
00:57:42 they will add 13 too or CUDA 12.8 but this will probably won't be necessary
-
00:57:50 once they upgrade the Nvidia driver of all the machines. This is for now.
-
00:57:55 And as a next step I will show how to use SimplePod. For SimplePod we are going to use
-
00:58:00 RunPod SimplePod ComfyUI instructions file. You see the SimplePod is here. What is the difference?
-
00:58:06 Let's register from this link I appreciate that. Then go to your dashboard add some billing like
-
00:58:13 this. Then we are going to use this SimplePod template. This is faster than the RunPod. You
-
00:58:20 see it is showing this use template like this. This template already has all the features that
-
00:58:29 we need. Now you need to rent a machine. So we have RTX 5090 very cheap on SimplePod you
-
00:58:38 see it is only 47 cents. This is double price on RunPod. RTX Pro 6000 is only 72 cents. This
-
00:58:47 is almost triple price on RunPod. So let's go with the RTX Pro 6000 Blackwell GPU. This GPU
-
00:58:54 has 96 gigabyte of VRAM memory and it is same fast as the RTX 5090 even faster than that. And
-
00:59:04 then run. So this is going to have 51 gigabytes of disk space therefore you can edit it and increase
-
00:59:12 your disk space according to your need like 200 gigabytes. Then it will re-create the disk. When
-
00:59:21 you are first time creating the disk you can also select from there. So let's return back and click
-
00:59:27 here. The template is selected. You can edit and use. You see there is use template and edit
-
00:59:32 and use. Let's use the edit and use and there is also add persistent volume. SimplePod also
-
00:59:38 has persistent volume let me show you that. So go to storage add a new persistent volume like
-
00:59:44 200 gigabytes and the name is 200. It is like 3 times cheaper than the RunPod. Save. Let's return
-
00:59:51 back to our template selection double click. And I am going to select add persistent volume
-
00:59:57 and edit and use. Now it will let me select from here 200 make this workspace this is important.
-
01:00:06 And the system is ready. So this is going to use my persistent volume. Save and use. And
-
01:00:12 when I go to my servers you will see that it is going to start. Okay we made a mistake let's do
-
01:00:20 again. Add persistent volume edit and use. So 200 workspace. Okay save and use. Oh by the way
-
01:00:29 it changed the template it still didn't selected a machine so we are going to select this machine for
-
01:00:35 example and run. So this one is going to use my persistent volume. The other one is also started
-
01:00:42 here. They don't show the persistent volume but you will understand it actually there is
-
01:00:48 one hint of it. This shows 200 gigabytes and this shows disk volume and volume. When you have disk
-
01:00:54 volume and volume it means that you are using your persistent volume. So the rest is same how same?
-
01:01:01 You see that there is Jupiter interface direct link go to there. If it asks you to continue
-
01:01:08 continue because direct means that it is using direct connection. Then you see the RunPod like
-
01:01:14 interface started. Upload your ComfyUI SwarmUI whatever you want. By the way everything is much
-
01:01:21 faster on SimplePod that is the main difference and everything is much cheaper on SimplePod. So
-
01:01:27 you can use both SimplePod and RunPod then extract archive. And the rest is same let me just show you
-
01:01:35 the start of the terminal like this 1 and 100. Let's see the installation speed. Everything
-
01:01:43 is ready in this template. Okay the installation is just amazing amazingly fast as you are seeing
-
01:01:48 right now. Moreover you can download files much faster on SimplePod or upload. You see there is
-
01:01:55 file browser when you click this it will let you to download and upload files directly. Go
-
01:02:00 to workspace and every file will be here you can just select and download like this. Or you
-
01:02:07 can just upload from here. The upload allows you to upload file and folder. It is very very fast
-
01:02:14 compared to the RunPod. This way you can upload and download much faster than RunPod. So how to
-
01:02:22 terminate machines once you are done go back to your my servers there is delete options there
-
01:02:29 is no stop option here. So let's delete and let's delete. But all the data in your permanent storage
-
01:02:37 will remain there even if we deleted exactly as on RunPod. By the way on SimplePod we have forgotten
-
01:02:44 to show one more thing which was connecting the ComfyUI interface. For connecting the ComfyUI
-
01:02:51 interface we are going to use the 3000 port. My template already exposes that 3000 port so that
-
01:03:00 you will be able to connect from 3000 port on SimplePod to your ComfyUI instance. For SwarmUI
-
01:03:08 we are also using Cloudflared link as in the RunPod. So it is same as the RunPod actually.
-
01:03:14 So this is an alternative to RunPod much faster much cheaper but the disadvantage
-
01:03:19 is that it has limited amount of different GPUs. So when I go to rent GPU docker you
-
01:03:25 see 5090 and RTX Pro 6000 are the main GPUs that you can use. These two are the very best
-
01:03:31 GPUs actually that you can use for cheap but it is missing like A100 H200 B200 very
-
01:03:38 specific GPUs but it has the most commonly used amazing GPUs. So you were asking me to
-
01:03:43 make Vast AI scripts but instead of Vast AI I am providing you SimplePod. Cheaper better faster.
-
01:03:52 So as a next step let's do on the Massed Compute. To follow the Massed Compute open the Massed
-
01:03:58 Compute instructions read txt file. Please use this link to register I appreciate that. Once
-
01:04:05 you registered and logged in go to billing and add some credits as usual then go to deploy.
-
01:04:15 And in here it will show you the available GPUs. They have amazing prices so you can use H200 if
-
01:04:23 it is available and apply our coupon. Let's see if there is any availability. Oh actually there
-
01:04:29 is available yes nice. So you see currently it is 2.6 dollars per hour with our coupon SECourses it
-
01:04:36 becomes 1 dollar 95 cents. This is an amazing it is like 4 dollars on RunPod it is 2 dollars here.
-
01:04:45 So we can rent this or you can rent RTX Pro 6000 if it is available. Oh they have sold out all
-
01:04:52 the RTX Pro 6000s. They are adding new GPUs but they are very popular. So let's go with this one.
-
01:05:00 This is even better at training than inference. So if you are going to do training since this GPU
-
01:05:06 has HBM memory high bandwidth unified memory it is faster than the other GPUs like RTX Pro 6000
-
01:05:15 or 5090. By the way when you are deploying don't forget that select category as a creator select
-
01:05:22 image as an SECourses. This is our image. This is mandatory to select then apply our coupon and
-
01:05:28 pick any GPU you want. You need to pick any GPU spots are not working but all other GPUs
-
01:05:34 are working like this. You see it is 41 cents. For example L40S it is 61 cents. Like H100 it
-
01:05:44 is okay this is spot instance yes H100 it is 1.76 cents but I recommend H200 this is an amazing GPU.
-
01:05:54 So wait for GPU to be initialized it will take a little while. To connect we are going to use
-
01:06:00 ThinLinc client as usual so download the ThinLinc client from here for Windows Linux Mac. So just
-
01:06:07 download it. Then double click and start and next next install it will install and start.
-
01:06:14 From this ThinLinc client you need to go to options local devices just clipboard
-
01:06:19 synchronization and drives and add a shared folder with read and write permission I am
-
01:06:24 using this folder. Then move the downloaded ComfyUI and SwarmUI files into your shared
-
01:06:32 folder so that you will be able to access and use them. You can also put your other
-
01:06:37 files but you cannot use this for big files. For big files you have to use other storage
-
01:06:45 system like Hugging Face or Google Drive. Okay Massed Compute is getting initialized.
-
01:06:50 Machine initialization is now faster than before it should be done in 5 to 10 minutes.
-
01:06:56 Okay so the machine has been initialized let's connect it copy this login URL. Okay this was
-
01:07:02 done. Paste here copy password paste here copy username paste here connect. Continue. Okay it
-
01:07:11 is connecting then start. Okay it is starting and it is started. This is running on cloud on
-
01:07:19 Massed Compute. Go to home go to thin drives. You can also login your Patreon from here and
-
01:07:25 download links. Go to Massed Compute you see it is like your Windows computer there is
-
01:07:31 Chrome you can login download upload you can do whatever you want. Wait it to list the files in
-
01:07:38 your shared folder in here. So it will list all the files I have in here. This is synchronizing
-
01:07:45 with my computer but it is slow. So I am going to copy paste all the files version 78 and SwarmUI
-
01:07:54 latest version into downloads folder. Do not run anything in this shared folder. Always copy them
-
01:08:01 into downloads folder then we are going to run them from here. So wait for copy to be completed
-
01:08:08 it shows the progress here. The ComfyUI is done so extract here and Massed Compute instructions read
-
01:08:17 txt file. So to install we are going to use this command copy open a new terminal and paste it.
-
01:08:27 Then it will ask us to which ones to install 1 and 100 and hit enter. So the installation on Massed
-
01:08:36 Compute is like 10 times to 100 times faster than on RunPod the disk speed is like 100 times faster
-
01:08:42 than RunPod on Massed Compute. Their disadvantage is that they don't have permanent storage system
-
01:08:48 yet so every time you have to start from beginning but the speed is just mind blowingly fast so it
-
01:08:56 takes like 2 minutes to setup and start the Massed Compute. So installation is almost completed you
-
01:09:02 see. Let's also download the files to download the files we are going to use SwarmUI model downloader
-
01:09:07 extract here. Massed Compute already has SwarmUI so that you don't need to install but if you want
-
01:09:14 you can install. So Massed Compute model downloads here I will copy copy paste and open terminal and
-
01:09:23 paste. Okay it is going to install and start the model downloader this is really fast.
-
01:09:29 I will download files into the SwarmUI folder but I will make them to be used by the ComfyUI as well
-
01:09:39 like in the Windows. So I will have the access to the files from both applications. You see it is
-
01:09:45 currently using this as a base path this is where the SwarmUI installed in Massed Compute. So let's
-
01:09:54 download LTX 2 bundle download all. The download is also really fast so it started. Before I start
-
01:10:03 the ComfyUI instance I am going to do a change. Okay it is already installed. It took like 1
-
01:10:08 minutes. Go back to ComfyUI and you will see that there is extra model paths.yaml file open it and
-
01:10:18 this is the base path I am going to change this into this. Okay. Let me zoom in. Okay can I zoom
-
01:10:25 in? No. So this is the path where I want ComfyUI to read model files save it. Okay it is not saved
-
01:10:34 yet it is saving right now. Okay it is saved. Then copy this and paste into ComfyUI folder.
-
01:10:42 Then return back to Massed Compute instructions. I am first going to show ComfyUI usage then we
-
01:10:48 will see in SwarmUI. Okay then we are going to use this command to start this one so copy this
-
01:10:58 open a new terminal inside this folder ComfyUI version 78. This is our main installation folder
-
01:11:06 paste. And it will start the ComfyUI. You can see the messages in here. Always look for the
-
01:11:14 CMD to see the accurate messages. You see it is loading everything properly with torch 2.9.1 and
-
01:11:22 CUDA 13. It is loaded. How to access this from our computer? For accessing it I am going to use this
-
01:11:31 URL copy link and in here type 8888 like this and hit enter. Because it has started in this port as
-
01:11:41 you can see oh actually this is the port not 8888 so let's make it like this yes. So now I am able
-
01:11:49 to access the ComfyUI running in Massed Compute from my Windows computer. The rest is same as in
-
01:11:58 the Windows tutorial part I am going to just load my preset for example audio to lip sync.
-
01:12:05 Yes perfectly loaded this is running on the Massed Compute. I am going to choose my demo files like
-
01:12:13 this one. Let's choose our audio file this one. Then set this as in the Windows 45 seconds 1081
-
01:12:25 frames and I need to type my prompt as usual this is my prompt and run generate. Now let's
-
01:12:36 see the speed of beginning the model loading and everything. Okay it is already loaded.
-
01:12:43 It is going to start processing in a moment. The speed of Massed Compute is unchallenged.
-
01:12:49 Not SimplePod not RunPod nothing can match the disk speed of the Massed Compute it already
-
01:12:55 started generating so the whole process takes like 5 minutes once you done it. You will get
-
01:13:03 used to it you can use the Massed Compute all the time if you don't have a powerful GPU this is one
-
01:13:08 of my recommended cloud server provider. So we are almost done to get 45 seconds HD video. It
-
01:13:18 is taking like 1 minutes lesser than 1 minutes to generate. We can see the whole process live while
-
01:13:25 I am recording. Everything is so transparent. So it is now VAE decoding. Let's see the output
-
01:13:33 in here so it is going to do video combine this is running on cloud not on my machine. But I am
-
01:13:39 using it from my computer. And it is almost ready. Yes. Yes it is ready let's see it.
-
01:13:55 So you can right click and save as the video like right click and save
-
01:14:00 preview or it is also saved in the outputs folder inside ComfyUI inside output so the
-
01:14:09 video is here I can copy this and move it into ThinLinc client you see it is here
-
01:14:16 this is with audio. So save preview will also download the video into your computer as well.
-
01:14:26 So how to use SwarmUI with this ComfyUI installation in Massed Compute? I am going to
-
01:14:32 close my ComfyUI and Ctrl Alt D it will minimize everything. You see we have RunPod stable SwarmUI
-
01:14:41 update this is actually latest version of SwarmUI and it will update it alternatively you can also
-
01:14:47 use the new installer from here but this is working pretty good so it will update SwarmUI
-
01:14:56 then go to server backends and change this. You should use new installed ComfyUI always
-
01:15:02 so it is here ComfyUI Ctrl L it will select the path Ctrl C and edit this and paste it and then
-
01:15:12 main.py then in here this is like this main.py then in here --use-sage-attention and save. We are
-
01:15:23 ready but we need to check the models are already here since I have downloaded them already into the
-
01:15:54 SwarmUI folder I need to check the Lora folder where it is. So it is Lora here. Choose file.
-
01:16:00 The presets are here amazing SwarmUI presets import refresh sort by name.
-
01:16:07 Now I can use the preset that I want for example LTX 2 text to video direct apply
-
01:16:16 always quick tools reset params to default then direct apply. Then type your prompt a spaceship
-
01:16:23 going in space with cinematic background music. Okay let's select our aspect ratio and generate.
-
01:16:33 So the usage is exactly same as in the Windows I am not repeating in details please do not
-
01:16:39 skip those parts. But this is running on the cloud machine you can see that it will start
-
01:16:46 the processing immediately on Massed Compute. This is fast and I can generate very long videos
-
01:16:53 this way advanced options text to video. I can make this 481 frames let's play this one.
-
01:17:02 So you see it is working amazing. Please like and subscribe. Join
-
01:17:07 our Patreon to support us and get these amazing presets. See you later.
