>>31150
Utilizing the "Lora Double Block" node I got from YAW.
Long rant incoming. A few more thoughts after a couple of days using runpod for the first time. It's amazing this tech exists for the average joe to throw $25 at the internet and have access to top of the line tech at their fingertips. It's less amazing that cloud tech isn't exactly...perfect. From slow loading of ComfyUI elements, to random disconnect/reconnect messages, to even SSH and root access being cut off/reset at odd times, to a pod sometimes having all the GPU's rented out but your data is there so you just... have to wait for who knows who long for one to free up, it's not 100% smooth. Beyond that, while it is indeed cheaper by a significant margin electricity wise, it surprised me that an A40 is not all that much *faster* at generating, at least at the resolutions I was doing on my 3080. Is it faster? Sure, but I expected... like... way faster. It's like.. idk... 20-30% faster at the resolutions I've been working with in the YAW workflow. Nothing mind blowing. At least it can do additional up-scaling and interpolation if I want but... tbh it adds so much generation time I generally don't bother.
I'll admit I've been kind of stuck in a rut using ga1n3rb0t + the slime girl lora. The slime generally tends to cover up imperfections in hands and such. I really should branch out, but I've just been kind of... fascinated. Like watching a lava lamp with tits. I was disappointed that the breast growth lora didn't really work at all (though I did get one golden seed that did reverse breast growth).
For anyone wondering if the Fast/Quantized/Distilled/GGUF versions of Hunyuan are worth playing with on their home computers... no. Not really. All the smaller models are vastly inferior to the full model.
I tried out Nvidia Cosmos. It's.. interesting. Especially since you're capable of using the 14B model on the A40. But even being able to use a starting and ending image, the results were not great, at least for extreme anime fats. Maybe it would do ok with more realistic fats, but I didn't bother cause who wants to see that. It's nothing compared to what Kling/etc can do with Image2Video.
As for the YAW workflow, I think it's the best one ATM. It's easy to understand and well-documented (for the most part). I do change a couple of things...
I usually change the base (fast) resolution from 368x208 to 384x224 in both landscape and portrait (inputs 1 and 3 on step 4 in YAW). Almost no increase in generation time, better when doing intermediate i2v.
I also generally do 117 frames instead of 141. The difference between a 5 or 6 second video is minimal, and the extra second of clip time adds over a minute of total intial + i2v generation time. Not worth it in my opinion. I don't know how some of these guys are starting at a much higher resolution and upscaling from there, I don't have the patience to wait 15+ minutes for ONE 6 second clip.
I change the Teacache scheduler back to original (1x) on the initial generation. Using the cache here only speeds it up by 10-15 seconds, it's already pretty fast, and it presents a generally worse image for i2v. I do leave the teacache at 1.6x on the i2v though, the speed increase there really matters.
With these settings I can generate the initial "preview" clip in 20-25s, and if I choose to progress the clip, I can get a mid-res I2V upscale at 448x768 in around ~4.5m. For a total time-per-clip of less than 5 minutes at 448x768. If I choose to go further with the flat 2x upscale to 896x1536, it's an additional 3 minutes (the result is more or less the same as just zooming into the video manually). So 8 minutes total for a 5 second clip at 896x1536. Interpolation can only be done at 73 frames, as trying to do it at 117 or above OOMs the A40.
Using 141 frames makes the I2V take about 6.5 minutes, the additional upscale then takes about 3 minutes as well. So about 10 minutes for a 6 second clip at 896x1536. I tend to be zooming through generations, so I rarely want to wait upwards of 25% longer per generation for just 1 additional second per clip.
Definitely disappointing we have no real way to extend a video from the last frame yet ion Hunyuan. Hope that comes soon.
thanks for coming to my ted talk.