/bbwai/ - bbwai

Stable Diffusion, Loras, Chatbots

BBW-Chan is supported by simple text or static image ads from a-ads.com.
Please consider whitelisting us in your adblocker if you enjoy your stay!


Mode: Reply
Name
Subject
Message

Max message length: 9999

Files

Max file size: 10.00 MB

Max files: 6

Captcha
E-mail
Password

(used to delete files and postings)

Misc

READ THE *RULES* BEFORE YOU POST!

[ / / ]

(4.04 MB 459x336 004a.gif)
(5.27 MB 424x240 008a.gif)
Hunyuan Video 01/12/2025 (Sun) 20:09:38 No. 30947
If you haven't tried out Hunyuan Video yet, you're missing out. Out of the box it can produce some interesting results. With Loras, what it can produce is pretty damn incredible. I've trained a few Loras using the diffusion-pipe with runpod. I'm currently on version 5.00 of my project. I've been training using a combo of 512x512 videos (33 frames) and images. I'll post some examples. Note that they've been converted from (.mp4) to (.gif) and down-scaled for upload here, so there is quality loss. I'd love to get some prompt ideas for testing. Comment if you have something you'd like to see and I'll post the results. Hunyuan uses natural language prompts.
>>30965 Lol yeah... I'm assuming this is using the base model hunyuan? It actually understands naked bodies quite well as it was trained on nsfw, but jiggle physics are whack and you'll need a lora for that. Though it will understand "rubbing," "massaging," "caressing" out of the box. If you are wanting anime, there are lots of anime loras up on civitai as well as breast jiggle loras. Also for low vram I would recommend using this workflow: https://civitai.com/models/1048570/simple-gguf-hunyuan-text2video-workflow (.gguf) version of hunyuan is a lot lighter on vram and faster. There is some quality loss, but you can still get great results. I would recommend rendering 540x540 resolution or lower @20 steps for low vram.
>>30968 Damn :(. Well, the good news is that people are working on optimizations daily. If you have a 1080ti with 11gb vram, the (.gguf) versions of hunyuan should be feasible. Or in the meantime, feel free to send some prompts my way! I'd be more than happy to test them out!
(2.50 MB 336x459 010.gif)
(2.97 MB 392x392 012.gif)
(2.08 MB 392x392 011.gif)
A few more
I have no idea how any of this shit works but can character loras be used in making these videos?
>>30996 In short, yes.. I am using a custom lora I made which specifically for fat women + belly interaction + jiggle physics. I included close-up shots of Ana de Arma's face to capture her likeness and get facial consistency. I mixed that with videos of fat women interacting with their bellies, as well as some images I rendered using sdxl Ponyv6 which depicted very oversized bellies. To make a specific character lora with results like I am posting here - you would have a dataset which includes images of your desired character, along with (33 or 65 frames) 512x512 videos which capture the movement you are looking for. That's the method I've found successful. My latest dataset included 27 videos and 17 images. I could try mixing a character lora with my lora, it would be an interesting experiment. From what I have heard, stacking multiple loras with hunyuan (at the moment) doesn't produce great results - which is why I recommended creating a seperate lora. If you browse civitai hunyuan loras and see a character lora that catches your eye, let me know, I'm down to give it a whirl.
Would be nice to see two girls with round belly pressing each other on beach in the cow patterned bikini if that possible or feeding themself with chocolates
I'll take you up on your offer of linking me some tutorials. I need to know how to install, setup, and use ComfyUI in general. Then, I'd have to learn how to install and use the low VRAM version of Hunyuan. Thanks in advance.
>>31030 Link 1: https://www.youtube.com/watch?v=wVTZj-RGIXw This is the video I found most helpful for using Hunyuan and creating Loras. The video is sectioned out nicely and he takes things step-by-step. Link 2: https://blog.runpod.io/train-your-own-video-loras-with-diffusion-pipe/ This is a more skeleton step-by-step for training Loras using the diffusion-pipe + runpod. I would recommend using the first link in combination with this blog post if you are interested in training. Link 3: https://www.youtube.com/watch?v=KTPLOqAMR0s Basic ComfyUI installation and setup for PC. Watch this one if you just want Comfy installed on your local PC and don't want to mess around with Runpod. I like this dude, he posts good StableDiffusion/Comfy content regularly. Worth looking through his channel. The most important thing for ComfyUI is having the Comfy Manager installed, which Link 1+3 both cover. It simplifies keeping ComfyUI updated and installing missing custom nodes from new workflows you want to try. As soon as you import a workflow (.json) into your Comfy workspace it'll highlight any missing nodes and offer to install them for you. A recent important node for Hunyuan has been released called "WaveSpeed" which can speed up generation times by 1.5x with minimal quality loss. I would strongly recommend adding this node to any workflow. You simply connect it to your model loading node. Link 4: https://github.com/chengzeyi/Comfy-WaveSpeed As for which workflow to use, I would strongly recommend using the (.gguf) version(s) of Hunyuan: https://civitai.com/models/1048570/simple-gguf-hunyuan-text2video-workflow All of the necessary files can be found in that persons post. GGUF is a quantization file format, compressing the original file. There will be many GGUF versions available varying in size, simply choose whichever one works best for you/your hardware. Lastly, I will be posting my Lora to CivitAI shortly. I'm satisfied with how it's been performing and would love to see what people can make with it. I will update here with a link when it is live.
(4.37 MB 394x288 015.gif)
(4.61 MB 394x288 014.gif)
>>31027 Here you go :) Also! My Lora is now available for download on CivitAI! Link: https://civitai.com/models/1144518/ga1n3rb0t?modelVersionId=1287232 Give it a try, would love to see what we can create with it. Also drop a like if you enjoy it, it's free and earns me some buzz. Thanks!
Is there a way someoone could make a girl from slim in a gym locker room getting so fast and her belly get rounder so its even pop her breasts or the underpants are ripping out?
>>31050 A Lora would be needed for expansion/weight_gain/inflation content. I've tried prompting for such with base Hunyuan achieving zero success. My current Lora is not capable of it either - it only works well for belly play and women who are already fat. I have an inflation Lora in the works but it is still early. I'm currently building up a dataset. If you know of any good inflation/expansion vids I'd love if you could share links. Thanks!
>>31051 So okay could be atleast a hot girl jiggling her thights in a shorts pajama pants in a bra with a bow?
>>31051 I tried this thing and i was able to generate test video with elephant how to implement LORAs in Comfyui?
(188.10 KB 1724x757 lora example - Copy.png)
>>31054 To insert a lora you need to add a "load lora" node. To do this, simply double-click a blank area in your workspace and a node search should pop up. Start typing in "load lora" and it should pop up. Click on it and it will be added to your workspace. Next you will need to attach it to your model an text-encoders (clip). I have circled how you connect it to the model using red. I have used green for the clip connections. You'll notice you have the options for "strength_model" and "strength_clip" in the lora loader. You can leave these both at their default of 1.00. The last thing I want to mention is what I have circled in orange - VAE Decode (Tiled). The majority of people will need to decrease the default values here or they will run out of memory. Use a "tile_size" of 128 and "overlap" of 32. You also have to change the prompt from the default prompt the workflow comes with, that's why you're generating an elephant lol. Hunyuan uses natural language so just type out sentences normally with punctuation. It does a phenomenal job at this. If you are using my lora, make sure you begin your prompt with the trigger word "ga1n3rb0t" as this is what activates the lora. Hope this helps!
>>31051 Does the training videos need to be hight quality or it can be AI videos stuff too? Can be it trained on it?Where I could send you the ones?
Sorry, also forgot to mention you will want to change the "control_after_generate" value in the KSampler from "fixed" to "randomize". If you leave the value at fixed you will be generating the same image over and over because the model is being fed the same input noise. Randomizing the noise seed will output different images.
>>31056 Nope, doesn't need to be high-quality! Preferably not potato quality lol, but as long as it's clear what's going on Hunyuan should pick it up. I try to source HD videos, but they get downscaled and cropped to 512x512 for training so they lose quality. Doesn't seem to effect the final product.
Its possible to make vids about strongfat women? Like maybe a strongfat gal in cow pattetnef bikini jiggling like a waterballoon as she fights or smt like that?
Is there a way to make vids of stronget women? I mean, like a vid of a strongfat woman in cow pattern bikini fighting as she jiggles like a massive waterballoon or smt like that?
(3.74 MB 394x394 018 - Copy.gif)
(3.82 MB 394x394 016 - Copy.gif)
>>31052 Here's my attempts :) -drewski
>>31058 If I can ask, how much videos you already got and how much you still need to make the lora? I think I could try to send you some but I don't know where I can do that, there is few videos that might be useful.
(462.73 KB 512x512 video_00015_.webp)
This is my first attempt with LORA, sadly i have low VRAM thus low resolution, and unfortunately breasts are jiggling instead of belly
>>31087 What was your prompt and settings?
(33.95 KB 1198x358 Screenshot 2025-01-16 125834.png)
(77.17 KB 539x1213 Screenshot 2025-01-16 130623.png)
Ok, I'm not gonna finagle my way through alternatives.
>>31058 Would for you to be possible to train on this compilation for the belly expansion? Or it needs to be more of that or just need something better quality? How much you material you need and what kind? https://mega.nz/file/mEIw2aCC#1ZDclSwGEnU-c2ExqC_Do6_BqbnADMj9gobIybwJ6-E
(314.97 KB 512x512 Rem1.webp)
(359.00 KB 512x512 Rem2.webp)
Made Made a couple more attempts
>>31088 Mostly same as from 4 steps guide what OP gave earlier, but sized up as for Prompt: fat nude girl lift her belly and some description of the character
>>31091 This is good stuff, thanks for sharing! I could definitely use more like that.
>>31096 Do you mean you generated using 4 steps? I would recommend using a minimum of 20 steps for the GGUF and regular versions of hunyuan. Otherwise videos will output quite fuzzy like this. If you wish to use lower step counts, Kijai did release a distilled 4-6 step lightning-style version of hunyuan called "FastVideo". Link: https://huggingface.co/Kijai/HunyuanVideo_comfy/tree/main I have limited experience with it, but people seem to be enjoying it. I also edited my lora post on civitai to include the prompts I used for my example videos, if you wanted to get an idea of what I've been using. -drewski
>> Theres another several ones you could used that I managed to get, hope it would be enought to train on those already, I still gonna post more if needed https://mega.nz/file/KN4niKrD#FHrArp401g0sgeiVi5PJWT7wm3KLPkVr090KfY_w7C4
>>31110 I meant the post where you gave links for installing Comfy and Hynyuan which have 4 steps, for my gens i used 30 steps and 141 fps, and 69 lenght (sorry my interface in russian, so i can miss something)
>>31049 >>31065 >>30977 >>30948 >>30947 this is so far one of the most impressive use of the tech, really great job folks. can't wait to see how it improves esp. getting more hi res please share more gens
>>31117 Thanks bro! 1nfL4t3 v1.0 is in the oven as we speak. No idea how it'll turn out as I'm trying out something a little different with how the training chunks out the video samples. Hopefully in ~3 hours I'll be posting some quality inflation content. Stay tuned. -drewski
>>31120 I also generated another 4 videos based on same image as the prompt, it may look same but there few angle differences and you could use it too for training https://mega.nz/file/XQAzzSBB#flAYaxw0JxTGpDvW0GrQF_HwgJLDuEk_hkcQTlX3k7g
>>31133 Awesome! Can you also get the result of boobs + hips expansion / inflation combo?
(331.09 KB 736x416 vid_00002.webm)
(335.94 KB 736x416 vid_00016.webm)
(315.80 KB 736x416 vid_00018.webm)
(334.28 KB 736x416 vid_00020.webm)
Welp, I spent all day yesterday messing with this, here are some collected thoughts. One, it's amazing that I experienced two firsts on this site: Aurora Borealis AND the first fat-compatible Hunyuan video lora were both posted here, amazing. As for the rest. Well, wow. It's incredible that this tech has filtered down so quickly from giant super computer and subscription required to kling or pixverse or whatever to consumer machines. Does Huyuan completely blow other, older methods of animation like Pyramidflow, CogVideo, and AnimatedDiff out of the water? Absolutely. Is it as good as Kling/Pixverse/etc? Ehh, on mid-grade consumer hardware, I'd say no. Maybe it can compete on 24gb VRAM and above. Maybe. And that brings me to the real kind of "downer" here: me spending 8 hours running this stuff (with tons of fails and tests) probably cost me around $8 in electricity total where I live, and I have to make a lot of concessions: small initial res, conservative upscale, no interpolation, long (5+ minute) generation times for 3 seconds clips. The reality is, it would have been cheaper, and faster, and would have had better looking results, if I had rented a runpod to do this experimenting than to run it at home on what was, four years ago, the second best GPU on the market (3080 10gb). So that's a stark look at how quickly this tech has evolved and how fast the tech a lot of people have sitting around is beginning to get outclassed. Not just in video but most actually good LLMs as well don't really fit in 10GB. Thankfully it's so cold right now that any heat I make from generating stuff is probably just money saved by causing the heat pump to not run as much (both are 600w when running with load), so it *probably* evens out, but still. As for the examples, I am thrilled that already, even in these infancy stages, it's possible to combine loras in Hunyuan video and still get a decent result. That means that eventually, just like every other model, character loras will be coming to Hunyuan at an accelerated rate, and eventually some amazing things will come of it. As of RIGHT THIS SECOND though, unless you're renting a runpod or have a 3090 or above, I would say this is *mostly* just a fun toy, and nobody should go out of their way to make this work on their home machines unless they have cheap electricity and a beefy (nvidia) GPU. If anyone wanted to donate some money to the site specifically for a BBWChan runpod for Hunyuan, I'd certainly set it up and take requests, but as of now these little Jinx clips will probably be the only "I was here when it happened" examples from me. Fascinating tech. Thanks for cluing us in, and great work on that first lora drewski. Way to take the initiative. For anyone curious, I just ran a slightly modified version of this workflow: https://civitai.com/models/1134115/hunyuan-yawyet-another-workflow-fastt2vv2vmulti-resaudiopreview-pauseupscalemulti-lora-aio-interpolateteacacheoptions To make it work on 10GB below, just make sure to us the lowest initial res it comes with (might even have to reduce that a bit on 8GB cards), disable interpolation and audio, only use the intermediate V2V upscale and not the better upscaler. Oh and change both areas with the the TileSize parameter from 256 to 128, and the overlap from 64 to 32.
Edited last time by admin on 01/17/2025 (Fri) 13:31:10.
>>31136 This is awesome, can you please gtell your prompt and settings< cause i get little to no movements in in a video? if it's not hard for you
>>31137 From what I can tell, you basically need to throw all your old prompting knowledge out of the window. with Hunyuan. The examples posted and updated with their prompts on civitai like this one hold the key: https://civitai.com/images/51822368 "ga1n3rb0t. A fat woman with a huge belly is eating cake and rubbing her belly while wearing a bikini. Her extremely large belly dominates her figure. In a luxury apartment filled with decadent food. The woman has a beautiful face, long blonde hair, plump lips, hazel eyes. The woman has huge breasts, huge ass, wide hips and thick thighs." You'll notice we skip all the best quality, masterpiece, etc we used to do. That's mostly because the max token count (without llm fudging) is only 77 with Hunyuan. Which is frustrating, because Hunyuan definitely takes a "first in last out" approach, i.e., it does *roughly* try to follow the prompt as the video goes along. So you can kind of think of the first sentence as describing the first second of video, the second sentence for the next, etc. So to get "movement", ideally, each sentence will have "implied" movement. Like in this example, there is implied movement since "eating" was used first, then "rubbing" later in the prompt. This would imply a transition from eating to rubbing, which Hunyuan would try to implement. However, it is still a giant roll of the dice.
Edited last time by admin on 01/17/2025 (Fri) 13:44:28.
(334.85 KB 736x416 vid_00001.webm)
(334.90 KB 736x416 vid_00002.webm)
A couple more I forgot I had since it was late at night and rolled over into a new folder. These were using the example prompt, with the "eating cake" part removed, and of course the jinx trigger words to get her in there.
>>31136 Hey btw, if you were able to make vids of Jinx, you can also make vida of Vi?
>>31145 nah, not really. Jinx here is one of the only character loras available for Hunyuan video atm. Someone could (and probably will) make one for VI at some point the same way they obviously used footage from the show to train Jinx. There is an arcane "Style" lora however, so maybe with prompting you could get semi-close? https://civitai.com/models/1130085/hunyuanvideo-style-lora-arcane
(5.33 MB 394x394 003.gif)
(4.18 MB 394x394 002.gif)
>>31136 Awesome generations, very cool to see the results of mixing loras. Honestly better than I expected. Are you connecting two "load lora" nodes together, or using a "lora stacking" node? And yeah, completely valid take on Hunyuan. It has the potential to output mind-blowing results, but obtaining those results on current consumer hardware is time-consuming and a gamble. Hunyuan isn't perfect, it might be every 1/3 generations that'll be a keeper (just like sdxl, any other generative ai), and this can be frustrating. This is why I was really pushing runpod for anyone who wants to experience Hunyuan's true capabilities. Renting an A40 48gb costs 0.39usd/hr, and with an A40 you can generate 73 frames @650x650 with 20 steps in ~2min. The larger vram will also give you access to more advanced workflows and better hunyuan models. It's definitely worth it to load up $5 and give it a go, if this is something you want to play around with. Onto 1nfL4t3 v1.00...Results have been meh. Certainly better than my previous attempts of trying to merge inflation content in with ga1n3rb0t. Having a separate lora is the right direction, however, inflation is proving a tricky concept to teach Hunyuan. I think this is due to the training process. To train on videos, you use 33 frame clips, which is already short, and in training the full 33 frames aren't even used. The training program takes a clip from the 33 frames, making it even shorter. You have a few options on how it clips: single_middle (one clip from the middle of 33 frames), single_beginning (one clip from the beginning), and multiple_overlapping (multiple clips attempting to cover the full range of the 33 frames). Inflation tends to be pretty gradual, so it's difficult to capture that full range of motion in such short sequences. For my training attempt I went with multiple_overlapping in hopes of covering as much of the sequence as possible. I have had some success, some videos produced are displaying active inflation. However, the majority of generations are outputting the end result of inflation - large belly but not growing/increasing in size. Hopefully alternative training methods are released in the near future. Anywho, here are some of the better results I've gotten so far in testing. -drewski
(3.30 MB 358x358 005.gif)
(2.47 MB 307x307 008.gif)
(3.63 MB 307x307 009.gif)
A few more generations messing around with 1nfL4t3 v1.00. Far from perfect, but interesting. -drewski
>>31150 What keywords do you use for trigger? Because maybe iirc you could just use "skinny girl" and "belly expansion" to emphaze that the girl belly is small and gets inflated over time? Or it might be because the data training is still low, but I still can give it a try. Eventually you can generate some slim girls while only belly would get inflated so the AI would know which parts actually get bigger and train it on the generations you got. Could be worth a try
>>31170 I labelled the dataset with "belly inflation", "belly expansion", "rapid weight gain", etc. I used both natural language as well as tagging-style. I also included images that I had generated using pony that depicted fully inflated women and labelled them as "fully inflated belly", "huge inflated belly". Better labelling will always help, maybe there's a trick to it I'm not seeing. But I think the majority of the issue still lies with training limitations of video clip size. ga1n3rb0t was successful (although it took a few attempts) because jiggling was easy to capture - no matter where it sampled a clip from in the training data it would've found jiggling/motion. I could always try increasing the speed of the clips I'm using for expansion to make the motion more dramatic, but that might cause the lora to output very rapid movement.
>>31172 Yeah that could work, or cut it only betweenwhen the expansion is visible (not when its already slim or full). Sometimes outfit helps like "crop top"
btw if you need some more to train, hope they would be better now too https://mega.nz/file/aEBWQJZI#FPEmWB1ZpgZraX1PNWgJ6WnFyk4_qvtY2QaFKjzsNAA
(2.89 MB 461x307 s - Copy.gif)
(3.02 MB 461x307 s2 - Copy.gif)
(2.94 MB 461x307 s3 - Copy.gif)
more ga1n3rb0t
(444.98 KB 736x416 vid_00006.webm)
(775.16 KB 640x416 vid_00008.webm)
(678.29 KB 640x832 vid_00002.webm)
(603.29 KB 608x416 vid_00011.webm)
>>30947 Well I caved and bought some runpod time. Also, everyone, please try to post as webm instead of gif, gifs are horridly inefficient space wise. You can generate as webm right from most workflows in comfy.
>>31179 Hell yeah, nice work! Those look minty. Which loras are you using here? And noted, I'll post webm from now on.

Delete
Report