r/StableDiffusion – Telegram
Z-Image first generation time

Hi, I'm using ComfyUI/Z-image with a 3060 (12GB VRAM) and 16 GB RAM. Anytime I change my prompt, the first generation takes between 250-350 seconds, but subsequent generations for the same prompt are must faster, around 25-60 seconds.

Is there a way to reduce the generation of the first picture to be equally short? Since others haven't posted this, is it something with my machine? (Not enough RAM, etc?)

https://redd.it/1pk13tx
@rStableDiffusion
Old footage upscale/restoration, how to? Seedvr2 doesn't work for old footage
https://redd.it/1pk4m9m
@rStableDiffusion
What are the Z-Image Character Lora dataset guidelines and parameters for training

I am looking to start training character loras for ZIT but I am not sure how many images to use, how different angles should be, how the captions should look like etc. I would be very thankful if you could point me in the right direction.

https://redd.it/1pjyzs4
@rStableDiffusion
Realtime Lora Trainer now supports Qwen Image / Qwen Edit, as well as Wan 2.2 for Musubi Trainer with advanced offloading options.
https://redd.it/1pkdrzv
@rStableDiffusion