r/StableDiffusion – Telegram
Update to Repo for my AI Toolkit Fork + New Yaml Settings for I2V motion training

Hi, PR has already been submitted to Ostris but yeah... my last one hasn't even been looked at. So here is my fork repo:
[https://github.com/relaxis/ai-toolkit](https://github.com/relaxis/ai-toolkit)

Changes:

1. Automagic now trains separate LR per lora (high and low noise) if it detects MoE training - LR outputs now print to log and terminal. You can also train each lora according to different optimizer parameters:

​

optimizer_params:
lr_bump: 0.000005 #old
min_lr: 0.000008 #old
max_lr: 0.0003 #old
beta2: 0.999
weight_decay: 0.0001
clip_threshold: 1
high_noise_lr_bump: 0.00001 # new
high_noise_min_lr: 0.00001 # new
high_noise_max_lr: 0.0003 # new
low_noise_lr_bump: 0.000005 # new
low_noise_min_lr: 0.00001 # new
low_noise_max_lr: 0.0003 #new

2. Changed resolution bucket logic - previously this worked on SDXL bucket logic but now you can specify pixel count. The logic will allow higher dimension videos and images to be trained as long as they fit within the specified pixel count (allows for higher resolution low vram videos below your cut off resolution).

resolution: - 512
max_pixels_per_frame: 262144

https://redd.it/1oiyuzr
@rStableDiffusion
How do people use WAN for image generation?

I've read plenty comments mentioning how good WAN is supposed to be with image gen, but nobody shares any specific or details about it.

Do they use the default workflow and modify settings? Is there a custom workflow for it? If its apparently so good, how come there's no detailed guide for it? Couldn't be better than Qwen, could it?

https://redd.it/1oj8ubq
@rStableDiffusion
Your Hunyuan 3D 2.1 preferred workflow, settings, techniques?

Local only, always. Thanks.

They say start with a joke so..
How do 3D modelers say they're sorry?
They Topologize.

I realize Hunyuan 3D 2.1 won't produce as good a result as nonlocal options but I want to get the output as good as I can with local.

What do you folks do to improve your output?

My model and textures always come out very bad, like a playdoe model with textures worse than an NES game.

Anyway, I have tried a few different workflows such as Pixel Artistry's 3D 2.1 workflow and I've tried:

Increasing the octree resolution to 1300 and the steps to 100. (The octree resolution seems to have the most impact on model quality but I can only go so high before OOM).

Using a higher resolution square source image from 1024 to 4096.

Also, is there a way to increase the Octree Resolution far beyond the GPU VRAM limits but have the generation take longer? For example, it only takes a couple minutes to generate a model (pre texturing) but I wouldn't mind letting it run overnight or longer if it could generate a much higher quality model. Is there a way to do this?

Thanks fam

Disclaimer: (5090, 64GB Ram)

https://redd.it/1ojcfti
@rStableDiffusion
This media is not supported in your browser
VIEW IN TELEGRAM
Texturing using StableGen with SDXL on a more complex scene + experimenting with FLUX.1-dev

https://redd.it/1ojfsvv
@rStableDiffusion