r/StableDiffusion – Telegram
Update to Repo for my AI Toolkit Fork + New Yaml Settings for I2V motion training

Hi, PR has already been submitted to Ostris but yeah... my last one hasn't even been looked at. So here is my fork repo:
[https://github.com/relaxis/ai-toolkit](https://github.com/relaxis/ai-toolkit)

Changes:

1. Automagic now trains separate LR per lora (high and low noise) if it detects MoE training - LR outputs now print to log and terminal. You can also train each lora according to different optimizer parameters:

​

optimizer_params:
lr_bump: 0.000005 #old
min_lr: 0.000008 #old
max_lr: 0.0003 #old
beta2: 0.999
weight_decay: 0.0001
clip_threshold: 1
high_noise_lr_bump: 0.00001 # new
high_noise_min_lr: 0.00001 # new
high_noise_max_lr: 0.0003 # new
low_noise_lr_bump: 0.000005 # new
low_noise_min_lr: 0.00001 # new
low_noise_max_lr: 0.0003 #new

2. Changed resolution bucket logic - previously this worked on SDXL bucket logic but now you can specify pixel count. The logic will allow higher dimension videos and images to be trained as long as they fit within the specified pixel count (allows for higher resolution low vram videos below your cut off resolution).

resolution: - 512
max_pixels_per_frame: 262144

https://redd.it/1oiyuzr
@rStableDiffusion
How do people use WAN for image generation?

I've read plenty comments mentioning how good WAN is supposed to be with image gen, but nobody shares any specific or details about it.

Do they use the default workflow and modify settings? Is there a custom workflow for it? If its apparently so good, how come there's no detailed guide for it? Couldn't be better than Qwen, could it?

https://redd.it/1oj8ubq
@rStableDiffusion