How do you create truly realistic facial expressions with z-image?
https://redd.it/1q36whm
@rStableDiffusion
https://redd.it/1q36whm
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: How do you create truly realistic facial expressions with z-image?
Explore this post and more from the StableDiffusion community
Trellis 2 is already getting dethroned by other open source 3D generators in 2026
Today I saw two videos that show what 2026 will hold for 3D model generation.
A few days ago Ultrashape 1.0 released their model and can create much more detailed 3D geometry, then Trellis 2, without textures though, but an extra pass with the texture part of Trellis 2 might be doable.
https://github.com/PKU-YuanGroup/UltraShape-1.0
https://youtu.be/7kPNA86G\_GA?si=11\_vppK38I1XLqBz
Also the base models of Huyuan 3D and Sparc 3D, Lattice and FaithC, respectively are planed to release, together with other nice 3D goodness, already out or coming.
https://github.com/Zeqiang-Lai/LATTICE
https://github.com/Luo-Yihao/FaithC
https://youtu.be/1qn1zFpuZoc?si=siXIz1y3pv01qDZt
Also a new 3D multi part generator is also on the horizon with MoCa:
https://github.com/lizhiqi49/MoCA
Plus for auto rigging and text to 3d animations, here are some ComfyUi addons:
https://github.com/PozzettiAndrea/ComfyUI-UniRig
https://github.com/jtydhr88/ComfyUI-HY-Motion1
https://redd.it/1q3ijwo
@rStableDiffusion
Today I saw two videos that show what 2026 will hold for 3D model generation.
A few days ago Ultrashape 1.0 released their model and can create much more detailed 3D geometry, then Trellis 2, without textures though, but an extra pass with the texture part of Trellis 2 might be doable.
https://github.com/PKU-YuanGroup/UltraShape-1.0
https://youtu.be/7kPNA86G\_GA?si=11\_vppK38I1XLqBz
Also the base models of Huyuan 3D and Sparc 3D, Lattice and FaithC, respectively are planed to release, together with other nice 3D goodness, already out or coming.
https://github.com/Zeqiang-Lai/LATTICE
https://github.com/Luo-Yihao/FaithC
https://youtu.be/1qn1zFpuZoc?si=siXIz1y3pv01qDZt
Also a new 3D multi part generator is also on the horizon with MoCa:
https://github.com/lizhiqi49/MoCA
Plus for auto rigging and text to 3d animations, here are some ComfyUi addons:
https://github.com/PozzettiAndrea/ComfyUI-UniRig
https://github.com/jtydhr88/ComfyUI-HY-Motion1
https://redd.it/1q3ijwo
@rStableDiffusion
GitHub
GitHub - PKU-YuanGroup/UltraShape-1.0: High-Fidelity 3D Shape Generation via Scalable Geometric Refinement
High-Fidelity 3D Shape Generation via Scalable Geometric Refinement - PKU-YuanGroup/UltraShape-1.0
TwinFlow: Realizing One-step Generation on Large Models with Self-adversarial Flows
https://huggingface.co/inclusionAI/TwinFlow-Z-Image-Turbo
https://redd.it/1q3lrk6
@rStableDiffusion
https://huggingface.co/inclusionAI/TwinFlow-Z-Image-Turbo
https://redd.it/1q3lrk6
@rStableDiffusion
huggingface.co
inclusionAI/TwinFlow-Z-Image-Turbo · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Wan2.2 : better results with lower resolution?
Usually I do a test by generating at a low resolutions like 480x480 , if I like the results I generate at a higher resolution.
But in some cases I find the low resolution generations to be better in prompt adherence and looking more natural, higher resolutions like 720x720 some time look weird.
Anyone else notice the same?
https://redd.it/1q3lq5n
@rStableDiffusion
Usually I do a test by generating at a low resolutions like 480x480 , if I like the results I generate at a higher resolution.
But in some cases I find the low resolution generations to be better in prompt adherence and looking more natural, higher resolutions like 720x720 some time look weird.
Anyone else notice the same?
https://redd.it/1q3lq5n
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Release: Invoke AI 6.10 - now supports Z-Image Turbo
The new Invoke AI v6.10.0 RC1 now supports Z-Image Turbo... https://github.com/invoke-ai/InvokeAI/releases
https://redd.it/1q3ruuo
@rStableDiffusion
The new Invoke AI v6.10.0 RC1 now supports Z-Image Turbo... https://github.com/invoke-ai/InvokeAI/releases
https://redd.it/1q3ruuo
@rStableDiffusion
GitHub
Releases · invoke-ai/InvokeAI
Invoke is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The ...
This media is not supported in your browser
VIEW IN TELEGRAM
Time-lapse of a character creation process using Qwen Edit 2511
https://redd.it/1q3sb0z
@rStableDiffusion
https://redd.it/1q3sb0z
@rStableDiffusion
The Z-Image Turbo Lora-Training Townhall
Okay guys, I think we all know that bringing up training on Reddit is always a total fustercluck. It's an art more than it is a science. To that end I'm proposing something slightly different...
Put your steps, dataset image count and anything else you think is relevant in a quick, clear comment. If you agree with someone else's comment, upvote them.
I'll run training for as many as I can of the most upvoted with an example data set and we can do a science on it.
https://redd.it/1q3tcae
@rStableDiffusion
Okay guys, I think we all know that bringing up training on Reddit is always a total fustercluck. It's an art more than it is a science. To that end I'm proposing something slightly different...
Put your steps, dataset image count and anything else you think is relevant in a quick, clear comment. If you agree with someone else's comment, upvote them.
I'll run training for as many as I can of the most upvoted with an example data set and we can do a science on it.
https://redd.it/1q3tcae
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Turned myself into a GTA-style character. Kinda feels illegal
https://redd.it/1q3vjp7
@rStableDiffusion
https://redd.it/1q3vjp7
@rStableDiffusion
WAN2.2 SVI v2.0 Pro Simplicity - infinite prompt, separate prompt lengths
https://redd.it/1q3wjyo
@rStableDiffusion
https://redd.it/1q3wjyo
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: WAN2.2 SVI v2.0 Pro Simplicity - infinite prompt, separate prompt lengths
Explore this post and more from the StableDiffusion community
SVI: One simple change fixed my slow motion and lack of prompt adherence...
https://redd.it/1q45liy
@rStableDiffusion
https://redd.it/1q45liy
@rStableDiffusion
LTXV2 Pull Request In Comfy, Coming Soon? (weights not released yet)
https://github.com/comfyanonymous/ComfyUI/pull/11632
Looking at the PR it seems to support audio and use Gemma3 12B as text encoder.
The previous LTX models had speed but nowhere near the quality of Wan 2.2 14B.
LTX 0.9.7 actually followed prompts quite well, and had a good way of handling infinite length generation in comfy, you just put in prompts delimited by a '|' character, the dev team behind LTX clearly cares as the workflows are nicely organised, they release distilled + non distilled versions same day etc.
There seems to be something about Wan 2.2 that makes it avoid body horror/keep coherence when doing more complex things, smaller/faster models like Wan 5B, Hunyuan 1.5 and even the old Wan 1.3B CAN produce really good results, but 90% of the time you'll get weird body horror or artifacts somewhere in the video, whereas with Wan 2.2 it feels more like 20%.
On top of that some of the models break down a lot quicker with lower resolution, so you're forced into higher res, partially losing the speed benefits, or they have a high quality but stupidly slow VAE (HY 1.5 and Wan 5B are like this).
I hope LTX can achieve that while being faster, or improve on Wan (more consistent/less dice roll prompt following similar to Qwen image/z image, which might be likely due to gemma as text encoder) while being the same speed.
https://redd.it/1q49ulp
@rStableDiffusion
https://github.com/comfyanonymous/ComfyUI/pull/11632
Looking at the PR it seems to support audio and use Gemma3 12B as text encoder.
The previous LTX models had speed but nowhere near the quality of Wan 2.2 14B.
LTX 0.9.7 actually followed prompts quite well, and had a good way of handling infinite length generation in comfy, you just put in prompts delimited by a '|' character, the dev team behind LTX clearly cares as the workflows are nicely organised, they release distilled + non distilled versions same day etc.
There seems to be something about Wan 2.2 that makes it avoid body horror/keep coherence when doing more complex things, smaller/faster models like Wan 5B, Hunyuan 1.5 and even the old Wan 1.3B CAN produce really good results, but 90% of the time you'll get weird body horror or artifacts somewhere in the video, whereas with Wan 2.2 it feels more like 20%.
On top of that some of the models break down a lot quicker with lower resolution, so you're forced into higher res, partially losing the speed benefits, or they have a high quality but stupidly slow VAE (HY 1.5 and Wan 5B are like this).
I hope LTX can achieve that while being faster, or improve on Wan (more consistent/less dice roll prompt following similar to Qwen image/z image, which might be likely due to gemma as text encoder) while being the same speed.
https://redd.it/1q49ulp
@rStableDiffusion
GLM-Image AR Model Support by zRzRzRzRzRzRzR · Pull Request #43100 · huggingface/transformers
https://github.com/huggingface/transformers/pull/43100/files
https://redd.it/1q42gv8
@rStableDiffusion
https://github.com/huggingface/transformers/pull/43100/files
https://redd.it/1q42gv8
@rStableDiffusion
GitHub
GLM-Image AR Model Support by zRzRzRzRzRzRzR · Pull Request #43100 · huggingface/transformers
This PR is to adapt the implementation of the AR model for GLM-Image.