r/StableDiffusion – Telegram
Is it just me or has the subreddit been over run with the same questions?

Between this account and my other account I’ve been with this subreddit for a while.

At the start this subreddit was filled with people asking real questions about things. Like tips or tricks for making unique workflows or understanding something. Recommend nodes to help with something particularly they’re trying to achieve. Maybe help trying to find a certain models after spending time searching and not able to find it. Or recommend videos or tutorials for something.

Now since Zimg or that what it seems like. Maybe Qwen it kinda started. Now it’s nothing but. “Best this, best that or best everything. How to make adult content this or that”..No actual real question I can try and answer.

The best question to me is” I’m new and don’t know anything and wanting to jump straight to using high end complex and advanced models or workflows without learning the very basics. So show me how to use it”


This could just be me. Or has anyone else that been doing this awhile have the same feeling?

https://redd.it/1pt1sop
@rStableDiffusion
We need a pin linking to the wiki (a guide to getting started), which should be updated. Too many redundant "how do I install a1111???" posts.

Every day there is at least one post which is something along the lines of

\- "Guys I can't install stable diffusion!!!"

\- "Guys why isn't a1111 working????? Something broke when I updated!!!"

\- "Guys I tried using *model from the last 1.5 years* and it makes this strange pattern??? btw it's stable diffusion"

\- "Guys I have an AMD GPU, what do I do????"

In the last 2 hours alone there were 2 posts like this. This sentiment also exists in the comments of unrelated posts, like people going "oh woe is me I don't understand Scratch, a shame Comfy is the only modern UI...".

The sub's wiki is a bit old, but all it needs is a small update linking to Stability Matrix, SDNext, Forge Classic Neo, etc., a big fat disclaimer to not use a1111 and that it's abandoned, cull the links to A1111/DirectML (which nukes performance), and add links to relevant ZLUDA/ROCm install guides - SDNext literally has docs for that, don't even need to include any explanation in the sub's wiki itself, just links. 5 minute change.

A pinned "read this before you make a new thread" post linking to such an updated wiki should hopefully inform people of how to properly get started, and reduce the number of these pointless posts that always have the same answer. Of course, there will always be people who refuse to read, but better than nothing.

https://redd.it/1pt5o1l
@rStableDiffusion
Why do I get better results with Qwen Image Edit 4 Step lora than original 20 step?

4 step takes less time and output is being better. Isn't more steps supposed to provide better image? I'm not familiar with this stuff but I thought slower/bigger/more steps would result in better results. But with 4 steps, it creates everything including text and the second image i uploaded accurately compared to 20 where text and the second image i asked for it to include gets distorted

https://redd.it/1pt6fdn
@rStableDiffusion
Last week in Image & Video Generation

I curate a weekly multimodal AI roundup, here are the open-source diffusion highlights from last week:

TurboDiffusion - 100-205x Speed Boost

Accelerates video diffusion models by 100-205 times through architectural optimizations.
Open source with full code release for real-time video generation.
[GitHub](https://github.com/thu-ml/TurboDiffusion) | [Paper](https://arxiv.org/pdf/2512.16093)

https://reddit.com/link/1ptggkm/video/azgwbpu4pu8g1/player

Qwen-Image-Layered - Layer-Based Generation

Decomposes images into editable RGBA layers with open weights.
Enables precise control over semantic components during generation.
Hugging Face | Paper | Demo

https://reddit.com/link/1ptggkm/video/jq1ujox5pu8g1/player

LongVie 2 - 5-Minute Video Diffusion

Generates 5-minute continuous videos with controllable elements.
Open weights and code for extended video generation.
[Paper](https://huggingface.co/papers/2512.13604) | [GitHub](https://github.com/Vchitect/LongVie)

https://reddit.com/link/1ptggkm/video/8kr7ue8pqu8g1/player

WorldPlay(Tencent) - Interactive 3D World Generation

Generates interactive 3D worlds with geometric consistency.
Model available for local deployment.
Website | Model

https://reddit.com/link/1ptggkm/video/dggrhxqyqu8g1/player

Generative Refocusing - Depth-of-Field Control

Controls focus and depth of field in generated or existing images.
Open source implementation for bokeh and focus effects.
[Website](https://generative-refocusing.github.io/) | [Demo](https://huggingface.co/spaces/nycu-cplab/Genfocus-Demo) | [Paper](https://arxiv.org/abs/2512.16923) | [GitHub](https://github.com/rayray9999/Genfocus)

https://reddit.com/link/1ptggkm/video/a9jjbir6pu8g1/player

DeContext - Protection Against Unwanted Edits

Protects images from manipulation by diffusion models like FLUX.
Open source tool for adding imperceptible perturbations that block edits.
Website | Paper | GitHub

https://preview.redd.it/iuyeboy8pu8g1.png?width=1427&format=png&auto=webp&s=6e451e1336fcb8d5cebab46956605d42ecce8604

Flow Map Trajectory Tilting - Test-Time Scaling

Improves diffusion outputs at test time using flow maps.
Adjusts generation trajectories without retraining models.
[Paper](https://arxiv.org/abs/2511.22688) | [Website](https://flow-map-trajectory-tilting.github.io/)

https://preview.redd.it/7huqzj9bpu8g1.png?width=1140&format=png&auto=webp&s=baf5ee057c6c69d2cb1566f0a743c73419de99ad

StereoPilot - 2D to Stereo 3D

Converts 2D videos to stereo 3D with open model and code.
Full source release for VR content creation.
Website | Model | GitHub

LongCat-Video-Avatar - "An expressive avatar model built upon LongCat-Video"

[Website](https://meigen-ai.github.io/LongCat-Video-Avatar/) | [GitHub](https://github.com/meituan-longcat/LongCat-Video) | [Paper](https://arxiv.org/abs/2510.22200) | [ComfyUI](https://huggingface.co/Kijai/LongCat-Video_comfy/tree/main/Avatar)

TRELLIS 2 - 3D generative model designed for high-fidelity image-to-3D generation

Model | Demo (i saw someone playing with this in Comfy but i forgot to save the post)




Wan 2.6 was released last week but only to the API providers for now.

Checkout the full newsletter for more demos,
papers, and resources.

* Reddit post limits stopped me from adding the rest of the videos/demos.

(https://www.reddit.com/submit/?sourceid=t31ptfw0q)

https://redd.it/1ptggkm
@rStableDiffusion
Let's hope it will be Z-image base.
https://redd.it/1ptj1lo
@rStableDiffusion
What model or LoRA should I use to generate images that are closest to this style?
https://redd.it/1ptnqzp
@rStableDiffusion