[Release] New ComfyUI Node – Maya1_TTS 🎙️
Hey everyone! Just dropped a new ComfyUI node I've been working on – **ComfyUI-Maya1\_TTS** 🎙️
[https://github.com/Saganaki22/-ComfyUI-Maya1\_TTS](https://github.com/Saganaki22/-ComfyUI-Maya1_TTS)
This one runs the Maya1 TTS 3B model, an expressive voice TTS directly in ComfyUI. It's 1 all-in-one (AIO) node.
https://preview.redd.it/on7otvl7fizf1.png?width=1426&format=png&auto=webp&s=288d3e2ee0081fd789c7ae3c13f305f238e8a6e8
**What it does:**
* Natural language voice design (just describe the voice you want in plain text)
* 17+ emotion tags you can drop right into your text: `<laugh>`, `<gasp>`, `<whisper>`, `<cry>`, etc.
* Real-time generation with decent speed (I'm getting \~45 it/s on a 5090 with bfloat16 + SDPA)
* Built-in VRAM management and quantization support (4-bit/8-bit if you're tight on VRAM)
* Works with all ComfyUI audio nodes
**Quick setup note:**
* Flash Attention and Sage Attention are *optional* – use them if you like to experiment
* If you've got less than 10GB VRAM, I'd recommend installing `bitsandbytes` for 4-bit/8-bit support. Otherwise float16/bfloat16 works great and is actually faster.
Also, you can pair this with my [**dotWaveform node**](https://github.com/Saganaki22/ComfyUI-dotWaveform) if you want to visualize the speech output.
[Realistic male voice in the 30s age with american accent. Normal pitch, warm timbre, conversational pacing.](https://reddit.com/link/1oph2fi/video/w0ayr8gqiizf1/player)
[Realistic female voice in the 30s age with british accent. Normal pitch, warm timbre, conversational pacing.](https://reddit.com/link/1oph2fi/video/kal929sriizf1/player)
The README has a bunch of character voice examples if you need inspiration. Model downloads from HuggingFace, everything's detailed in the repo.
If you find it useful, toss the project a ⭐ on GitHub – helps a ton! 🙌
https://redd.it/1oph2fi
@rStableDiffusion
Hey everyone! Just dropped a new ComfyUI node I've been working on – **ComfyUI-Maya1\_TTS** 🎙️
[https://github.com/Saganaki22/-ComfyUI-Maya1\_TTS](https://github.com/Saganaki22/-ComfyUI-Maya1_TTS)
This one runs the Maya1 TTS 3B model, an expressive voice TTS directly in ComfyUI. It's 1 all-in-one (AIO) node.
https://preview.redd.it/on7otvl7fizf1.png?width=1426&format=png&auto=webp&s=288d3e2ee0081fd789c7ae3c13f305f238e8a6e8
**What it does:**
* Natural language voice design (just describe the voice you want in plain text)
* 17+ emotion tags you can drop right into your text: `<laugh>`, `<gasp>`, `<whisper>`, `<cry>`, etc.
* Real-time generation with decent speed (I'm getting \~45 it/s on a 5090 with bfloat16 + SDPA)
* Built-in VRAM management and quantization support (4-bit/8-bit if you're tight on VRAM)
* Works with all ComfyUI audio nodes
**Quick setup note:**
* Flash Attention and Sage Attention are *optional* – use them if you like to experiment
* If you've got less than 10GB VRAM, I'd recommend installing `bitsandbytes` for 4-bit/8-bit support. Otherwise float16/bfloat16 works great and is actually faster.
Also, you can pair this with my [**dotWaveform node**](https://github.com/Saganaki22/ComfyUI-dotWaveform) if you want to visualize the speech output.
[Realistic male voice in the 30s age with american accent. Normal pitch, warm timbre, conversational pacing.](https://reddit.com/link/1oph2fi/video/w0ayr8gqiizf1/player)
[Realistic female voice in the 30s age with british accent. Normal pitch, warm timbre, conversational pacing.](https://reddit.com/link/1oph2fi/video/kal929sriizf1/player)
The README has a bunch of character voice examples if you need inspiration. Model downloads from HuggingFace, everything's detailed in the repo.
If you find it useful, toss the project a ⭐ on GitHub – helps a ton! 🙌
https://redd.it/1oph2fi
@rStableDiffusion
AI communities be cautious ⚠️ more scams will poping up using specifically Seedream models
This is an just awareness post.
Warning newcomers to be cautious of them,
Selling some courses on prompting, I guess
https://redd.it/1opn965
@rStableDiffusion
This is an just awareness post.
Warning newcomers to be cautious of them,
Selling some courses on prompting, I guess
https://redd.it/1opn965
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Mixed Precision Quantization System in ComfyUI most recent update
https://redd.it/1opw64u
@rStableDiffusion
https://redd.it/1opw64u
@rStableDiffusion
Infinite Length AI Videos with no Color Shift (Wan2.2 VACE-FUN)
https://youtu.be/f82CZl23OOo
https://redd.it/1oq0xgl
@rStableDiffusion
https://youtu.be/f82CZl23OOo
https://redd.it/1oq0xgl
@rStableDiffusion
YouTube
Create Infinite Length AI Videos with Wan VACE-FUN | No Jerky Motion, No Color Shifting
For consulting and business inquiries, email the.artofficial.trainer@gmail.com — revolutionize your pipeline with next-generation AI video tools before the rest of the industry catches on!
Taking advantage of Wan-Fun VACE 2.2’s advanced video extension properties…
Taking advantage of Wan-Fun VACE 2.2’s advanced video extension properties…
Has anyone tried the newer video model Longcat yet?
Hugging Face: https://huggingface.co/meituan-longcat/LongCat-Video
GitHub: https://github.com/meituan-longcat/LongCat-Video
Would be nice to have some more examples.
https://redd.it/1oq7egc
@rStableDiffusion
Hugging Face: https://huggingface.co/meituan-longcat/LongCat-Video
GitHub: https://github.com/meituan-longcat/LongCat-Video
Would be nice to have some more examples.
https://redd.it/1oq7egc
@rStableDiffusion
huggingface.co
meituan-longcat/LongCat-Video · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Thank you SD sub
I just really wanted to say thank you to all of you folks in here who have been so helpful and patient and amazing regardless of anyone's knowledge level.
This sub is VERY different from "big reddit" in that most everyone here is civil and does not gate-keep knowledge. In this day and age, that is rare.
Context:
I was in the middle of creating a workflow to help test a prompt with all of the different sampler and scheduler possibilities. I was thinking through how to connect and I remade the workflow a few times until I figured out how to do it while reusing as few nodes as possibles, then using less visible wires, etc etc.
Anyway, I paused and I realized I just hit my 2 month mark of using ComfyUI and AI in general, outside of ChatGPT. When I first started ComfyUI seemed incredibly complex and I thought, "there's no way I'm going to be able to make my own workflows, I'll just spend time searching for other people's workflows that match what I want instead". But now it's no problem and far better because I understand the workflow I'm creating.
I just wanted to thank you all for helping me get here so fast.
Thanks fam.
https://redd.it/1oq9fzi
@rStableDiffusion
I just really wanted to say thank you to all of you folks in here who have been so helpful and patient and amazing regardless of anyone's knowledge level.
This sub is VERY different from "big reddit" in that most everyone here is civil and does not gate-keep knowledge. In this day and age, that is rare.
Context:
I was in the middle of creating a workflow to help test a prompt with all of the different sampler and scheduler possibilities. I was thinking through how to connect and I remade the workflow a few times until I figured out how to do it while reusing as few nodes as possibles, then using less visible wires, etc etc.
Anyway, I paused and I realized I just hit my 2 month mark of using ComfyUI and AI in general, outside of ChatGPT. When I first started ComfyUI seemed incredibly complex and I thought, "there's no way I'm going to be able to make my own workflows, I'll just spend time searching for other people's workflows that match what I want instead". But now it's no problem and far better because I understand the workflow I'm creating.
I just wanted to thank you all for helping me get here so fast.
Thanks fam.
https://redd.it/1oq9fzi
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community