Training and inference codes here: https://github.com/NVlabs/Sana
Note: I still can't find the model on huggingface yet, according to the code, it should be on hugging face under Efficient-Large-Model/Sana_1600M_1024px
[link] [comments]
Training and inference codes here: https://github.com/NVlabs/Sana
Note: I still can't find the model on huggingface yet, according to the code, it should be on hugging face under Efficient-Large-Model/Sana_1600M_1024px
I'm getting into YouTube and wanted to find a free TTS software that doesn't sound robotic and then learned about RVCs.I know about Elevenlabs but paying for anything isn't an option for me right now.
So I'm just looking for a fully free RVC AI software that I can use to copy & combine voices and then use TTS to generate basically an uncapped number of words.
Thanks.
submitted by /u/Dismal-Rich-7469 [link] [comments] |
Is there an updated beginner friendly tutorial that actually teaches you everything you need to know about ComfyUI, i watch everyday tutorials people showing workflows on useful ways you can use comfy, don't want to just follow along, i want to understand be able to adjust the workflow, know what am doing. Any recommendations?
Hi guys! I keep getting these weird lines when I try to resize (as in, make larger) images in img to img when using Flux. The checkpoint is flux1-dev-bnb-v2. The prompt I used for this one was: a mother kneels on a blanket holding her small child at sundown, hyperdetailed, analog style, subsurface scattering, realistic, heavy shadow, masterpiece, best quality Any suggestions much appreciated! As I say, it’s been happening loads [link] [comments] |
Error message below. Reinstalling xformers does not help. Also, the file in question is present at the place in the error message. I feel like I'm following every single guide available to the letter and anyway end up with a failure. Also, trying to run SD on rx 5700xt.
The procedure entry point ?dtype@TensorOptions@c10@@QEBA?AU12@V?$optional@W4ScalarType@c10@@@std@@@Z could not be located in the dynamic link library C:\Users\vinmai\Desktop\stable-diffusion-webui-amdgpu\venv\Lib\site-packages\xformers\_C.pyd
I just wanted to download this model and stumbled across this "Suspicious" file. What is this all about? Is it safe to run this model locally on my machine? How bad is this "Architectural Backdoor" and what does it do? [link] [comments] |
I'm running some AI services in H100 runpod for a while, but lately runpod usually says H100 is unavailable is there any options? [link] [comments] |
Hi guys, forgive me if I sound like a noob but this is all new and exciting for me to learn SD. It took me a couple of days to install and made it work, I'm not a technical person but Googling all the errors and updating/installing stuff seems to help. Anyways my issue is how to save a larger image than I one I bring it? I'm using img2img, my original is ~500px x 375px, if I want the final image to be 5000px wide, how can I do that? Do I put the size in the "resize" section? I am still testing different sampling methods and they take hours for one image lol. I want see if outing larger image size is possible and how to do it the right way. Thanks for any help on this! image - first thing I tested! So cute ;) [link] [comments] |
That's the only UI I manage to get running (I have AMD GPU), and I can't wait to use it
Hey StableDiffusion community! Been getting into FLUX finetuning recently and noticing a lot of the output images are stretched, particularly the faces. Wondering what the root cause of this is and what I can do to resolve it? Thanks for any tips/explanations!
Im trying to get stable diffusion 3.5 running. I followed the instructions on git but when trying to generate an image is gives a ssl certificate error. I cant find a solid answer to this. This is my first time trying to run SD locally. Im quite new and would appreciate help. For reference I'm running python 10. Using brave. Sd 3.5 checkpoint. Thanks.
I need to build a PC to train SDXL using Kohya (and generate a video using warpfusion). The machine we currently have uses an A5000 with 24GB of VRAM - however, it's been kinda cramped for our training (that's a whole other story) - and we share it with others. I need at least 30-32 GB of VRAM to avoid spilling into shared memory (which TANKS performance), but I also wanna keep the budget reasonable
Does it make sense to replace an A5000 with two A4000s or A4500s? Using Accelerate we should be able to combine them and get more than enough memory, but I need to know if they'd be too inferior to the A5000 and by how much
EDIT
Additional questions: does this make sense in general? are there GPUs more sutited to work in parallel/train SDXL?