Me trying to test every new AI video model
![]() | submitted by /u/runboli [link] [comments] |
![]() | submitted by /u/idk155 [link] [comments] |
https://huggingface.co/datasets/opendiffusionai/laion2b-23ish-woman-solo
From the dataset page:
All images have a woman in them, solo, at APPROXIMATELY 2:3 aspect ratio. (and at least 1200 px in length)
Some are just a little wider, not taller. Therefore, they are safe to auto crop to 2:3
These images are HUMAN CURATED. I have personally gone through every one at least once.
Additionally, there are no visible watermarks, the quality and focus are good, and it should not be confusing for AI training
There should be a little over 15k images here.
Note that there is a wide variety of body sizes, from size 0, to perhaps size 18
There are also THREE choices of captions: the really bad "alt text", then a natural language summary using the "moondream" model, and then finally a tagged style using the wd-large-tagger-v3 model.
Here's my list:
Thoughts?
![]() | submitted by /u/Wrong_Rip5185 [link] [comments] |
Hello!
I want to fine-tune an Illustrious/NoobAI base model (checkpoint) with a few hundreds/thousands images, so that it will be able to reproduce styles like Arcane, Incase, Bancin, CptPopcorn and many more out of the box. Also I want to "westernize" the model so that it could produce european/american faces/styles aswell, because it really gets boring to see only anime-like images everywhere - and they almost look like they have the same style.
I looked for some training parameters/settings, but I couldn't find anything for Illu/NoobAI fine-tuning. I even downloaded some of the best "trained" Illu/NoobAI models from Civitai and I inspected their metadata and everything and guess what. They weren't even "trained/fine-tuned" but only merged or having injected LoRAs into them. So there are lots of liars on civitai.
I know for sure that by fine-tuning you reach the maximum quality possible, that's why I don't want to train LoRAs and inject them afterwards into the checkpoint.
I have access to some 24-48 GB VRAM GPUs.
Kohya SS GUI settings/parameters are appreciated as I'm more familiar with this (or kohya ss scripts).
Thanks!
The people wanting/willing to help or to contribute to this project (and I mean being a part of it, not contributing monetarily) with knowledge and other ideas are welcomed!
Let's make a community fine-tune better than what we have right now!
Discord: tekeshix_46757
Gmail: [tekeshix1@gmail.com](mailto:tekeshix1@gmail.com)
Edit: Not LoRA training, not Dreambooth training but only full fine-tuning.
Dreambooth is better than LoRA, but still inferior to full fine-tune.
like prompt-to-image inference with easy deployment
Soo I played with ai and find out that I love tinkering with it and that my 1070 gpu is really bad at it. I want to understand what's better for me from this criteria: mainly gaming but don't really play AAA titles have 1080p monitor and want to switch to 1440p 240hrz (mostly for fps marvel rivals rn), want to tinker with ai and to do it faster than flicking 1min for 512x512 img. And want to try flux donw the road. What I was considering: - used 3090 - 4080 supper - the less likely 4090 - is there any chance to go for amd?
What to hear any pros and cons, any suggestions etc Ty
![]() | I'm new to image generation, i started with comfyui, and I'm using flux schnell model and sdxl. Flux Schnell is incredibly slow, Am I doing something wrong? I'm using the default workflows given in comfyui. EDIT: [link] [comments] |
![]() | Look guys this is only a minor thing but I am LOSING my mind over it. I cant get a single image of an anime character sitting while having their legs stretched out. I tried everything. Right now even openpose decide to completly ignore everything. [link] [comments] |
![]() | Hey everyone, I'm experiencing an issue with Flux Fill Pro when using the outpainting function from the original API of black forest labs via replicate. Instead of smoothly extending the image, the AI generates two completely different scenes instead of naturally continuing the background. Interestingly, when we use x1.5 and x2 scaling, the expansion works correctly without breaking the continuity. However, when selecting Right, Top, Left, or Bottom, the AI seems to lose coherence and creates new elements that don't follow the original composition. We've tried several adjustments to fix the issue, including:
Despite these efforts, the problem still occurs when using Right, Top, Left, or Bottom. Has anyone else encountered this issue? Any ideas on how to fix it? 🚀 Thanks in advance for your help! [link] [comments] |
Lately skyeels seem to be the latest one, is it the best?
Couple weeks ago I saw unofficial hunyuan i2v support. Are those better?
Link me workflows/threads to follow like an ape :3
Is there a way to generate a character with different poses, but in the same environment? Currently, I am using ControlNet generate some characters that mimic the pose similar to a reference image.
However, background environment will always slightly change a bit for every run, even though I have a detailed prompt about the environment. I would like to have the same background for each run. I tried searching online but couldn't find anything similar to this.