Vista Normal

Hay nuevos artículos disponibles. Pincha para refrescar la página.
Hoy — 20 Febrero 2025Salida Principal

15k hand-curated portrait images of "a woman"

20 Febrero 2025 at 02:10

https://huggingface.co/datasets/opendiffusionai/laion2b-23ish-woman-solo

From the dataset page:

Overview

All images have a woman in them, solo, at APPROXIMATELY 2:3 aspect ratio. (and at least 1200 px in length)
Some are just a little wider, not taller. Therefore, they are safe to auto crop to 2:3

These images are HUMAN CURATED. I have personally gone through every one at least once.

Additionally, there are no visible watermarks, the quality and focus are good, and it should not be confusing for AI training

There should be a little over 15k images here.

Note that there is a wide variety of body sizes, from size 0, to perhaps size 18

There are also THREE choices of captions: the really bad "alt text", then a natural language summary using the "moondream" model, and then finally a tagged style using the wd-large-tagger-v3 model.

submitted by /u/lostinspaz
[link] [comments]

What would you consider to be the most significant things that AI Image models cannot do right now (without significant effort)?

20 Febrero 2025 at 10:48

Here's my list:

  • Precise control of eyes / gaze
    • Even with inpainting, this can be nearly impossible
  • Precise control of hand placement and gestures, unless it corresponds to a well known particular pose
  • Lighting control
    • Some models can handle "Dark" and "Blue Light" and such, but precise control is impossible without inpainting (and even with inpainting, it's hard)
  • Precise control of the camera
    • Most models can do "Close-up", "From above", "Side view", etc... but specific zooms and angles that are not just 90 degree rotations, are very difficult and require a great deal of luck to achieve

Thoughts?

submitted by /u/_BreakingGood_
[link] [comments]

Illustrious/NoobAI full model fine-tuning project

20 Febrero 2025 at 00:42

Hello!

I want to fine-tune an Illustrious/NoobAI base model (checkpoint) with a few hundreds/thousands images, so that it will be able to reproduce styles like Arcane, Incase, Bancin, CptPopcorn and many more out of the box. Also I want to "westernize" the model so that it could produce european/american faces/styles aswell, because it really gets boring to see only anime-like images everywhere - and they almost look like they have the same style.

I looked for some training parameters/settings, but I couldn't find anything for Illu/NoobAI fine-tuning. I even downloaded some of the best "trained" Illu/NoobAI models from Civitai and I inspected their metadata and everything and guess what. They weren't even "trained/fine-tuned" but only merged or having injected LoRAs into them. So there are lots of liars on civitai.

I know for sure that by fine-tuning you reach the maximum quality possible, that's why I don't want to train LoRAs and inject them afterwards into the checkpoint.

I have access to some 24-48 GB VRAM GPUs.

Kohya SS GUI settings/parameters are appreciated as I'm more familiar with this (or kohya ss scripts).

Thanks!

The people wanting/willing to help or to contribute to this project (and I mean being a part of it, not contributing monetarily) with knowledge and other ideas are welcomed!

Let's make a community fine-tune better than what we have right now!

Discord: tekeshix_46757
Gmail: [tekeshix1@gmail.com](mailto:tekeshix1@gmail.com)

Edit: Not LoRA training, not Dreambooth training but only full fine-tuning.

Dreambooth is better than LoRA, but still inferior to full fine-tune.

submitted by /u/TekeshiX
[link] [comments]

Need help in gpu choice

20 Febrero 2025 at 09:09

Soo I played with ai and find out that I love tinkering with it and that my 1070 gpu is really bad at it. I want to understand what's better for me from this criteria: mainly gaming but don't really play AAA titles have 1080p monitor and want to switch to 1440p 240hrz (mostly for fps marvel rivals rn), want to tinker with ai and to do it faster than flicking 1min for 512x512 img. And want to try flux donw the road. What I was considering: - used 3090 - 4080 supper - the less likely 4090 - is there any chance to go for amd?

What to hear any pros and cons, any suggestions etc Ty

submitted by /u/yar4ik
[link] [comments]

Why is Flux "schnell" so much slower than SDXL?

20 Febrero 2025 at 11:04
Why is Flux "schnell" so much slower than SDXL?

I'm new to image generation, i started with comfyui, and I'm using flux schnell model and sdxl.
I heard everywhere, including this subreddit that flux is supposed to be very fast but I've had a very different experience.

Flux Schnell is incredibly slow,
for example, I used a simple prompt
"portrait of a pretty blonde woman, a flower crown, earthy makeup, flowing maxi dress with colorful patterns and fringe, a sunset or nature scene, green and gold color scheme"
and I got the following results

https://preview.redd.it/rzky9t860ake1.png?width=403&format=png&auto=webp&s=67471f77fabf4600ecb2fc2161f25355c7793f5a

Am I doing something wrong? I'm using the default workflows given in comfyui.

EDIT:
A sensible solution:
Use q4 models available at
flux1-schnell-Q4_1.gguf · city96/FLUX.1-schnell-gguf at main
and follow (5) How to Use Flux GGUF Files in ComfyUI - YouTube
to setup

submitted by /u/BeetranD
[link] [comments]

Anime sitting pose with stretched legs

20 Febrero 2025 at 14:24
Anime sitting pose with stretched legs

Look guys this is only a minor thing but I am LOSING my mind over it. I cant get a single image of an anime character sitting while having their legs stretched out. I tried everything. Right now even openpose decide to completly ignore everything.

https://preview.redd.it/l8un8oq30bke1.png?width=902&format=png&auto=webp&s=69acae11aab2d50a323e34d3865efec3abfc6f38

submitted by /u/Daedalus_the_wise
[link] [comments]

Outpainting Continuity Issue in Flux Fill Pro

20 Febrero 2025 at 08:09
Outpainting Continuity Issue in Flux Fill Pro

Hey everyone,

I'm experiencing an issue with Flux Fill Pro when using the outpainting function from the original API of black forest labs via replicate. Instead of smoothly extending the image, the AI generates two completely different scenes instead of naturally continuing the background.

Interestingly, when we use x1.5 and x2 scaling, the expansion works correctly without breaking the continuity. However, when selecting Right, Top, Left, or Bottom, the AI seems to lose coherence and creates new elements that don't follow the original composition.

We've tried several adjustments to fix the issue, including:

  • Modifying the prompt to ensure the AI maintains the lighting, colors, and composition of the original image: "Extend the image while maintaining the lighting, colors and composition. Continue existing elements without adding new scenes."
  • Adjusting guidance (from 60 to high and low levels) to balance adherence and flexibility.
  • Changing diffusion steps to test differences in detail levels.
  • Using a mask with smooth transitions to avoid abrupt cuts.
  • Reducing the expansion area and making small iterations instead of a single large expansion.

Despite these efforts, the problem still occurs when using Right, Top, Left, or Bottom.

Has anyone else encountered this issue? Any ideas on how to fix it? 🚀

Thanks in advance for your help!

https://preview.redd.it/m0rrop6559ke1.png?width=462&format=png&auto=webp&s=8c028bf0d33d83f850a2165e0601aa64b95c37a1

submitted by /u/FirstWorld1541
[link] [comments]

Creating Different Poses in Same Environment with ControlNet

20 Febrero 2025 at 13:34

Is there a way to generate a character with different poses, but in the same environment? Currently, I am using ControlNet generate some characters that mimic the pose similar to a reference image.

However, background environment will always slightly change a bit for every run, even though I have a detailed prompt about the environment. I would like to have the same background for each run. I tried searching online but couldn't find anything similar to this.

submitted by /u/VickNicks
[link] [comments]
❌
❌