Vista de Lectura

Hay nuevos artículos disponibles. Pincha para refrescar la página.

Comparison - Juggernaut SDXL - from two years ago to now. Maybe the newer models are overcooked and this makes human skin worse

Comparison - Juggernaut SDXL - from two years ago to now. Maybe the newer models are overcooked and this makes human skin worse

Early versions of SDXL, very close to the baseline, had issues like weird bokeh on backgrounds. And objects and backgrounds in general looked unfinished.

However, apparently these versions had a better skin?

Maybe the newer models end up overcooking - which is useful for scenes, objects, etc., but can make human skin look weird.

Maybe one of the problems with fine-tuning is setting different learning rates for different concepts, which I don't think is possible yet.

In your opinion, which SDXL model has the best skin texture?

submitted by /u/More_Bid_2197
[link] [comments]

So, custom SDXL models were trained using millions of joycaption tagged images BUT prompt understanding did not improve. Does anyone know why? Which SDXL models have the best prompt understanding?

There was a "legend" that SDXL and SD 1.5 were trained with wrong captions

So, the custom models used millions of images with long descriptions - but apparently it was not enough. Or it changed very little. Am I wrong?

1) Which SDXL models have better adherence to the prompt? (I am not interested in illustritious and pony)

2) Which SDXL models work best with lora?

3) Which models are best for training lora?

4) Do any models have special strengths? For example, Lustify is very good with erotic images and can do male anatomy. Are there any models who are exceptionally good with skin? Or any other strengths?

submitted by /u/More_Bid_2197
[link] [comments]

I don't know if I'm wrong. But checkpoints created from a mix of other checkpoints or loras are not suitable for training, correct? Do you train SDXL loras with the base or custom models? Which models work best for training SDXL loras ?

Until today I have always trained with the base

Some custom checkpoints were created by mixing several loras and checkpoints - as far as I know, they have strange results when applying other loras

submitted by /u/More_Bid_2197
[link] [comments]

I don't know if open source generative AI will still exist in 1 or 2 years. But I'm proud of my generations. Training a lora, adjusting the parameters, selecting a model, cfg, sampler, prompt, controlnet, workflows - I like to think of it as an art

I don't know if open source generative AI will still exist in 1 or 2 years. But I'm proud of my generations. Training a lora, adjusting the parameters, selecting a model, cfg, sampler, prompt, controlnet, workflows - I like to think of it as an art

But I don't know if everything will be obsolete soon

I remember Stable Diffusion 1.5. It's fun to read posts from people saying that dreambooth was realistic. And now 1.5 is completely obsolete. Maybe it still has some use for experimental art, exotic stuff

Models are getting too big and difficult to adjust. Maybe the future will be more specialized models

The new version of Chatgpt came out and it was a shock because people with no knowledge whatsoever can now do what was only possible with control net / ipadapter.

But even so, as something becomes too easy, it loses some of its value. For example, midjorney and gpt look the same

submitted by /u/More_Bid_2197
[link] [comments]

Is Prodigy the best option for training loras ? Or is it possible to create better loras by manually choosing the learning rate ?

apparently the only problem with the prodigy is that it loses flexibility

But in many cases this was the only efficient way I found to train and obtain similarity. Maybe other optimizers like lion and adafactor are "better" in the sense of generating something new, because they don't learn properly.

submitted by /u/More_Bid_2197
[link] [comments]

1 million questions about training. For example, if I don't use the prodigy optimizer, lora doesn't learn enough and has no facial similarity. Do people use prodigy to find the optimal learning rate and then retrain? Or is this not necessary ?

Question 1 - dreambooth vs lora, locon, loha, lokr.

Question 2 - dim and alpha.

Question 3 - learning rate and optmizers and functions (cosine, constant, cosine with restart)

I understand that it can often be difficult to say objectively which method is best.

Some methods become very similar to the data set, but they lack flexibility, which is a problem.

And this varies from model to model. Sd 1.5 and SDXL will probably never be perfect because the model has more limitations, such as small objects distorted by Vae.

submitted by /u/More_Bid_2197
[link] [comments]

Apparently, the perpetrator of the first stable diffusion hacking case (comfyui LLM vision) has been discovered by FBI and pleaded guilty (1 to 5 years sentence). Through this comfyui malware a Disney computer was hacked

https://www.justice.gov/usao-cdca/pr/santa-clarita-man-agrees-plead-guilty-hacking-disney-employees-computer-downloading

https://variety.com/2025/film/news/disney-hack-pleads-guilty-slack-1236384302/

LOS ANGELES – A Santa Clarita man has agreed to plead guilty to hacking the personal computer of an employee of The Walt Disney Company last year, obtaining login information, and using that information to illegally download confidential data from the Burbank-based mass media and entertainment conglomerate via the employee’s Slack online communications account.

Ryan Mitchell Kramer, 25, has agreed to plead guilty to an information charging him with one count of accessing a computer and obtaining information and one count of threatening to damage a protected computer.

In addition to the information, prosecutors today filed a plea agreement in which Kramer agreed to plead guilty to the two felony charges, which each carry a statutory maximum sentence of five years in federal prison.

Kramer is expected to make his initial appearance in United States District Court in downtown Los Angeles in the coming weeks.

According to his plea agreement, in early 2024, Kramer posted a computer program on various online platforms, including GitHub, that purported to be computer program that could be used to create A.I.-generated art. In fact, the program contained a malicious file that enabled Kramer to gain access to victims’ computers.

Sometime in April and May of 2024, a victim downloaded the malicious file Kramer posted online, giving Kramer access to the victim’s personal computer, including an online account where the victim stored login credentials and passwords for the victim’s personal and work accounts.

After gaining unauthorized access to the victim’s computer and online accounts, Kramer accessed a Slack online communications account that the victim used as a Disney employee, gaining access to non-public Disney Slack channels. In May 2024, Kramer downloaded approximately 1.1 terabytes of confidential data from thousands of Disney Slack channels.

In July 2024, Kramer contacted the victim via email and the online messaging platform Discord, pretending to be a member of a fake Russia-based hacktivist group called “NullBulge.” The emails and Discord message contained threats to leak the victim’s personal information and Disney’s Slack data.

On July 12, 2024, after the victim did not respond to Kramer’s threats, Kramer publicly released the stolen Disney Slack files, as well as the victim’s bank, medical, and personal information on multiple online platforms.

Kramer admitted in his plea agreement that, in addition to the victim, at least two other victims downloaded Kramer’s malicious file, and that Kramer was able to gain unauthorized access to their computers and accounts.

The FBI is investigating this matter.

submitted by /u/More_Bid_2197
[link] [comments]

Is there any method to train lora with medium/low quality images but the model does not absorb jpeg artifacts, stains, sweat ? A lora that learns the shape of a person's face/body, but does not affect the aesthetics of the model - is it possible ?

Apparently this doesn't happen with flux because the loras are always undertrained

But it happens with SDXL

I've read comments from people saying that they train a lora with SD 1.5, generate pictures and then train another one with SDXL

Or change the face or something like that

The dim/alpha can also help. apparently if the sim is too big, the blonde absorbs more unwanted data

submitted by /u/More_Bid_2197
[link] [comments]

Any new discoveries about training ? I don't see anyone talking about dora. I also hear little about loha, lokr and locon

At least in my experience locon can give better skin textures

I tested dora - the advantage is that with different subtitles it is possible to train multiple concepts, styles, people. It doesn't mix everything up. But, it seems that it doesn't train as well as normal lora (I'm really not sure, maybe my parameters are bad)

I saw dreambooth from flux and the skin textures looked very good. But it seems that it requires a lot of vram, so I never tested it

I'm too lazy to train with flux because it's slower, kohya doesn't download the models automatically, they're much bigger

I've trained many loras with SDXL but I have little experience with flux. And it's confusing for me the ideal learning rate for flux, number of steps and optimizer. I tried prodigy but bad results for flux

submitted by /u/More_Bid_2197
[link] [comments]
❌