Vista de Lectura

Hay nuevos artículos disponibles. Pincha para refrescar la página.

2DN NAI - highly detailed NoobAI v-pred model

2DN NAI - highly detailed NoobAI v-pred model

I thought I’d share my new model, which consistently produces really detailed images.

After spending over a month coaxing NoobAI v-pred v1 into producing more coherent results+ I used my learnings to make a more semi-realistic version of my 2DN model

CivitAI link: https://civitai.com/models/520661

Noteworthy is that all of the preview images on CivitAI use the same settings and seed! So I didn’t even cherry pick from successive random attempts. I did reject some prompts for being boring or too samey to the other gens, that’s all.

I hope people find this model useful, it really does a variety of stuff, without being pigeonholed into one look. It uses all of the knowledge of NoobAI’s insane training but with more details, realism and coherency. It can be painful to first use a v-pred model, but they do way richer colours and wider tonality. Personally I use reForge after trying just about everything.

submitted by /u/advo_k_at
[link] [comments]

Full Breakdown: The bghira/Simpletuner Situation

I wanted to provide a detailed timeline of recent events concerning bghira, the creator of the popular LoRA training tool, Simpletuner. Things have escalated quickly, and I believe the community deserves to be aware of the full situation.

TL;DR: The creator of Simpletuner, bghira, began mass-reporting NotSFW LoRAs on Hugging Face. When called out, he blocked users, deleted GitHub issues exposing his own project's severe license violations, and took down his repositories. It was then discovered he had created his own NotSFW FLUX LoRA (violating the FLUX license), and he has since begun lashing out with taunts and false reports against those who exposed his actions.

Here is a clear, chronological breakdown of what happened:


  1. 2025-07-04 13:43: Out of nowhere, bghira began to spam-report dozens of NotSFW LoRAs on Hugging Face.

  2. 2025-07-04 17:44: u/More_Bid_2197 called this out on the StableDiffusion subreddit.

  3. 2025-07-04 21:08: I saw the post and tagged bghira in the comments asking for an explanation. I was promptly blocked without a response.

  4. Following this, I looked into the SimpleTuner project itself and noticed it severely broke the AGPLv3 and Apache 2.0 licenses it was supposedly using.

  5. 2025-07-04 21:40: I opened a GitHub issue detailing the license violations and started a discussion on the Hugging Face repo as well.

  6. 2025-07-04 22:12: In response, bghira deleted my GitHub issue and took down his entire Hugging Face repository to hide the reports (many other users had begun reporting it by this point).

  7. bghira invalidated his public Discord server invite to prevent people from joining and asking questions.

  8. 2025-07-04 21:21: Around the same time, u/atakariax started a discussion on the StableTuner repo about the problem. bghira edited the title of the discussion post to simply say "Simpletuner creator is based".

  9. I then looked at bghira's Civitai profile and discovered he had trained and published an NotSFW LoRA for the new FLUX model. This is not only hypocritical but also a direct violation of FLUX's license, which he was enforcing on others.

  10. I replied to some of bghira's reports on Hugging Face, pointing out his hypocrisy. I received these two responses:

    2025-07-05 12:15: In response to one comment:

    i think it's sweet how much time you spent learning about me yesterday. you're my number one fan!

    2025-07-05 12:14: In response to another:

    oh ok so you do admit all of your stuff breaks the license, thanks technoweenie.

  11. 2025-07-05 14:55: bghira filed a false report against one of my SD1.5 models for "Trained on illegal content." This is objectively untrue; the model is a merge of models trained on legal content and contains no additional training itself. This is another example of his hypocrisy and retaliatory behavior.

  12. 2025-07-05 16:18: I have reported bghira to Hugging Face for harassment, name-calling, and filing malicious, false reports.

  13. 2025-07-05 17:26: A new account has appeared with the name EnforcementMan (likely bghira), reporting Chroma.


I'm putting this all together to provide a clear timeline of events for the community.

Please let me know if I've missed something.

(And apologies if I got some of the timestamps wrong, timezones are a pain).

Mirror of this post in case this gets locked: https://www.reddit.com/r/comfyui/comments/1lsfodj/full_breakdown_the_bghirasimpletuner_situation/

submitted by /u/TechnoByte_
[link] [comments]

How come there isn’t a popular peer-to-peer sharing community to download models as opposed to Huggingface and Civitai?

Is there a technical reason why the approach to hoarding and sharing models hasn’t gone the p2p route? That seems to be the best way to protect the history of these models and get around all the censorship concerns.

Or does this exist already and it’s just not popular yet?

submitted by /u/mccoypauley
[link] [comments]

No humans needed: AI generates and labels its own training data

No humans needed: AI generates and labels its own training data

We’ve been exploring how to train AI without the painful step of manual labeling—by letting the system generate its own perfectly labeled images.

The idea: start with a 3D mesh of a human body, render it photorealistically, and automatically extract all the labels (like body points, segmentation masks, depth, etc.) directly from the 3D data. No hand-labeling, no guesswork—just pixel-perfect ground truth every time.

Here’s a short video showing how it works.

Let me know what you think—or how you might use this kind of labeled synthetic data.

submitted by /u/YuriPD
[link] [comments]

What's up with Pony 7?

The lack of any news over the past few months can't help but give rise to unpleasant conclusions. In the official Discord channel, everyone who comes to inquire about the situation and the release date gets a stupid joke about "two weeks" in response. Compare this with Chroma, where the creator is always in touch, and everyone sees a clear and uninterrupted roadmap.

I think that Pony 7 was most likely a failure and AstraliteHeart simply does not want to admit it. The situation is similar to Virt-A-Mate 2.0, where after a certain time, people were also fed vague dates and the release was delayed under various formulations, and in the end, something disappointing came out, barely even pulling for alpha.

It could easily happen that when Pony comes out, it will be outdated and no one needs it.

submitted by /u/from_monitor
[link] [comments]

I documented every Fooocus 2.5.5 setting for 6 months. Then taught AI to configure them perfectly.

I documented every Fooocus 2.5.5 setting for 6 months. Then taught AI to configure them perfectly.

Look at this portrait. First generation, 40 seconds. Zero post-processing.

Getting here took 6 months of documenting every Fooocus setting, model comparison, and LoRA weight that actually matters.

Then I built HyperFocus-AI - a meta-prompt that turns any LLM into a Fooocus 2.5.5 configuration expert.

What HyperFocus-AI does:

My simple input: "A young woman with pink hair and freckles"

What it returned:

Enhanced Prompt: "A young woman with pastel pink hair and soft freckles across her cheeks and nose, facing the camera with a calm, serene expression, hyper-detailed skin texture with visible pores and fine peach fuzz, soft window light from the side, smooth bokeh background, natural makeup, delicate eyelashes, subtle catchlights in the eyes, captured with an 85mm lens at f/1.4, ultra high-resolution portrait photography style"

Complete Technical Configuration:

  • Base Model: RealVisXL_V5.0
  • LoRA Stack: Add_Details_v1.2.safetensors (0.5), PerfectEyesXL.safetensors (0.3), warm_light_style_v2 (0.5)
  • CFG Scale: 7.5
  • Image Sharpness: 5.0
  • Sampler: dpmpp 2m sde gpu
  • Selected Styles: ["Fooocus V2", "Fooocus Photograph", "Fooocus Enhance"]

Plus: Carefully crafted negative prompt to eliminate common artifacts

First generation using these exact settings. This level of photorealism used to take me 50+ attempts.

Why this changes everything:

The prompt transformed my basic idea into:

  • Professional photography language
  • Specific technical details (85mm f/1.4)
  • Skin texture optimization
  • Lighting direction
  • Every parameter calculated

No more:

  • "Which model for portraits?"
  • "What CFG scale?"
  • Plastic-looking skin
  • Wasted generations

The Meta-Prompt:

[Fooocus 2.5.5 Optimizer]

How to use:

  1. Copy HyperFocus-AI into ChatGPT
  2. Give it any image idea
  3. Copy the complete configuration into Fooocus 2.5.5
  4. Generate photorealistic perfection

Pro Tip: For optimal results, tell HyperFocus-AI which base models and LoRAs you have installed. It will adapt its recommendations to YOUR available resources rather than suggesting models you don't have.

Not selling anything. No course. Just sharing 6 months of research.

P.S. - From "pink hair and freckles" to museum-quality portrait settings. This is what systematic optimization looks like.

P.P.S. - Yes, I'm using AI to optimize AI. Welcome to 2025.

submitted by /u/Prestigious-Fan118
[link] [comments]

The Fat Rat - Myself & I - AI Music Video

The Fat Rat - Myself & I - AI Music Video

a video I've made for a uni assignment Decided to make another music video this time about a song from "The Fat Rat" it does basically include almost all of the new stuff that came out in the last 3 or 4 months, up until the day FusionX got released i've used:

  • Flux distilled with some loras,
  • Wan T2V, I2V, Diffusion Forcing, VACE Start End Frame, Fun Style Transfer, Camera Loras,
  • Adiff with AudioReact,
submitted by /u/Small_Light_9964
[link] [comments]

Am I Missing Something? No One Ever Talks About F5-TTS, and it's 100% Free + Local and > Chatterbox

I see Chatterbox is the new/latest TTS tool people are enjoying, however F5-TTS has been out for awhile now and I still think it sounds better and more accurate with one-shot voice cloning, yet people rarely bring it up? You can also do faux podcast style outputs with multiple voices if you generate a script with an LLM (or type one up yourself). Chatterbox sounds like an exaggerated voice actor version of the voice you are trying to replicate yet people are all excited about it, I don't get what's so great about it

submitted by /u/StuccoGecko
[link] [comments]

V2V workflow for improving quality?

Hi there, I hope you can help me.
TLDR: I have a video of different clips stitched together. The fact that they are different clips make the actors in the clips move in a weird way. Is there a way to give a V2V the clip and make it have more coherent movements, while preserving the likeness and outfit of the character, possibly improving the overall quality too?

Lately with Kontext I started experimenting with I2V with first and last frame guidance, it is great!
I can upload an image of my DnD warrior to Kontext and create another image of him surprised if front of a dragon, then create an animation from those key frames. I noticed that unfortunately if the two images are too different the model does not understand the request well, so I have to create many 2 seconds long videos with different key frames.
Doing so, though, makes the character move in short bursts of movement, and the final result is weird to watch.
Is there a way to feed the final video to a Video to Video model (WAN, HY, anything is fine, I don't care if it is censored or not) and have it recreate the scene with more coherent movements? Also, if I manage to create such a video, would it be possible to enhance the quality / resolution?

Thanks in advance :)

submitted by /u/Tomorrow_Previous
[link] [comments]
❌