Vista Normal

Hay nuevos artículos disponibles. Pincha para refrescar la página.
Hoy — 9 Julio 2025StableDiffusion

What better way to test Multitalk and Wan2.1 than another Will Smith Spaghetti Video

9 Julio 2025 at 05:26
What better way to test Multitalk and Wan2.1 than another Will Smith Spaghetti Video

Wanted try make something a little more substantial with Wan2.1 and multitalk and some Image to Vid workflows in comfy from benjiAI. Ended up taking me longer than id like to admit.

Music is Suno. Used Kontext and Krita to modify and upscale images.

I wanted more slaps in this but A.I is bad at convincing physical violence still. If Wan would be too stubborn I was sometimes forced to use hailuoai as a last resort even though I set out for this be 100% local to test my new 5090.

Chatgpt is better at body morphs than kontext and keeping the characters facial likeness. There images really mess with colour grading though. You can tell whats from ChatGPT pretty easily.

submitted by /u/prean625
[link] [comments]

Flux Kontext Workflow

9 Julio 2025 at 11:53
Flux Kontext Workflow

Workflow: https://pastebin.com/HaFydUvK

Came across a bunch of different Kontext workflows and I tried to combine the best of all here!

Notably, u/DemonicPotatox showed us the node "Flux Kontext Diff Merge" that will preserve the quality when the image is reiterated (Output image is taken as input) over and over again.

Another important node is "Set Latent Noise Mask" where you can mask the area you wanna change. It doesnt sit well with Flux Kontext Diff Merge. So I removed the default flux kontext image rescaler (yuck) and replaced it with "Scale Image (SDXL Safe)".

Ofcourse, this workflow can be improved, so if you can think of something, please drop a comment below.

submitted by /u/RickyRickC137
[link] [comments]

Lets's discuss LORA naming standardization proposal. Calling all lora makers.

9 Julio 2025 at 07:06

Hey guys , I want to suggest a format for lora naming for easier and self-sufficient use. Format is:

{trigger word}_{lora name}V{lora version}_{base model}.{format}

For example- Version 12 of A lora with name crayonstyle.safetensors for sdxl with trigger word cray0ns would be:

cray0ns_crayonstyleV12_SDXL.safetensors

Note:- {base model} could be- SD15, SDXL, PONY, ILL, FluxD, FluxS, FluxK, Wan2 etc. But it MUST be standardized with agreements within community.

"any" is a special trigger word which is for loras dont have any trigger words. For example: any_betterhipsV3_FluxD.safetensors

By naming your lora like this. There are many benefits:

  1. Self-sufficient names. No need to rely on external sites or metadata for general use.

  2. Trigger words are included in lora. "any" is a special trigger word for lora which dont need any trigger words.

  3. If this style catches on, it will lead to loras with concise and to the point trigger words.

  4. Easier management of loras. No need to make multiple directories for multiple base models.

  5. Changes can be made to Comfyui and other apps to automatically load loras with correct trigger words. No need to type.

submitted by /u/Diskkk
[link] [comments]

Real HDRI with Flux Kontext

8 Julio 2025 at 23:10
Real HDRI with Flux Kontext

Really happy with how it turned out. Workflow is in the first image - it produces 3 exposures from a text prompt, which can then be combined in Photoshop into HDR. Works for pretty much anything - sunlight, overcast, indoor, night time

Workflow uses standard nodes, except for GGUF and two WAS suite nodes used to make an overexposed image. For whatever reason, Flux doesn't know what "overexposed" means and doesn't make any changes without it.

LoRA used in the workflow https://civitai.com/models/682349?modelVersionId=763724

submitted by /u/ifilipis
[link] [comments]

train loras on community models?

9 Julio 2025 at 06:05

hi,

  • what do you guys use to train your loras on community models? eg cyberrealistic pony i will mainly need XL fine tuned models.

i saw some use onetrainer, or kohya. personally i can’t use kohya locally.

  • you guys train in cloud, if yes, is it like a kohya on colab?
submitted by /u/DragonsWFlamingPearl
[link] [comments]

Wan 2.1 vs Flux Dev for posing/Anatomy

9 Julio 2025 at 00:47
Wan 2.1 vs Flux Dev for posing/Anatomy

Order: Flux sitting on couch with legs crossed (4X) -> Wan sitting on couch with legs crossed (4X), Flux Ballerina with leg up (4X)-> Wan Ballerina with leg up (4X)

I cant speak for anyone else but Wan2.1 as an image model flew clean under my radar until yanokushnir made a post about it yesterday https://www.reddit.com/r/StableDiffusion/comments/1lu7nxx/wan_21_txt2img_is_amazing/

I think it has a much better concept of anatomy because videos contain temporal data on anatomy. Ill tag one example on the end which highlights the photographic differences between the base models (i don't have enough slots to show more)

Additional info: Wan is using a 10 step Lora which i have to assume reduces quality. It takes 500 seconds to generate a single image for Wan2.1 with my 1080 and 1000 for Flux at the same resolution (20 steps)

submitted by /u/Ok-Application-2261
[link] [comments]

Is there any site alternative to Civit? Getting really tired of it.

9 Julio 2025 at 02:02

I upload and post a new model, include ALL metadata and prompts on every single video yet when I check my model page it just says "no image" getting really tired of their mid ass moderation system and would love an alternative that doesn't hold the entire model post hostage until it decides to actually post it. No videos on the post are pending verification it says.

EDIT: It took them over 2 fucking hours to actually post the model and im not even a new creator I have 8.6k downloads (big whoop just saying its not a brand new account) yet they STILL suck ass. Would love it if we could get a site as big as civit but not suck ass.

submitted by /u/Commercial-Celery769
[link] [comments]

How do people achieve this cinematic anime style in AI art ?

8 Julio 2025 at 16:35
How do people achieve this cinematic anime style in AI art ?

Hey everyone!

I've been seeing a lot of stunning anime-style images on Pinterest with a very cinematic vibe — like the one I attached below. You know the type: dramatic lighting, volumetric shadows, depth of field, soft glows, and an overall film-like quality. It almost looks like a frame from a MAPPA or Ufotable production.

What I find interesting is that this "cinematic style" stays the same across different anime universes: Jujutsu Kaisen, Bleach, Chainsaw Man, Genshin Impact, etc. Even if the character design changes, the rendering style is always consistent.

I assume it's done using Stable Diffusion — maybe with a specific combination of checkpoint + LoRA + VAE? Or maybe it’s a very custom pipeline?

Does anyone recognize the model or technique behind this? Any insight on prompts, LoRAs, settings, or VAEs that could help achieve this kind of aesthetic?

Thanks in advance 🙏 I really want to understand and replicate this quality myself instead of just admiring it in silence like on Pinterest 😅

submitted by /u/Unlikely-Drive5770
[link] [comments]

I compared Kontext BF16, Q8 and FP8_scaled

8 Julio 2025 at 21:40
I compared Kontext BF16, Q8 and FP8_scaled

More examples with prompts in article: https://civitai.com/articles/16704

TLDR - nothing new, less details, Q8 is closer to BF16. Changing seed has bigger variations. No decrease in instruction following.

Interestingly I found random seed that basicaly destoys backgrounds. Also sometimes FP8 or Q8 performed sligtly better than others.

submitted by /u/shapic
[link] [comments]

LTX-Video 13B Control LoRAs - The LTX speed with cinematic controls by loading a LoRA

8 Julio 2025 at 15:20
LTX-Video 13B Control LoRAs - The LTX speed with cinematic controls by loading a LoRA

We’re releasing 3 LoRAs for you to gain precise control of LTX-Video 13B (both Full and Distilled).

The 3 controls are the classics - Pose, Depth and Canny. Controlling human motion, structure and object boundaries, this time in video. You can merge them with style or camera motion LoRAs, as well as LTXV's capabilities like inpainting and outpainting, to get the detailed generation you need (as usual, fast).

But it’s much more than that, we added support in our community trainer for these types of InContext LoRAs. This means you can train your own control modalities.

Check out the updated Comfy workflows: https://github.com/Lightricks/ComfyUI-LTXVideo

The extended Trainer: https://github.com/Lightricks/LTX-Video-Trainer

And our repo with all links and info: https://github.com/Lightricks/LTX-Video

The LoRAs are available now on Huggingface: 💃Pose | 🪩 Depth | ⚞ Canny

Last but not least, for some early access and technical support from the LTXV team Join our Discord server!!

submitted by /u/ofirbibi
[link] [comments]

Is there an easy way to convert pixel art into anime style?

9 Julio 2025 at 14:09

I have a pixel art upscaled to 512*512, and I want to convert it into anime style img2img. What is the best way to do this? Would it also help if I have a proper anime style reference image? Happy to hear any thoughts!

For example pixel art version: https://ibb.co/p6qbsBFW Anime art version: https://static.divine-pride.net/images/items/cards/4413.png

submitted by /u/Business_Force_9395
[link] [comments]

An update of my last post about making an autoregressive colorizer model

8 Julio 2025 at 15:49
An update of my last post about making an autoregressive colorizer model

Hi everyone;
I wanted to update you about my last lost about me making an autoregressive colorizer AI model that was so well received (which I thank you for that).

I started with what I thought was an "autoregressive" model but sadly was not really (Still line by line training and inference but was missing the biggest part which is "next line prediction based on previous one").

I saw that with my actual code it's reproducing in-dataset images near perfectly but sadly out-dataset images only makes glitchy "non-sense" images.

I'm making that post because I know my knowledge is very limited (I'm still understanding how all this works) and that I may just be missing a lot here. So I made my code online at github so you (the community) can help me shape it and make it work. (Code Repository)

As it may sounds boring (and FLUX Kontext dev got released and can do the same), I see that "fun" project as a starting point for me to train in the future an open-source "autoregressive" T2I model.

I'm not asking for anything but if you're experienced and wanna help a random guy like me, it would be awesome.

Thank you for taking time to read that useless boring post ^^.

PS: I take all criticism on my work even bad ones as long as It helps me understand more of this world and do better.

submitted by /u/YouYouTheBoss
[link] [comments]

What’s the best way to faceswap?

9 Julio 2025 at 12:50

Hi all, totally new to this!

So I have a set of faces for a character I’m building. What’s the best to say have it appear on a different image of a body?

Also, would this be possible with an existing short video?

I am currently using WebUI Forge - is this doable?

Thanks so much in advance!

submitted by /u/jinzo_the_machine
[link] [comments]

Creature Shock Flux LoRA

8 Julio 2025 at 20:45
Creature Shock Flux LoRA

My Creature Shock Flux LoRA was trained on approximately 60 images to excel at generating uniquely strange creatures with distinctive features such as fur, sharp teeth, skin details and detailed eyes. While Flux already produces creature images, this LoRA greatly enhances detail, creating more realistic textures like scaly skin and an overall production-quality appearance, making the creatures look truly alive. This one is a lot of a fun and it can do more than you think, prompt adherence is pretty decent, I've included some more details below.

I utilized the Lion optimizer option in Kohya, which proved effective in refining the concept and style without overtraining. The training process involved a batch size of 2, 60 images (no repeats), a maximum of 3000 steps, 35 epochs and a learning rate of 0.0003. The entire training took approximately 4 hours. Images were captioned using Joy Caption Batch, and the model was trained with Kohya and tested in ComfyUI.

The gallery will feature examples with workflows attached, I'm running a very simple 2-pass workflow for most of these, drag and drop the first image into ComfyUI to see the workflow. (It's being analyzed right now, may take a few hours to show up past the filter.)

There are a couple of things with variety that I'd like to improve. I'm still putting the model through its paces, and you can expect v1, trained with some of its generated outputs from v0, to drop soon. I really wanted to share this because I think we, as a community, often get stuck just repeating the same 'recommended' settings without experimenting with how different approaches can break away from default behaviors.

renderartist.com

Download from CivitAI

Download from Hugging Face

submitted by /u/renderartist
[link] [comments]
❌
❌