Vista Normal

Hay nuevos artículos disponibles. Pincha para refrescar la página.
Hoy — 10 Abril 2025StableDiffusion

Comparison of HiDream-I1 models

10 Abril 2025 at 13:33
Comparison of HiDream-I1 models

There are three models, each one about 35 GB in size. These were generated with a 4090 using customizations to their standard gradio app that loads Llama-3.1-8B-Instruct-GPTQ-INT4 and each HiDream model with int8 quantization using Optimum Quanto. Full uses 50 steps, Dev uses 28, and Fast uses 16.

Seed: 42

Prompt: A serene scene of a woman lying on lush green grass in a sunlit meadow. She has long flowing hair spread out around her, eyes closed, with a peaceful expression on her face. She's wearing a light summer dress that gently ripples in the breeze. Around her, wildflowers bloom in soft pastel colors, and sunlight filters through the leaves of nearby trees, casting dappled shadows. The mood is calm, dreamy, and connected to nature.

submitted by /u/thefi3nd
[link] [comments]

Converted my favorite scene from Spirited Away to 3D using the Depthinator, a free tool I created that convert 2D video to side-by-side and red-cyan anaglyph 3D. Cross-eye method kinda works but looks phenomenal on a VR headset.

10 Abril 2025 at 10:21
Converted my favorite scene from Spirited Away to 3D using the Depthinator, a free tool I created that convert 2D video to side-by-side and red-cyan anaglyph 3D. Cross-eye method kinda works but looks phenomenal on a VR headset.

Download the mp4 here

Download the Depthinator here

Looks amazing on a VR headset. The cross-eye method kinda works, but I set the depth-scale too low to really show off the depth using that method. I recommend viewing through a VR headset. The Depthinator uses video depth anything via comfyui to get the depth then the pixels are shifted using an algorithmic process that doesn't use AI. All locally run!

submitted by /u/kingroka
[link] [comments]

Structure-Preserving Style Transfer (Flux[dev] Redux + Canny)

10 Abril 2025 at 08:01
Structure-Preserving Style Transfer (Flux[dev] Redux + Canny)

This project implements a custom image-to-image style transfer pipeline that blends the style of one image (Image A) into the structure of another image (Image B).We've added canny to the previous work of Nathan Shipley, where the fusion of style and structure creates artistic visual outputs. Hope you check us out on github and HF give us your feedback : https://github.com/FotographerAI/Zen-style and HuggingFace : https://huggingface.co/spaces/fotographerai/Zen-Style-Shape

We decided to release our version when we saw this post lol : https://x.com/javilopen/status/1907465315795255664

submitted by /u/Comfortable-Row2710
[link] [comments]

What would be the best tool to generate facial images from the source?

10 Abril 2025 at 10:25
What would be the best tool to generate facial images from the source?

I've been running a project that involves collecting facial images of participants. For each participant, I currently have five images taken from the front, side, and 45-degree angles. For better results, I now need images from in-between angles as well. While I can take additional shots for future participants, it would be ideal if I could generate these intermediate-angle images from the ones I already have.

What would be the best tool for this task? Would Leonardo or Pica be a good fit? Has anyone tried Icons8 for this kind of work?

Any advice will be greatly appreciated!

submitted by /u/talkinape888
[link] [comments]

Dear Anyone who ask a question for troubleshoot

10 Abril 2025 at 12:57

Buddy, for the love of god, please help us help you properly.

Just like how it's done on GitHub or any proper bug report, please provide your full setup details. This will save everyone a lot of time and guesswork.

Here's what we need from you:

  1. Your Operating System (and version if possible)
  2. Your PC Specs:
    • RAM
    • GPU (including VRAM size)
  3. The tools you're using:
    • ComfyUI / Forge / A1111 / etc. (mention all relevant tools)
  4. Screenshot of your terminal / command line output (most important part!)
    • Make sure to censor your name or any sensitive info if needed
  5. The exact model(s) you're using

Optional but super helpful:

  • Your settings/config files (if you changed any defaults)
  • Error message (copy-paste the full error if any)
submitted by /u/Altruistic_Heat_9531
[link] [comments]

Stubborn toilet

10 Abril 2025 at 12:26
Stubborn toilet

Hello everyone, I generated this photo and there is toilet in the background (I zoomed in). I tried to inpaint this in flux for 30 min and no matter what I do it just generates another toilet. I know my workflow works because I inpainted seamlessly countless time. Now I don’t even care about it I just want to know why it doesn’t work and what am I doing wrong?

There is mask on whole toilet and its shadow and I tried a lot of prompts like „bathroom wall seamlessly blending with the background”

submitted by /u/Nervous-Ad-7324
[link] [comments]

Did your ComfyUI generations degrade in quality when you use a LoRA in the last few weeks?

10 Abril 2025 at 11:25
Did your ComfyUI generations degrade in quality when you use a LoRA in the last few weeks?

A few weeks ago, I noticed a sudden degradation in quality when I generate FLUX images with LoRAs.

Normally, the XLabs FLUX Realism LoRA, if configured in a certain way, used to generate images as crisp and beautiful as this one:

https://preview.redd.it/h2azhzxylzte1.jpg?width=1344&format=pjpg&auto=webp&s=ff81abf47b304c54dfc722db34e2ec5d427397c7

I have many other examples of images of this quality, with that LoRA and many others (including LoRAs I trained myself). I have achieved this quality since the first LoRAs for FLUX were released by the community. The quality has not changed since Aug 2024.

However, some time between the end of January and February* the quality suddenly decreased dramatically, despite no changes to my workflow or my Pytorch environment (FWIW configured with Pytorch 2.5.1+CUDA12.4 as I think it produces subtly better images than Pytorch 2.6).

Now, every image generated with a LoRA looks slightly out of focus / more blurred and, in general, not close to the quality I used to achieve.

Again: this is not about the XLabs LoRA in particular. Every LoRA seems to be impacted.

There are a million reasons why the quality of my images might have degraded in my environment, so a systematic troubleshooting is a very time-consuming exercise I postponed so far. However, a brand new ComfyUI installation I created at the end of February showed the same inferior quality, and that made me question if it's really a problem in my system.

Then, today, I saw this comment, mentioning an issue with LoRA quality and WanVideo, so I decided to ask if anybody noticed something slightly off.

I maintained APW for ComfyUI for 2 years now, and I use it on a daily basis to generate images at an industrial scale, usually at 50 steps. I notice changes in quality or behavior immediately, and I am convinced I am not crazy.

Thanks for your help.

*I update ComfyUI (engine, manager, and front end) on a daily basis. If you noticed the same but you update them more infrequently, your timeline might not align with mine.

submitted by /u/GianoBifronte
[link] [comments]

A word of thanks to the Stable Diffusion community

9 Abril 2025 at 23:14

You will occasionally see me post a URL to my latest release of my desktop application AI Runner. If you look through my history you'll see many posts over the years to /r/stablediffusion - this is because I made the app specifically for Stable Diffusion and the /r/stablediffusion community.

I don't know if any of the OGs are around, but many of you provided feedback, opened bugs and even donated, so I just wanted to say thank you for your support. If you weren't one of those people, that's fine too - I just enjoy building AI tools and I pay a lot of attention to the things you all say in comments about the tools that you use, so that's very valuable as well.

I've started putting more effort into the app again recently and will have a new packaged version available soon and of course I'll post about it here when its available.

submitted by /u/w00fl35
[link] [comments]

We already have t5xxl's txt condition in flux, why it still uses clip's vec guidance in generation?

10 Abril 2025 at 14:03
We already have t5xxl's txt condition in flux, why it still uses clip's vec guidance in generation?

Hi guys. I'm just wondering since we already have t5xxl for txt condition, why flux still uses clip's guidance. I'm new to this area, can anyone explain this to me?

And I actually did a little test, in the flux forward function, I add this:

 img = self.img_in(img) vec = self.time_in(timestep_embedding(timesteps, 256)) if self.params.guidance_embed: if guidance is None: raise ValueError("Didn't get guidance strength for guidance distilled model.") vec = vec + self.guidance_in(timestep_embedding(guidance, 256)) y = y * 0 # added so l_pooled is forced to be plain zeros vec = vec + self.vector_in(y) 

and I compared the results when force vec to be zero or not, the seed is 42, resolution (512,512), flux is quantized to fp8e4m3, and prompt is "a boy kissing a girl.":
use vec as usual:

https://preview.redd.it/x5dr7nxqk0ue1.png?width=512&format=png&auto=webp&s=95748622ed4c4210d389a8ad80a510704a69070a

force vec to be zeros:

https://preview.redd.it/dm8eelcvk0ue1.png?width=512&format=png&auto=webp&s=cde34c6ff9606e427cb8f7f7251dd32b61de6bb1

For me the differences between these results are tiny. So I really rope someone can explain this to me. Thanks!

submitted by /u/Creepy_Astronomer_83
[link] [comments]

[Question] Is there a model that I can use to add colors onto a clay render?

10 Abril 2025 at 04:26
[Question] Is there a model that I can use to add colors onto a clay render?

[Images only for reference]

I make things to 3D print, but I want my clay renders to look more eye catching at times. are there any tools that can add color to a clay render, like to the one on the left, to make it look more like the image on the right without changing the geo at all? bonus points if I can mess with the style, or make it look painted. But keeping the geo consistent is important.

All other tools I've found change the features of the model.

submitted by /u/TheArtCallstoMe
[link] [comments]
❌
❌