The same prompt used for all of the images and I feel like the de-distilled one wins by a long shot after adding the realism, detail, turbo and fast Loras. And not to forget detail daemon on top of everything. I feel when adding a negative prompt, it switches into another mode, where things look quite fine grained but also a bit rougher but has way more fidelity than without. And the great part is the base image was generated in about 10 seconds on an RTX 4090 thanks to the turbo and fast Loras where only 8 steps were used. I don’t really see anything degraded from the turbo Lora where for example in SD 1.5 the LCM Lora was way more obvious. [link] [comments] |
Vista de Lectura
Flux Dev De-distilled VS Flux Pro VS Flux 1.1 Pro VS Flux 1.1 Pro Ultra Raw
ComfyUI: Isn't it time to put some sanity back for us users?
I think it's good that ComfyUI is becoming more and more of the default tool for implementing various AI repos. For one, it is so much nicer to run everything from one place rather than keep installing Torch and all the requirements for each and every repo. But it also comes at a cost.
Recently, I had to upgrade CUDA to 12.4 and Torch to 2.5 to make them compatible with Triton for Mochi1. Then, Ecomimic2 dropped which comes with ComfyUI custom nodes from the get-go. After looking at their requirements.txt, I decided not to install the custom nodes and went with their original repo installation instead. The problem? Dependency conflicts starting from Torch to all the dependencies.
But this kind of problem isn't new to the point of my acquiring a paranoia of checking requirements.txt before implementing any new custom nodes. Now that ComfyUI has gathered enough clout, isn't it time to have a dependency policy that says either you follow the dependency policy in making custom nodes or get the hell out?
Also, I have noticed that there are duplicate models and files as different custom nodes for different repos use different file management structures. I don't know but SSD space is a rather rare commodity these days for me and I just hate having all these files duplicated in different folders. Wouldn't it be too difficult to have a file management policy as well?
[link] [comments]
me these few days, cogVideoX > cogVideoX 1.5 > LTX video > cogVideoX 1.5
like rollar coaster, keep try out the best way on I2V. cogVideoX still most reliable,
already give up on LTX I2V, it just give me still image most of the time.
[link] [comments]
Testing LTX Video on my GTX3060 12GB
First of all, it's blazingly fast. It only takes about 90s to generate a 768x512 video for 97 frames, that is 1 second per frame, generating an image with Flux could take longer than that, so wow. And here's some of my findings from the test I did this morning.
edit: the videos were still for some reasons, replaced them with live images [link] [comments] |
KFC style crispy chicken, wings
i realized it is so difficult to get crispy chicken wings or deep fried crispy chicken right in SDXL/FLUX... all i get are mutated disguising looking meat.... so far only firefly did a good job. what will be your approach to make a better crispy chicken wings
[link] [comments]
Flux Tools - A compilation of cleaned up workflows for beginners
Hi guys,
Flux Tools just came out, and it is hands down one of the most consistent image control tools I've used for Flux (duh, given it comes from BlackForestLabs).
One of the pain points I had while learning how to use each of these tools is not having a place with just the most fundamental and cleaned up workflows (no extraneous nodes, dependencies etc) to begin with. So, I spent the day trying out and cleaning up every one of these workflows so beginners will find a more friendly starting point.
I've also included all the links to the files you need to download in each workflow repo on openart, so here they are:
Flux Fill (Inpaint) - https://openart.ai/workflows/odam_ai/flux-fill-inpaint---official-flux-tools-by-bfl---beginner-friendly/8wIPSZy0aOuXsGfdfIVp
Flux Fill (Outpaint) - https://openart.ai/workflows/odam_ai/flux-fill-outpaint---official-flux-tools-by-bfl---beginner-friendly-edit/6CeBgmyrVDP35r4pO4S9
Flux Depth ControlNet - https://openart.ai/workflows/odam_ai/flux-tools-best-depth-controlnet---official-flux-tools-by-bfl---beginner-friendly/2UDeSn35mPGIEqT1tgYu
Flux Canny ControlNet - https://openart.ai/workflows/odam_ai/flux-tools-best-canny-controlnet---official-flux-tools-by-bfl---beginner-friendly/O8aLfWdCOKGCyJX79Jm0
Flux Redux - https://openart.ai/workflows/odam_ai/flux-redux---official-flux-tools-by-bfl---beginner-friendly/tgGYqY7Kri5bMzaulHiI
Have fun!
Stonelax
[link] [comments]
Sticker Designs
I’ve been experimenting with prompts to generate clean and outlined Sticker designs. Here are some of the prompts I used: A bold, graphic representation of the Joker's face, featuring exaggerated facial features with a wide, sinister grin and vibrant green hair. The design uses high contrast black and white elements, ensuring clarity in smaller sizes. The text "Why So Serious?" is integrated into the design, arched above the Joker's head in a playful yet menacing font. The sticker has a die-cut shape around the character's outline, with a 1/8 inch border. Ideal for both glossy and matte finishes, with clear knock-out spaces around the text. Bold, stylized "Wakanda Forever" text in an intricate, tribal-inspired font, surrounded by a powerful black panther silhouette. The panther has sharp, clean outlines and features vibrant green and gold accents, symbolizing vibrancy and strength. The design is die-cut into the shape of the panther, with a thick, contrasting black border. The background is transparent to enhance the focus on the text and panther, ensuring clarity at 1-3 inches. The color scheme is high contrast, working beautifully in glossy and matte finishes. Incorporate a layered effect, with the text appearing to emerge from the panther, designed for optimal visibility on both print and digital platforms. A stylized baby Groot character with oversized expressive eyes and a playful stance, surrounded by vibrant, oversized leaves. The text "I Am Groot" is bold and playful, integrated into the design as if Groot is playfully holding it. Die-cut shape with organic edges, ensuring the design stands out. High contrast colors of deep greens and warm browns against a white background, maintaining clarity at sizes of 1-3 inches. Plan for a glossy finish to enhance color vibrancy. Mortal Kombat Skorpion in a dynamic pose with his iconic yellow and black costume, holding a flaming spear, surrounded by jagged orange and red flames. The text "Finish Him!" in bold, stylized typography arcs above him, contrasting in white with a black outline. The design is die-cut in a jagged shape following the outline of Skorpion and the flames. High contrast colors ensure visibility at small sizes, with negative space around the character enhancing clarity. Suitable for glossy or matte finishes. [link] [comments] |
Noob needs help. Trying to recreate image from civitai and failing
I've been trying for 8 and failing. I try to do iOS so many times I found an image I liked on civitai (https://civitai.com/models/929922/princess-peach-from-mario-and-luigi-brothership-style-pony-xl) and I am trying to recreate it here. I try normally and it doesn’t work, I tried even alternative ways and I'm not getting the same image (see below)! What am I doing wrong? Because lm need to make it 100% same Here what I, trying to recreate exactly same https://civitai.com/images/39044876 And hear what I was get https://orchestration.civitai.com/v2/consumer/blobs/PZSXQ67BE55AA03VPMAHEY1BH0
[link] [comments]
frustrated with flux1
frustrated with flux1 ... sometime it just wont do what you ask most of the time... jumped back to SDXL and got it right instantly... the Prompt: side view of a white cat, long fur, head looking upwards, mouth open wide, highly detailed, rim lighting, dslr, SDXL SDXL cat look good with head look upward FLUX [link] [comments] |
What’s your process for showing two characters in a specific intertwined double pose?
I'm curious—how do you approach generating characters in specific double poses that might come up in storytelling? For example
- One person giving another a piggyback ride.
- One person slung over another’s shoulder, like a sack of potatoes.
- One person propping up a tipsy friend, by holding them around the waist, with their friend's arm draped over them
- One person cornering another, hands on the wall to either side of their head, in a menacing way.
(I'll skip over poses like two people fighting, kissing, hugging, or having sex, since reference images for those tend to be fairly easy to find.)
How do you go about this?
Do you start by creating the double pose and then swap the faces using inpainting? If so, what’s your process for creating the initial pose?
Do you use a specific model that’s particularly good at understanding text prompts for double poses? Then just generate a bunch of images until you find one you like, then refine it from there?
Do you sketch the double pose in a drawing, and then gradually develop it into what you want to see, and in what style, through image to image plus inpainting? (That’s what I have been doing so far.)
Do you use Blender or PoseMyArt to create a depthmap, which you then use for a depth controlnet?
(Am I correct in thinking that using an openpose controlnet with two skeletons with a lot of overlap isn’t really working?)
Do you use regional prompting, or attention masking? Is “attention masking” actually all that useful for double poses that go beyond two people just standing next to each other, given that the masks can’t intertwine? Can it be used for any of the above examples?
Do you use omnigen?
What apps / combination of apps do you use for this? Krita? Blender? Invoke? Comfy?
Would love to hear about your workflows, tips, or tools—anything that helps you create complex double poses effectively.
[link] [comments]
LoRa and PullID
I just discovered PullID and found it to be a really good tool as it can basically create similar things to LoRa without training. I'm wondering what advantage does training a LoRa have over PullID?
[link] [comments]
Is there an easy to use OpenPose UI
I've found workflows that will take a picture and generate an openpose in comfyui, but what I am really looking for is a UI that lets me create custom poses or manipulate the poses and then output the poses as PNG files like this: [link] [comments] |