Every OpenAI image.
![]() | At least we do not need sophisticated gen AI detectors. [link] [comments] |
![]() | At least we do not need sophisticated gen AI detectors. [link] [comments] |
![]() | prompt was `http://127.0.0.1:8080\` so if you're using this IP address, you have skynet installed and you're probably going to kill all of us. [link] [comments] |
https://github.com/mit-han-lab/nunchaku/discussions/236
![]() | submitted by /u/thoughtlow [link] [comments] |
![]() | part 2 of my wan vid2vid workflow with real life footage and style transfer using wan control [link] [comments] |
![]() | submitted by /u/Leading_Hovercraft82 [link] [comments] |
![]() | ostris's newly released Redux model touts a better vision encoder and a more permissive license than Flux Redux. [link] [comments] |
![]() | I made this block building app in 2019 but shelved it after a month of dev and design. In 2024, I repurposed it to create architectural images using Stable Diffusion and Controlnet APIs. Few weeks back I decided to convert those images to videos and then generate a 3D model out of it. I then used Model-Viewer (by Google) to pose the model in Augmented Reality. The model is not very precise, and needs cleanup.... but felt it is an interesting workflow. Of course sketch to image etc could be easier. P.S: this is not a paid tool or service, just an extension of my previous exploration [link] [comments] |
I think we've reached a point where some of us could give some useful advice how to design a Wan 2.1 prompt. Also if the negative prompt(s) makes sense. And has someone experience with more then 1 lora? Is this more difficult or doesnt matter at all?
I do own a 4090 and was creating a lot in the last weeks, but I'm always happy if the outcome is a good one, I'm not comparing like 10 different variations with prompt xyz and negative 123. So I hope the guys who rented (or own) a H100 could give some advice, cause its really hard to create "prompt-rules" if you havent created hundreds of videos.
I was about to test out i2v 480p fp16 vs fp8 vs q8, but I can't get fp16 loaded even with 35 block swaps, and for some reasons my GGUF loader is broken since about a week ago, so I can't quite do it myself at this moment.
So, has anyone done a quality comparison of fp16 vs fp8 vs q8 vs 6 vs q4 etc?
It'd be interesting to know whether it's worth going fp16 even though it's going to be sooooo much slower.
![]() | This post is to motivate you guys out there still on the fence to jump in and invest a little time learning ComfyUI. It's also to encourage you to think beyond just prompting. I get it, not everyone's creative, and AI takes the work out of artwork for many. And if you're satisfied with 90% of the AI slop out there, more power to you. But you're not limited to just what the checkpoint can produce, or what LoRas are available. You can push the AI to operate beyond its perceived limitations by training your own custom LoRAs, and learning how to think outside of the box. Stable Diffusion has come a long way. But so have we as users. Is there a learning curve? A small one. I found Photoshop ten times harder to pick up back in the day. You really only need to know a few tools to get started. Once you're out the gate, it's up to you to discover how these models work and to find ways of pushing them to reach your personal goals. \"It's okay. They have YouTube tutorials online.\" Comfy's "noodles" are like synapses in the brain - they're pathways to discovering new possibilities. Don't be intimidated by its potential for complexity; it's equally powerful in its simplicity. Make any workflow that suits your needs. There's really no limitation to the software. The only limit is your imagination. Same artist. Different canvas. I was a big Midjourney fan back in the day, and spent hundreds on their memberships. Eventually, I moved on to other things. But recently, I decided to give Stable Diffusion another try via ComfyUI. I had a single goal: make stuff that looks as good as Midjourney Niji. Ranma 1/2 was one of my first anime. Sure, there are LoRAs out there, but let's be honest - most of them don't really look like Midjourney. That specific style I wanted? Hard to nail. Some models leaned more in that direction, but often stopped short of that high-production look that MJ does so well. Mixing models - along with custom LoRAs - can give you amazing results! Comfy changed how I approached it. I learned to stack models, remix styles, change up refiners mid-flow, build weird chains, and break the "normal" rules. And you don't have to stop there. You can mix in Photoshop, CLIP Studio Paint, Blender -- all of these tools can converge to produce the results you're looking for. The earliest mistake I made was in thinking that AI art and traditional art were mutually exclusive. This couldn't be farther from the truth. I prefer that anime screengrab aesthetic, but maxed out. It's still early, I'm still learning. I'm a noob in every way. But you know what? I compared my new stuff to my Midjourney stuff - and the latter is way better. My game is up. So yeah, Stable Diffusion can absolutely match Midjourney - while giving you a whole lot more control. With LoRAs, the possibilities are really endless. If you're an artist, you can literally train on your own work and let your style influence your gens. So dig in and learn it. Find a method that works for you. Consume all the tools you can find. The more you study, the more lightbulbs will turn on in your head. Prompting is just a guide. You are the director. So drive your work in creative ways. Don't be satisfied with every generation the AI makes. Find some way to make it uniquely you. In 2025, your canvas is truly limitless. Tools: ComfyUI, Illustrious, SDXL, Various Models + LoRAs. (Wai used in most images) [link] [comments] |
![]() | submitted by /u/photobombolo [link] [comments] |
![]() | I'm currently working on a script for my workflow on modal. Will release the Github repo soon. [link] [comments] |
![]() | ✅Workflow link (free no paywall) ✅Video tutorial [link] [comments] |
![]() | Images created in Flux and animated with kling [link] [comments] |
I sell fantasy lingerie. I want to take the same outfit from my existing photo and show it on a new AI-generated model (different pose, face, background).
What tools or workflow can do this? Thanks!