Making DnD Images Make me happy - Using Stable Diffusion
submitted by /u/Glacionn [link] [comments] |
submitted by /u/BlueMetaMind [link] [comments] |
I am not sure if this one was posted already but this looks stunning. Details are fantastic and video animation brought it to next level. If someone knows more about methods how this one was made... please share with us 😉 [link] [comments] |
You can try it out on tensor (or just download it from there), I didn't know Tensor was blocked but it's there under Cave Paintings. If you do try it, for best results try to base your prompts on these, https://www.bradshawfoundation.com/chauvet/chauvet_cave_art/index.php Best way is to paste one of them to your fav ai buddy and ask him to change it to what you want. Lora weight works best at 1, but you can try +/-0.1, lower makes your new addition less like cave art but higher can make it barely recognizable. Same with guidance 2.5 to 3.5 is best. [link] [comments] |
submitted by /u/Rich-Assistant-4657 [link] [comments] |
A short WIP trailer for a technical trading course, using several workflows like steerable motion and vid2vid. Lmk what i can improve on [link] [comments] |
So, as always I am at the cutting edge of SD info, so I only recently found out you can have random options in a prompt.
Just for Gits & Shiggles, I tried a prompt with containing 10 random options. It could be useful if you have prompters block ;)
Here's the prompt -> one {Happy | Sad | Angry | Scared} , {man | woman | child} with {long | Short | Updo | Permed} , {Ginger | Blond | Black | Brown} hair, wearing {dark | bright} , {casual | formal | Sports | Evening} clothing, looking {up | Down | Left | Right}, infront of a {wall | Bush | Shop | building} , {neon| Bright | Dim} lighting , {day | night}
Example - https://imgur.com/a/SKPuwD9
(Posting this here, because people here are more talented than some other freelance platform).I’m looking to hire a freelance AI Video Generation Specialist who’s experienced with ComfyUI, ControlNet, and other AI tools to create videos similar to Varun Mayya’s style. This is a consultancy-based role (hourly pay), but full-time is also an option if you’re interested. If you’re up-to-date with the latest AI image/video generation models and can create stunning video presentations, please DM me or reply with your portfolio and rate. Let’s make something awesome!
I want to upscale a batch of old videos - is there a clever way of loading a video batch or increment the files loaded in ComfyUI? Right now I'm manually entering the file path in the VHS nodes, but it's really time consuming to enter each file path :)
I could use some help with finding the right parameters for training my Lora with Flux. This is the object that I'm trying to reproduce with a Lora. And these are the best results so far (I handpicked the best, I also get some unusable results): I used the guide from this repo: https://github.com/geocine/flux to have an environment to train my Lora in. I'm using this for my config.yaml
This is the output config.yaml
Here are some images that I have used to train with And here are some prompts I have used together with the images
A whiteboard with a red boarder and an open [trigger] attached to it. Above the [trigger] on the whiteboard is a blue and white sticker that reads "ROOM". To the left of the board is a stand with papers on it, and the floor is covered with a carpet. A whiteboard with a red border on a ledge leaning against a partially green and orange wall with white text on it. On the whiteboard, there is a closed [trigger] attached to it. In front of the whiteboard, on the ledge, is a bottle of whiteboard eraser fluid. The red border of the whiteboard also has some whiteboard markers placed into it. On the right side there is a smaller whiteboard with a gray border leaning against the bigger whiteboard. A whiteboard with a red border and a green wall behind it. On the whiteboard, there is a open [trigger], as well as a bottle and other smaller whiteboard with a gray border in front of the board. A white wall with a variety of items on it, including marker holders, whiteboard markers, a closed [trigger], 4 whiteboard erasers in different colors and 2 wooden objects. A wooden stand with a closed [trigger] sitting on top of it. The background is a white wall. A wooden stand with a closed [trigger] on top of it standing in front of a white wall. Any help is truly appreciated! [link] [comments] |
I have a few different faceswap models that I compiled using FaceSwapLab. In the last few years, I vaguely remember using some extension that allowed me to drop my safetensors file into it and it would display all the training images used to create it. Does this actually exist, or am I hallucinating? Thanks for any help :)
I'm by no means an expert regarding AI stuff, so please correct me if I get something wrong; that way I can learn.
Recently I've been seeing a lot of videos with a watermark referencing the company ViggleAi.
ViggleAI offers a service which lets you replace characters in videos. To do this, it will ask for two things:
they call it the "motion"
the reference character that replaces one in the video/motion
What is so mind blowing to me is how smooth and coherent their output is. I mean sure, it does get jittery sometimes, but generally, the motion and even the facial expressions all look incredibly smooth.
I'm curious on how this is done and if there is any open source solution offering something similar.
I know this example is kind of immature and all of that, but it's the best one that I have found
I've been looking at https://huggingface.co/models, but I have found no model that could do this out of the box.
I was thinking that perhaps they process the video frame by frame with a model like Leffa or SDXL-inpaint, but if each frame is processed with no context of the surrounding frames, must that not cause a lot of incoherency? I certainly believe so.
Perhaps the video is segmented into masks of objects and then this is used to map the input character to a character in the video? Sounds to me like quite a reasonable way to go about it.
Let's say we have a mask of a character that stretches throughout a certain scene. Now what? Well I'm not too sure to be honest.
What strikes me as challenging is:
How can the orientation be predicted from a single image? Is a 3d-representation of the input character created and mapped to a character's mask from the video?
Anyone got any advice?