Vista de Lectura

Hay nuevos artículos disponibles. Pincha para refrescar la página.

I successfully 3D-printed my Illustrious-generated character design via Hunyuan 3D and a local ColourJet printer service

I successfully 3D-printed my Illustrious-generated character design via Hunyuan 3D and a local ColourJet printer service

Hello there!

A month ago I generated and modeled a few character designs and worldbuilding thingies. I found a local 3d printing person that offered colourjet printing and got one of the characters successfully printed in full colour! It was quite expensive but so so worth it!

i was actually quite surprised by the texture accuracy, here's to the future of miniature printing!

submitted by /u/Neggy5
[link] [comments]

HiDream-I1: New Open-Source Base Model

HiDream-I1: New Open-Source Base Model

HuggingFace: https://huggingface.co/HiDream-ai/HiDream-I1-Full
GitHub: https://github.com/HiDream-ai/HiDream-I1

From their README:

HiDream-I1 is a new open-source image generative foundation model with 17B parameters that achieves state-of-the-art image generation quality within seconds.

Key Features

  • Superior Image Quality - Produces exceptional results across multiple styles including photorealistic, cartoon, artistic, and more. Achieves state-of-the-art HPS v2.1 score, which aligns with human preferences.
  • 🎯 Best-in-Class Prompt Following - Achieves industry-leading scores on GenEval and DPG benchmarks, outperforming all other open-source models.
  • 🔓 Open Source - Released under the MIT license to foster scientific advancement and enable creative innovation.
  • 💼 Commercial-Friendly - Generated images can be freely used for personal projects, scientific research, and commercial applications.

We offer both the full version and distilled models. For more information about the models, please refer to the link under Usage.

Name Script Inference Steps HuggingFace repo
HiDream-I1-Full inference.py 50 HiDream-I1-Full🤗
HiDream-I1-Dev inference.py 28 HiDream-I1-Dev🤗
HiDream-I1-Fast inference.py 16 HiDream-I1-Fast🤗
submitted by /u/latinai
[link] [comments]

Agent Heroes - Automate your characters with images and videos

Hi community :)

I love creating pictures and video on socials using things like ChatGPT and Mid-journey and convert it to video on Replicate and Fal.

But I realized it's super time consuming 😅

So I created a AgentHeroes, a repository to train models, generate pictures, video and schedule it on social media.

https://github.com/agentheroes/agentheroes

Not sure if it's something anybody needs so happy for feedback.

Of course a star would be awesome too 💕

Here is what you can do:

  • Connect different services like Fal, Replicate, ChatGPT, Runway, etc.
  • Train images based on models you upload or using models that create characters.
  • Generate images from all the models or use the trained model.
  • Generate video from the generated image
  • Schedule it on social media (currently I added only X, but it's modular)
  • Build agents that can be used with an API or scheduler (soon MCP):
    • Check reddit posts
    • Generate a character based on that post
    • Make it a video
    • Schedule it on social media

Everything is fully open-source AGPL-3 :)

Some notes:

Backend is fully custom, no AI was used but the frontend is fully vibe code haha, it took me two weeks to develop it instead of of a few months.

There is a full-working docker so you can easily deploy the project.

Future Feature:

  • Connect ComfyUI workflow
  • Use local LLMs
  • Add MCPs
  • Add more models
  • Add more social medias to schedule to

And of course, let me know what else is missing :)

submitted by /u/Mean_Preparation_364
[link] [comments]

I built an image viewer that reads embedded prompts from AI images (PNG/JPEG), maybe someone is interested :)

I built an image viewer that reads embedded prompts from AI images (PNG/JPEG), maybe someone is interested :)
Hey, I built a image viewer that automatically extracts prompt data from PNG and JPEG files — including prompt, negative prompt, and settings — as long as the info is embedded in the image (e.g. from Forge, ComfyUI, A1111, etc.). You can browse folders, view prompts directly, filter, delete images, and there’s also a fullscreen mode with copy functions. If you have an image where nothing is detected, feel free to send it to me along with the name of the tool that generated it. The tool is called ImagePromptViewer. GitHub: https://github.com/LordKa-Berlin/ImagePromptViewer Feel free to check it out if you're interested. 

https://preview.redd.it/6m116qebylte1.png?width=2560&format=png&auto=webp&s=1c77f7a5c981ba7312d7170e5f3c74107f90728a

https://preview.redd.it/z6jmfj6cylte1.png?width=2560&format=png&auto=webp&s=ef50c3472c8dc7e3c5635fd62ae446d79aa880a3

submitted by /u/Ok_Heron8703
[link] [comments]

TripoSF: A High-Quality 3D VAE (1024³) for Better 3D Assets - Foundation for Future Img-to-3D? (Model + Inference Code Released)

TripoSF: A High-Quality 3D VAE (1024³) for Better 3D Assets - Foundation for Future Img-to-3D? (Model + Inference Code Released)

Hey community! While we all love generating amazing 2D images, the world of Image-to-3D is also heating up. A big challenge there is getting high-quality, detailed 3D models out. We wanted to share TripoSF, specifically its core VAE (Variational Autoencoder) component, which we think is a step towards better 3D generation targets. This VAE is designed to reconstruct highly detailed 3D shapes.

What's cool about the TripoSF VAE? * High Resolution: Outputs meshes at up to 1024³ resolution, much higher detail than many current quick 3D methods. * Handles Complex Shapes: Uses a novel SparseFlex representation. This means it can handle meshes with open surfaces (like clothes, hair, plants - not just solid blobs) and even internal structures really well. * Preserves Detail: It's trained using rendering losses, avoiding common mesh simplification/conversion steps that can kill fine details. Check out the visual comparisons in the paper/project page! * Potential Foundation: Think of it like the VAE in Stable Diffusion, but for encoding/decoding 3D geometry instead of 2D images. A strong VAE like this is crucial for building high-quality generative models (like future text/image-to-3D systems).

What we're releasing TODAY: * The pre-trained TripoSF VAE model weights. * Inference code to use the VAE (takes point clouds -> outputs SparseFlex params for mesh extraction). * Note: Running inference, especially at higher resolutions, requires a decent GPU. You'll need at least 12GB of VRAM to run the provided examples smoothly.

What's NOT released (yet 😉): * The VAE training code. * The full image-to-3D pipeline we've built using this VAE (that uses a Rectified Flow transformer).

We're releasing this VAE component because we think it's a powerful tool on its own and could be interesting for anyone experimenting with 3D reconstruction or thinking about the pipeline for future high-fidelity 3D generative models. Better 3D representation -> better potential for generating detailed 3D from prompts/images down the line.

Check it out: * GitHub: https://github.com/VAST-AI-Research/TripoSF * Project Page: https://xianglonghe.github.io/TripoSF * Paper: https://arxiv.org/abs/2503.21732

Curious to hear your thoughts, especially from those exploring the 3D side of generative AI! Happy to answer questions about the VAE and SparseFlex.

submitted by /u/pookiefoof
[link] [comments]

Will this thing work for Video Generation? NVIDIA DGX Spark with 128GB

Will this thing work for Video Generation? NVIDIA DGX Spark with 128GB

Wondering if this will work also for image and video generation and not just LLMs. With LLMs we could always groupt our GPUs together to run larger models, but with video and image generation, we are mostly limited to a single GPU, which makes this enticing to run larger models, or more frames and higher resolution videos. Doesn't seem that bad, considering the possibilities we could do with video generation with 128GB. Will it work or is it just for LLMs?

submitted by /u/Prestigious-Use5483
[link] [comments]

Anybody got any tips and tricks to try keep or match the same face used as the refrence image in generated images using wan2.1 i2v

Seem to be having a hard time trying to keep the resemblance to the face in my reference images using wan.. it always seems to get it wrong where for the most part the person's face is completely different, I tried different models and denonising ammounts but there's so many options here, you could literally spend months messing around by the time a video generation is done to see any difference, I understand that it can't get it very accurate, but what's the general best sampler model and tweaks to get a decent enough similarity?

submitted by /u/AutomaticChaad
[link] [comments]

Creating Before/After Beaver Occupancy AI Model

Creating Before/After Beaver Occupancy AI Model

Howdy! Hopefully this is the right subreddit for this - if not please tell refer me to a better spot!

I am an ecology student working with a beaver conservation foundation and we are exploring possibilities of creating an AI model that will take a before photo of a landowner's stream (see 1st photo) and modify it to approximate what it could look like with better management practices and beaver presence (see next few images). The key is making it identifiable, so that landowners could look at it and be better informed at how exactly our suggestions could impact their land.

Although I have done some image generation and use LLMs with some consistency, I have never done anything like this and am looking for some suggestions on where to start! From what I can tell, I should probably fine-tune a model and possibly make a LoRA, since untrained models do a poor job (see last photo). I am working on making a database with photos such as the ones I posted here, but I am not sure what to do beyond that.

Which AI model should I train? What platform is best for training? Do I need to train it on both "before" and "after" photos, or just "after"?

Any and all advice is greatly appreciated!!! Thanks

submitted by /u/spencerarnold
[link] [comments]

Combining multiple GPUs

Hello all!

I've been recently experimenting with SDXL+LCM running off ComfyUI on my rig, which has a 1080 8gb card, and I've been getting pretty good results, I'm able to generate 1216*832 images in about 45-60 seconds.

This got me thinking about getting a second card to upgrade performance, I was thinking a 3080 10gb card. Would this be a viable upgrade, as in would I be able to use both cards at the same time in ComfyUI? What would a ballpark performance gain be? Finally, I would love to hear what GPUs in the $300-200 dollar range would y'all recommend? I'm pretty constrained budgetwise so I'd really appreciate some suggestions.

Thanks!

submitted by /u/Rubendarr
[link] [comments]
❌