❌

Vista de Lectura

Hay nuevos artΓ­culos disponibles. Pincha para refrescar la pΓ‘gina.

BIG Midjourney Updates & New (Free) Lipsync Tool!!

πŸ’Ύ

Today, I'm diving into some HUGE Midjourney V7 updates that almost feel like a V7.1. Plus, I've got a first look at a seriously impressive new lip sync model you can try right now!We'll also check out a cool new feature from today's sponsor, Recraft – don't worry, it's actually pretty neat! What's Inside This Video: Midjourney V7's Big Update: While we're still waiting for the much-anticipated Omni Reference (aka Subject Reference or working Siri!), Midjourney dropped some significant improvements to V7. I'll break down what's new and if it fixes those 'unrealistic poses' and 'blotchy background characters'.   NEW Experimental Aesthetics Command: Meet --exp! I tested this new command that works like 'stylize' to add dynamic elements and creativity. Find out the best values to use (hint: less might be more!).   Editor Improvements: Took a quick look at the updated Midjourney editor, focusing on the texture feature which is pretty cool for visual explorations. Got editor tips? Share them below!   V7 Model Polish: Are the Gumby characters and background weirdness gone? We'll see how the updated V7 model stacks up.   Omni Reference Sneak Peek: It's delayed, but I got a glimpse of the upcoming Omni Reference feature using CEO David Holtz as the subject! Looks promising and retains that classic Midjourney aesthetic.   Tavis Hummingbird Lip Sync: Checked out the new Hummingbird lip sync model from Tavis. It's available on their site (with a free tier!) and Fall. See how it handles my head bobbles vs. an AI character!   Chapters: 00:00 – Intro: Big Midjourney Updates & New Lip Sync! 00:35 – Tracking V7 Since Launch 01:11 – V7's Initial Quirks 01:30 – The Latest V7 Update (Feels like 7.1?) 01:41 - Where is OmniReference? 02:16 – NEW Command: Experimental Aesthetics (--exp) 02:31 - What is --exp? 02:41- Toast Test! 03:10 – Testing --exp Values (Toast Experiments!) 03:37 - Experimental Toast! 04:04 - Maximum Toast! 04:22 – Finding the Sweet Spot for --exp 04:34 - v7 old prompts now with --exp! 05:32 – Midjourney Editor Updates & Texture Feature 05:57 -Edit and Retexture Tutorial 06:46 – V7 Model Improvements: Less Jank? 07:23- But is it Perfect? No it is not. 07:51 – Omni Reference SNEAK PEEK! 08:43 – Other Reference Options (Runway Recap) 08:55 - Recraft New Feature 09:14 - Recraft Infinite Style Library 11:43 – NEW Lip Sync: Tavis Hummingbird Model 11:54 -Hummingbirds at the White Lotus 12:31 - Where you can use it for FREE! 12:59 – Testing Hummingbird (Me vs. AI Character) 13:16 - Result - AI TIM is here! 13:43 - Where Hummingbird flies! 14:18 – Final Thoughts & Call for Lip Sync Models! Resources Mentioned: Recraft.AI: https://go.recraft.ai/theoreticallymedia_styles use code: β€œMEDIA11” for $11 OFF any paid plan!! Runway References Video: https://youtu.be/umhFIUudEwo Tavis Lip Sync: Try Tavis Hummingbird (Free Tier!) : https://www.tavus.io/model/hummingbird I promised to cover the rolling Midjourney updates, and while Omni Reference keeps us waiting, this V7 refresh and the new --exp command are definitely worth exploring. The Tavis Hummingbird lip sync also looks like a strong contender!What do you think of these updates? And what other lip sync models should I check out for a future shootout? Let me know in the comments below! Don't forget to Like this video if you found it helpful, Subscribe for more AI deep dives, and hit that notification bell!

Runway's GAME CHANGER for AI Video Is Here!

πŸ’Ύ

Ever struggled with keeping characters and objects consistent in your AI videos? Well, guess what? Runway just dropped a potential game-changer: the References feature! It might just be the solution we've all been waiting for. In this video, I'm diving deep into Runway's new References tool. We'll explore: What References does: How it uses Runway Frames to generate a first frame with your character/object reference.   Getting Started: Simple steps to use the feature (hint: don't just drag and drop!).   Putting it to the Test: See how it handles single characters, multiple characters (like our man in the blue suit and his wolf companion!), and different styles.   Tips & Tricks: Learn how to potentially avoid issues like attribute bleed and use styles effectively.   Advanced Techniques: Combining references with specific locations, using character sheets, and a sneaky trick for establishing shots.   Community Showcase: Check out awesome examples from Dave Clark, Tom Likes Robots, and Andy McNamara.   Is it perfect? Not yet. But it's a massive step forward for AI video creation and character consistency. Ready to see if Runway cooked with this one? Let's taxi down the runway and find out! Chapters: 00:00 - Intro: The AI Video Consistency Problem Solved? 00:23 - Runway's Recent Updates (Frames, Gen4)   00:38 - Introducing the References Feature   00:51 - Reference First Release and Congrats to the 48 gang! 01:13 - References Wide Release 01:26 - How References Works (Using Frames)   02:02 - Simple Example: Man in Blue Buinesss Suit   02:37 - How to Use References Correctly   02:56 - One Character Reference 03:32 - Video Output 03:58 - Multiple References: Adding the Wolf   04:12 - Prompting the Wolf 04:37 - Video Output 05:00 - Using Styles with References 05:26- CSI K9   05:39 - Pushing Styles to Break 06:00 - Generating a Crime Movie 06:25- Challenges: Bad Outputs 07:00 - Attribute Bleed? 07:10 - The solve with a new Character 07:33 - Creating Cinematic Scenes (Master & OTS Shots)   07:52- Creating Locations to use as Reference 08:31- Video Output 08:52 - Gen4 vs. Turbo Model Comparison   09:13 - Does Generating in Frames Matter? 09:38 - Video Output 10:12 - Pro Tip: Using Character Sheets & Establishing Shots   11:00 - Sneaky Tip! 11:38 - Video Output 11:41 - Community Examples   12:44 - Final Thoughts: Did Runway Cook?   Resources Mentioned: My Full Breakdown on Runway Frames: https://youtu.be/7jwSNb4qq_E Let me know in the comments: What do you think of Runway's References feature? Have you tried it yet? What are you excited to create?

One Minute AI Video Is HERE & It’s FREE/Open Source!

πŸ’Ύ

AI Video's ten second generation wall officially SMASHED! I'm diving into something truly game-changing today: FramePack. This open-source tool lets you generate AI videos up to (and even beyond!) ONE MINUTE long, right now, for FREE. Forget those short clips – we're talking serious length here, and it's compatible with generators like Wan, Hunan, and more.   In this video, I'll break down: How FramePack overcomes the old drifting and coherence issues using cool tech like anti-drifting sampling.   How YOU can get it running, whether you have an Nvidia GPU (even with just 6GB VRAM!) using Pinokio, or if you're on a Mac using Hugging Face.   Step-by-step guides for both installation methods.   Tips for using the tool, including dealing with Tea CachΓ© for better results (or maybe turning it off!).   Lots of examples, including successes and some… well, let's call them "learning experiences" (dancing girl goes exorcist, anyone?).   Limitations I found, like issues with tracking shots.   This tech is brand new and evolving fast, but it's already opening up incredible possibilities for longer-form AI video. Let's explore it together!   Chapters: 00:00 - Intro: Breaking the AI Video Length Barrier!   00:46 - Meet FramePack: The Game Changer?   01:24 - How FramePack Works (No More Drifting!)   02:01 - What questions does Framepack ask? 02:44 - Hardware Requirements (You Might Already Have It!)   03:08 - Method 1: Installing with Pinocchio (Nvidia Users)   04:18 - Method 2: Using Hugging Face (Mac & Others)   05:49 - Using Frame Pack: Settings & Tips (TeaCachΓ© Explained)   07:27 - Generation Examples & Experiments (Hourglass Timer!)   08:36 - More Examples: Detective 08:56 - Dancing Girls 09:42 - TeaCachΓ© On vs. Off Comparison   10:10 - Known Limitations (Tracking Shots)   10:45 - Some great use cases (Moving Pictures) 11:20 - Endless Lofi Girl 11:54 - This is not the only one: The TTT Model   Resources Mentioned: Pinokio Installer: https://pinokio.computer/ Hugging Face Space (Duplicate This!): https://huggingface.co/spaces FramePack Paper: https://github.com/lllyasviel/FramePack TTT (Tom & Jerry) Paper: https://test-time-training.github.io/video-dit/

The AI Video Tool for WILD Camera Shots! Higgsfield Deep Dive & Review!

πŸ’Ύ

Some WILD AI Camera Here! Start your 7-day free trial with HoneyBook and see the difference - https://bit.ly/HoneyBook-Theoretically_media Today I'm diving deep into HiggsField, an AI video platform that's making waves with some seriously wild virtual camera movements! Ever seen those cool, hyper-stylized AI videos popping up on your feed? Chances are, they came from HiggsField.   I spent some time exploring whether this platform is just a gimmick or a genuinely useful tool for creators. Is it all just cherry-picking, or can it really deliver? Let's find out!   LINKS: HiggsField: https://higgsfield.ai escape Platform: https://escape.ai/ Paige Piskin : https://x.com/PaigePiskin/status/1912313290409062697 My thanks to today's sponsor: HoneyBook is more than just a CRM platform. It's more of a behind-the-scenes partner to eliminate busywork, reclaim your time, and keep you focused on what truly matters. https://bit.ly/HoneyBook-Theoretically_media In this video, I'll walk you through: What is HiggsField? A look at the platform and its niche in hyper-stylized shots.   Exploring the Presets: From explosions and bullet time to the... unique... tentacle effect. (Spoiler: I get a little Lovecraftian later! )   Testing the Camera Movements: Putting presets like Zoom Out, Kiss POV (yes, really!), 360 spins, Snorri Cam, Disintegration, and Glass Shatter to the test.   Bullet Time Deep Dive: My personal favorite effect! We even recreate a Matrix scene (with a surprise Dwight Schrute lookalike!).   Mixing & Matching Effects: Combining presets like Building Explosion and Arc for dynamic shots.   Overcoming Limitations: Tackling the 5-second clip limit with the last-frame-first-frame trick and discussing potential extension solutions.   Style Consistency: Does it maintain styles like anime, or does it lean towards realism?   9:16 Format: Why Higgs Field shines for TikTok and Instagram content.   Is it Worth It? My thoughts on using Higgs Field for cinematic projects vs. social media, plus a breakdown of the pricing.   Chapters: 00:00 - Intro: Wild Camera Moves! 00:44 - Exploring Higgs Field Presets 01:34 - Useful vs. Gimmicky Effects 02:00 - Testing: Noir Detective Zoom Out 02:35 - Zoom and Arc 03:15 - Michael Bay Level Explosions! 03:54 - The Virtual Smooch: Kiss Preset 04:46 - 360 Spins 05:05 - Snorri Cam Showcase 05:58 - Disintegration 06:22: Tentacles (Lovecraft Time!) 07:08 - HoneyBook 08:52 - Bullet Time Fun (Matrix BTS!) 09:48 - Side Note: John Gaeta & escape.io 10:24 - Cool Effect: Glass Shatter 10:49 - Mixing Presets: Explosions + Arc 11:41 - Extending Clips: The 5-Second Limit 12:19 - Last First Frame 13:15 - Style Consistency 13:51- 9:16, where this flies 14:10 - But still good at 16:9 14:35 - Pricing & Final Thoughts HiggsField offers some genuinely impressive (and sometimes hilarious) tools for creating eye-catching AI video content, especially for shorter formats. While the 5-second limit is a hurdle, the unique camera movements and effects make it a platform worth exploring, even if just for adding some 'zazz' to your projects.   What do you think? Is Higgs Field a useful tool or just a fun gimmick? Let me know in the comments below! πŸ‘‡ Don't forget to LIKE πŸ‘ if you found this helpful and SUBSCRIBE for more AI video explorations!

Kling 2.0 is HERE! Is it STILL the BEST AI Video Generator?!

πŸ’Ύ

Kling 2.0 just dropped, and you know I had to dive right in to see if the reigning champ of AI video generation got even better. In this video, I'm putting Kling 2.0 through its paces, comparing it to the previous 1.6 model (which was already crushing it!) and seeing how it stacks up against the competition like Google's VEO 2. We'll look at text-to-video, image-to-video, new features, and find out what's still missing and what's coming next. Let's see if Kling keeps the crown! πŸ‘‘   Chapters: 00:00 - Intro: Kling 2.0 Arrives!   00:41 - First Look: Fidelity & Coherence Boost   01:03 - Text-to-Video: Game of Thrones Direwolves (and Real Life!)   01:26 - Direwolf Results 01:59 - Text To Video Test 2 02:13 - Image-to-Video Quick Test (vs. VEO 2 Style)   02:59 - Overall Text To Video Impressions 03:15 - Text-to-Video: Strengths & Weaknesses (Cowboy Cosplay?!)   03:43 - Coherent Fast Motion: Kung Fu Fight Scenes!   04:41 - More Examples (Thanks Justin!): Punching Bags & Blizzards   05:19 - Running Test 05:50 - Generation Specs & Camera Control   06:09 - Lens and Camera Motion Callouts 07:14 - Image-to-Video Deep Dive (ft. Kolors 2.0) 07:50 - Sci-Fi Vibe Short 08:32 - Saucing Up the Footage 09:02 - Less Generations 09:22 - Impressive Walking Cycles   09:58 - Stylistic Retention with Image to Video 10:49 - AI Girlfriends and Crusty Pirates 11:45 - Final Frame Trick 12:26 - NEW Multi-Elements Feature   13:55 - Final Verdict: Still the King?   What I Cover: Comparing Kling 2.0 vs 1.6 and Google VEO 2.   Testing Text-to-Video: Examples range from direwolves and stock footage to kung fu and slightly awkward cowboys.   Exploring Image-to-Video: Checking consistency, character animation, style adherence, and complex motion like walking and fighting.   Camera Control & Lens Simulation: Does asking for an "85mm lens" work?   New Multi-Elements Feature: A quick look at swapping elements between videos (using the 1.6 model for now!).   Creating a short sci-fi piece using Kling 2.0, Udio for music, and Topaz for upscaling.   What do YOU think of Kling 2.0? Is it still the best AI video generator out there, or are you using something else? Let me know down in the comments! πŸ‘‡   Don't forget to LIKE πŸ‘ if you found this helpful and SUBSCRIBE for more AI explorations! Thanks for watching! Tim!

Massive Veo-2 News From Google & $300 of FREE Use?!

πŸ’Ύ

Google's Cloud Next event just wrapped, and wow, did they unload a ton of AI news! It felt a bit like a shotgun blast of announcements, but I'm honing in on the cool creative AI stuff – the latest on Veo 2, Imagen 3, and a brand new AI music generator called Lyra. Did Google cook? Yeah, they did! But can you get a plate easily? Well... maybe it's more like the drive-thru for now.   In this video, I break down the biggest creative AI takeaways from Google Cloud Next: What's New: We got updates on the already impressive Veo 2 text-to-video, the next iteration of their image generator (Imagen 3), Chirp 3 for text-to-speech, and the new Lyra AI music tool. Google's flexing that they're the only platform doing video, music, image, and speech natively... sort of.   Veo 2 Power-Ups: I dive into the new Veo 2 features like inpainting (think magic eraser for video!), aspect ratio changes, camera controls, and extending video length. Plus, a cool look at how Veo 2 helped bring The Wizard of Oz to the Sphere!   Imagen 3 & Chirp 3: Imagen 3 gets better detail and even image editing capabilities now. Chirp 3 lets you create custom voices with just seconds of audio.   Lyra Music: Google's answer to AI music? It generates short, instrumental jingles – maybe not dropping fire beats just yet, but useful for content creators.   Vertex AI Access: Here's the catch – most of this cool stuff lives on Google's Vertex AI platform, which looks enterprise-y but is actually accessible.   FREE AI Fun?: The best part? I show you how you might be able to test drive Imagen 3 and the powerful base Veo 2 on Vertex AI for FREE right now using their trial credits (seriously, my cost is $0 so far!). No guarantees on how long this 'open cookie jar' lasts, though!   Gemini 1.5 Pro: Also, Gemini 1.5 Pro is widely available and has honestly become my daily driver LLM. I even show a neat trick using it with Veo 2. GOOGLE VERTEX: https://cloud.google.com/vertex-ai?hl=en   Chapters: 00:00 - Intro: Google Cloud Next AI Overload! 01:06 - Veo 2 & The Wizard of Oz Sphere Project 02:13 - Gemini 1.5 Pro Goes Wide 02:32 - Creative AI Suite: Veo 2, Imagen 3, Chirp 3, Lyra 03:19 - Lyra AI Music Generator: First Look 04:17 - Imagen 3 Gets Updates & Editing 04:38 - Chirp 3: Custom Text-to-Speech 04:50 - Veo 2 NEW Features: Inpainting, Camera Controls & More! 05:30 - Inpainting in Veo-2 05:52 - Expanding the Frame in Veo-2 06:21 - Camera Controls in Veo-2 06:40 - First Frame Last Frame 06:57 - Extend Video in Veo-2 07:27 - Editing Toolbar in Veo-2 07:43 - Accessing Google's AI: The Vertex Platform 08:48 - How to Use Vertex AI (FREE Veo 2/Imagen 3?!) 10:43 - Cool Trick: Gemini + Veo 2 Prompts 12:01 - Final Thoughts: Grab the Free AI While You Can! What do you think about these Google AI updates? Are you going to try accessing Veo 2 or Imagen 3 on Vertex AI? Let me know in the comments below! Don't forget to LIKE this video if you found it helpful and SUBSCRIBE for more AI adventures!

AI QUAKE Is Here & You Can Play It! Is it a "GameChanger?"

πŸ’Ύ

Microsoft just dropped an AI bomb on an Id Software classic! They've unleashed an AI-powered version of the legendary Quake II, and YES, you can play it RIGHT NOW! But should you? πŸ€” Let's dive into this wild experiment, ask the big "why," and try to figure out where this whole AI gaming thing is headed. Grab your Railgun and some Quad Damage – it's gonna be a trip! What's Going On with AI Quake II?! This isn't your daddy's Quake (unless your dad is super into cutting-edge AI!). Microsoft Research, the wizards behind the curtain, have unleashed "Muse," their AI world agent, to recreate Quake II. Following in the footsteps of Google's AI Doom, this project is seriously mind-bending. But hold on a sec... is it actually Quake II? Let's break it down! Chapters: 00:00 – The AI Quake II Drop! - Microsoft's surprise release and initial thoughts. 00:23 – What Exactly IS AI Quake II? - Diving into the tech and its predecessors like Google's AI Doom. 00:45 – A Quick Quake II History Lesson - A blast from the past on the original game's impact. 01:15 – From Xbox to AI: The Evolution - Tracing Quake II's journey to this AI iteration. 01:56 – Meet Muse: The AI Behind the Magic - Exploring Microsoft's AI world agent and its upgrades. 02:21 – Tech Specs & What This Isn't - Clearing up confusion: it's research, not a polished game! 02:52 – AI-Generated Worlds: The Core Concept - Understanding how this interactive AI video works. 03:10 – Muse's Level Up: Faster and Bigger! - Examining the improvements in frame rate and resolution. 03:42 – Training Time: From Years to a Week?! - The surprisingly short training period for this AI model. 04:12 – The Big Question: Is It a Good Game? - Spoiler alert: not quite yet! 04:21 – The Janky Reality: Object Permanence & More - Facing the current limitations of AI game generation. 04:56 – Hands-On (Kind Of): My Hilarious Gameplay Experience - Trying out the keyboard-only controls and weird glitches! 06:05 – Game Over (Literally): The Short Play Session - Highlighting the limited playtime. 06:15 – Why Does This Exist?! The Big Picture - Exploring Microsoft's vision for AI and game preservation. 07:00 – The Future of AI in Gaming: My Thoughts - Reflecting on the progress and what's to come. 07:56 – The "Can It Run Crisis?" of AI Games! - My proposal for a new AI benchmark. LINK: Previous Video On MUSE: https://youtu.be/wbfhdq7f7jA Play AI QUAKE 2: https://copilot.microsoft.com/wham?features=labs-wham-enabled The Nitty-Gritty (Tech Specs & More): This AI Quake II is a research project by Microsoft's "Muse" (formerly WHAM!). It generates interactive video at roughly 10 frames per second (up from 1!). The output resolution is 640x360 (doubled from the previous model). The Quake II model was trained on about one week of Quake II video. Controls are currently limited to keyboard input (WASD for movement, arrow keys to look, F to fire). Object permanence is a major challenge – if you look away for less than a second, things disappear! Health values and other in-game counters aren't super reliable. Gameplay sessions are currently limited in duration. Why Bother with AI Quake II? While it's not exactly replacing your gaming rig anytime soon, this project hints at a fascinating future. Phil Spencer from Microsoft Gaming has talked about using AI to learn old games and make them playable on any platform, regardless of the original hardware. Think game preservation on a whole new level! The Future is Now (Kind Of): We're seeing rapid advancements in generative AI for games, from AI-powered NPCs to 3D asset creation. While a fully AI-generated AAA title is still a ways off, projects like Muse and Google's Genie are laying the groundwork. Could this be the beginning of the "ChatGPT moment" for video games? I think we might just be seeing the very first sparks

Midjourney V7 Is Here! Was It Worth The Wait?

πŸ’Ύ

Midjourney V7 is finally here, and I'm diving deep! I'm testing out the new features, including the personalization profile and the super-fast draft mode. Is it worth the hype? Let's find out! In this video, I'll cover: My first impressions of Midjourney V7. The new personalization profile setup. A look at the draft mode and its speed. Prompt tests and image quality comparisons. What's missing and what's coming next. If you're into AI art, you won't want to miss this! Hit like, subscribe, and share your V7 thoughts below! Chapters: 00:00 – Intro: Midjourney V7 Arrival! 00:24 – Just How Long Has It Been? 00:55 - The Landscape Has Changed 01:32 – Key Features Overview 01:45 - Getting Started With Personalize 02:35 - Options Toolbar Changes 03:10 – Prompt Testing Toast 03:46 - Prompt Test Beauty 04:11 - Prompt Testing Surreal 05:03- Prompt Testing Wizards 06:09 - Hard Prompt Test 06:44 - New Feature: Draft Mode 07:32 - Draft Mode Test 07:55- Missing Features 08:13 - More Hard Prompts 08:45 - Blurry Faces 09:04 - Impressions 09:36 - What is Coming Up for Midjourney? 10:04 - Midjourney Video 10:42 - Is Midjourney the Best Image Generator Around? 10:53 – Does "the best" matter? 11:08 - Coming up Next

Runway COOKED for GEN-4! Ultimate Deep Dive & Review!

πŸ’Ύ

Today, we're diving deep into RunwayML's Gen 4, the latest update to their AI video generation model. Runway claims that Gen 4 boasts improved fidelity and output quality. More importantly, Gen 4 introduces better consistency in character, location, and color grading.   While the highly anticipated reference feature isn't available at launch, there's still plenty to explore. I'm putting Gen 4 through its paces with a series of tests, checking out community outputs, and sharing some best practices.   We'll take a look at everything from image to video generation, character consistency, and even some fun experiments with text and effects I’ll also be comparing Gen 4 to previous Runway versions like Gen 3 and Gen 2.   Runway's Gen 4 is a significant step forward, and I'm excited to share my findings with you. There are still more updates on the way. So, buckle up, and let's jump into the world of Gen 4!   RunwayML: https://runwayml.com/ Runway's Frames Video: https://youtu.be/7jwSNb4qq_E 00:00 - Intro to Runway Gen 4 00:29 - Gen 4 Overview and Key Features 01:17 - Gen 4 at Launch 01:48 - Frames to Video (Man In the Blue Suit) 01:59 - First Video Test 02:24 - Frames Image Generation 02:44 - London Test 02:58 - Gen 4 vs. Gen 3 vs. Gen 2 04:02 - Gen-4 Text (Woman in the Red Dress) 05:13 - Impressive Wide Angle Shot 05:48 - Gen 4 Wonk 06:29 - Expand Feature With Gen-4 07:00 - Fight Sequences (Raging Bull) 07:32- Noir Test & What Reference Might REALLY Be 08:15 - Noir Test 2 08:42 - How Powerful Will Reference Be? 09:55 - Image to Video 10:29 - No Prompt with Image To Video Test 11:07 - Working with Multiple Faces 12:13 - Grandma with a Flamethrower 12:21 Style Consistency with Image to Video 12:52 - Cyberpunk Woman 13:55 - Gen-4 To Restylize 13:59 - Community Outputs Showcase 15:58 - Final Thoughts on Gen 4

This AI Image/Video Platform Opens Up So Many Possibilities (And You Can Try It For Free!)

πŸ’Ύ

Let's take a tour of one of the most interesting AI generation platforms I've come across: Flora. If you're into AI art, image editing with Gemini, and even video generation, this platform has it all.   I know the workflow might look a bit intimidating at first – all those Blocks and connections! But trust me, it's not as scary as it seems. I'll walk you through the basics, from generating images with text prompts to using cool features like style reference and video-to-video generation.   We'll be playing around with Gemini's editing features, creating our own character references, and even training a custom style. Plus, I’ll show you how to take those images and turn them into videos!   Whether you're into clean, organized workflows or a bit of creative chaos, Flora has something for you. And the possibilities? Let’s just say they got me thinking about fight sequences and character turnarounds!   Check out Flora for yourself! They offer 2000 free monthly credits so you can explore the platform and see if it’s a good fit for you.   LINKS: FLORA: https://tinyurl.com/floratim CHAPTERS 00:00 - Intro to Flora: An AI Playground   00:28 - The AI Murderboard Isn't Scary 00:52 - Understanding the Workflow   01:19 - Text to Image Generation with LLMs   01:55 - Image Generation Options 02:42 - Video Generation Options 03:21 - Organization on a Canvas  03:53 - Gemini's Image Editing Features   04:46 - Flora Styles 05:25- Cinematic Crime Film Style 06:30 - Training Your Own Style   07:33 - Testing Out a Custom Style 08:27 - Video Generation Tests 08:57 - Off Book Video to Video Generation   09:50 - Character Edits With Gemini 11:26 - Impressive Turnarounds! 12:37 - Thoughts on Turnarounds Use Cases 12:57 - Advanced Flora Technique 13:25 - Storyboarding with AI   13:54- Flora's Free and Paid Plans 14:12 - Next Up for Flora

Behind The Scenes Of An AI Film (Masterclass & Cost Breakdown!)

πŸ’Ύ

Hey everyone! I'm super excited to finally share a full production breakdown of my AI short film, "The Bridge." The response has been incredible (almost 400,000 views across platforms!) and I'm here to spill all the secrets on how I made it. In this video, I’ll walk you through the entire process, from pre-production and AI tools, all the way to final post-production. I’ll also dive into the costs and compare them to traditional filmmaking.   If you haven’t seen the short yet, don’t worry, I’ve included it in the video! My Thanks to Recraft for Sponsoring Today's Video! Check out Recraft – https://go.recraft.ai/media. Use my code MEDIA for $12 off any paid plan. LINKS: Jason Zeda's Wu-Tang Clan Video: https://youtu.be/ZBTb_xJBh5c?si=kqZPfc0h30F2oTBN Henry’s Prompt Tweet: https://x.com/henrydaubrez/status/1894513057109348405 Google Labs Discord: https://discord.gg/googlelabs Topaz Upscale: https://topazlabs.com/ref/2518/ ElevenLabs: https://try.elevenlabs.io/w5183uo8pydc Hume: https://www.hume.ai/ Hedra: https://www.hedra.com/ CHAPTERS 00:00 – Introduction   00:37 - The Bridge Views 01:20 - The Bridge 03:37 – Pre-Production   04:11 – AI Tool Selection   05:10 - Re-Inspired 05:31 – Prompt Engineering   06:13- Working with an LLM on a Film 06:33 – Veo-2's Super Power 07:50 – Achieving Visual Consistency   09:10 – Audio Production   09:52 – Lip Syncing with Hedra   10:32 - Upscaling 11:16 – More Upscaling 11:52 - Making a Poster w/ Recraft 13:23 – Post-Production in Premiere Pro   13:55 - Death's Voice 14:40 – Is it Perfect? 15:25 - Cost Breakdown 17:01 - Comparison To Traditional Filmmaking 18:11 – Make Movies 18:55 - It's just not there yet

OpenAI's Stunning Image Model Has Some Cool Tricks!

πŸ’Ύ

OpenAI has dropped a brand NEW AI Image generation model, and it's NOT Dall-E 4! In this video, I'm taking you on a deep dive into this remarkable AI Model and showing you EVERYTHING it can do. We'll explore its impressive capabilities, uncover some hidden features, and yes, even talk about its quirks. Is this the end for avocado chairs?! In this video, I cover: * OpenAI's new AI image generator (it's not Dall-E!) * Image generation AI tools and comparisons * AI art techniques and workflows * Text-to-image AI generation * AI character design and consistency * Sora AI video generation updates * AI art community highlights * New AI tools and software This new AI is seriously impressive, but it also has some limitations. I'll show you what I've discovered in my testing, including how it handles text, image referencing, and creating consistent characters. Plus, we'll take a look at the latest Sora updates and some incredible AI art from the community. If you're interested in AI art, image generation, or just the latest in AI tech, you NEED to watch this! LINKS: THE BRIDGE: https://youtu.be/YDlME4qvER8 OpenAI: https://openai.com/ Reve: https://preview.reve.art/app Ideogram: https://ideogram.ai/ #AI #ImageGeneration #OpenAI #Dalle #Dalle3 #Sora #ArtificialIntelligence #TechReview #AIArt #Stablediffusion #Midjourney #AItools #TexttoImage #CreativeAI #aimusic Chapters: 00:00 - Intro: OpenAI's New Image Generator - Is Dall-E Dead? 00:18 - Quick Announcement: My AI Short Film - The Bridge 00:41 - Goodbye Dall-E? First Look at the NEW AI Model 01:28 - Image Generation Tests: Blue Suit Guy, Samurai & More! 02:11- Controls and Options 03:04 - Samurai Test 03:15 - Clown with a chainsaw 03:45 - AI Challenge: Complex Prompts with the Woman in the Red Dress 04:17 - Remixing for Different Angles 04:49 - Creative AI: Underwater Scenes, 90s Nostalgia & VHS Tapes 05:24 - The VHS Tape 05:48 - Text Generation is INSANE! Alan Wake Novelization Test 06:56 - GTA 7 Box Art Test 07:52 - Image Referencing EXPLAINED: How to Use Reference Images 08:04 - Scrambling Faces 08:26 - Not John Wick 09:04 - Multiple Image References 10:24 - Illustrated to Photo with Image Reference 11:13 - Sora Time 11:48 - Community Outputs 13:04 - More AI Tools You NEED to Know: Reve, Ideogram 3.0

THE BRIDGE: A Stunning AI Film Created with Veo-2.

πŸ’Ύ

Presenting: The Bridge. An AI Short film utilizing Google’s Veo-2. I’m really proud of this one, as my goal (as always) is to push storytelling, performance, and narrative in this emerging art form, and I feel like I managed to pull something close that off here. Every shot here utilized Veo-2, although there were a few post-generation tricks here and there. I’ll cover everything in detail in an upcoming video. In the meantime, I hope you enjoy the short.

Adobe Goes AI Dirty with Flux, Runway, & Google!

πŸ’Ύ

Adobe AI & Flux! Build your own web app with Hostinger Horizons: http://hostinger.com/tmedia & use Code TMEDIA for 10% off! In this video, I dive deep into some HUGE updates in the world of AI image and video generation! Adobe is surprisingly opening up its ecosystem to third-party models like Black Forest Labs (Flux), Google's Imagen 3, and RunwayML's Frames – right inside Adobe Express and Project Concept! I explain what this means for creators, especially the ability to finally "get dirty" with models beyond Firefly. Plus, I'm testing out Stability AI's brand new FREE virtual camera tool and showing you how it could revolutionize how we control camera angles in generated content. I even use it in conjunction with google's Gemini and Runway! Key Topics: Adobe's surprising partnerships with external AI model providers. What Flux, Imagen 3, and Runway Frames bring to the Adobe ecosystem. The implications for Photoshop and Premiere Pro (and what about Sora?). Hands-on with Stability AI's Stable Virtual Camera. Combining Stable Virtual Camera with Gemini 2.0 and Runway Gen 3 for enhanced control. Project Concept Beta Waitlist: https://concept.adobe.com/discoverStability AI Virtual Camera (Hugging Face): https://huggingface.co/spaces/stabilityai/stable-virtual-camera Previous Video (Gemini 2.0): https://youtu.be/llvyFBTyiGs Chapters 00:00 - Intro: Adobe Enters the Black Forest! 00:25 Adobe Embraces Third-Party AI Models 01:18 My Experience at Adobe's AI Summit 01:45 Adobe's New Partnerships: Flux, Imagen 3, Runway Frames 02:22 Where These Models Will Appear First 03:14 The Interesting Inclusion of Video: Veo2 04:04 - Potential Cost Savings with Integration 04:26 - Hostinger Horizon AI Web App Builder 08:27 Stability AI's Stable Virtual Camera 08:28 How to Use Stable Virtual Camera (Hugging Face Demo) 10:36 Limitations and Research Preview Status 10:48 Combining with Gemini 2.0 and Runway Gen 3 11:15 The Future of Camera and Subject Control 11:48 - A real look at what is coming

Google's FREE AI Image Game Changer Is KILLER

πŸ’Ύ

Is Google's Gemini 2.0 about to revolutionize image generation and editing? In this video, I'm diving deep into Google's latest AI release and explore its powerful capabilities. Some are calling it a Photoshop killer, but is it really? We'll break down what Gemini 2.0 CAN do, its limitations, and how you can start using it for FREE right now! We'll cover: Image generation quality and prompt coherence   Editing existing images (even Midjourney!)   Creating image sequences for AI video   Generating consistent characters for AI training (Laura)   Cool community examples & use cases!   Sneak peek at upcoming video generation features!   Whether you're an AI enthusiast, content creator, or just curious about the future of image technology, this video is for you! πŸ‘‡ Links & Resources: Google AI Studio: https://aistudio.google.com/prompts/new_chat πŸ‘ Like, Subscribe, and hit the notification bell for more AI content! Community Outputs Bilawal Sidhu: https://x.com/bilawalsidhu Victor M: https://x.com/victormustar Min Choi: https://x.com/minchoi Aiba Keiworks: https://x.com/AibaKeiworks Umesh: https://x.com/umesh_ai #Gemini2.0 #AIImageGeneration #AIArt #GoogleAI #Midjourney #Dalle3 #AITools #FreeAI #PhotoshopKiller #ArtificialIntelligence CHAPTERS 0:00 - Intro 0:31 - Gemini Goes Multimodal 0:45 - This is What OpenAI implied 1:33 - How To Access AI Studio 1:55 - Rate Limits 2:15 - Image Tests With The Man In a Blue Business Suit 3:20 - Generating Video With Luma Labs 3:38 - Cinematic Angles with Midjourney Images 04:31 - Image Fidelity Loss and How To Overcome 05:17 - Things can still be wonky 05:30 - Tips and Tricks With Rerolling 06:19 - Using Real Photos 6:37 - Using This For Video Keyframes 7:14 - Using 3 keyframes in Runway 7:35 - Speedramping 8:02 - The Gamechanger for LoRAs 9:02 - Community Outputs 10:27 - Video is Coming!

AI's Next Horizon Is Real Time Game Characters

πŸ’Ύ

Is AI the future of gaming? Sony's AI demo sparked a huge debate! We're diving deep into the tech, the controversy, and how YOU can create even BETTER AI-powered characters using readily available tools. Join us as we explore the exciting (and sometimes hilarious) possibilities of AI NPCs! In this video, we'll break down: Sony's AI demo and why it got mixed reactions. The technology behind the demo: OpenAI's Whisper, ChatGPT, Llama, and more. How to create your own AI characters with realistic voices and lip-sync using tools like Hume and Hedra. The future of AI in gaming, including real-time reskinning and interactive NPCs. Get ready to see how AI is transforming the gaming landscape! Thanks to Hume for Sponsoring Today's Video: Go get some Free credits: Hume AI: https://try.hume.ai/yooyhnds8jtd Links: Hume AI: https://try.hume.ai/yooyhnds8jtd Hedra: https://www.hedra.com/ Krea.ai: https://www.krea.ai/ Magnific: https://magnific.ai/ Previous Hume Video: https://youtu.be/KQjl_iWktKk Tyyle's Cyberpunk Reskin: https://x.com/tyyleai/status/1898483733243502713 Runway Restyle Video: https://youtu.be/EsPGN6NdIgM 00:00 - Intro 00:21 - Sony's AI Controversy 00:53 - Horizon Zero Dawn: A Brief History 02:01 - The Demo Clip 02:25 - A Note on Fair Use 02:44 - Analysis of the Demo 03:07 - Tech Behind the AI Demo 03:34 - Face Animation and PS5 Capabilities 04:07 - This is a Glimpse of What Is Possible 04:34 - Building a Higher Fidelity AI Aloy 05:38 - Upscaling With Magnific 05:50 - AI Voice Generation 06:18 - Creating Original Voices With Hume 07:16 - The Horizon Zero Dawn Audiobook with Hume 07:40 - Hume Pricing (Free Tier!) 08:14 - Creating the Conversation 08:28 - Insane Lipsync with Hedra 08:42 - Hedra Quick Demo 09:06 - My Convo with AI Aloy 10:02 - Differences 10:13 - A Glimpse Into the Future 10:54 - Outro

This free AI Speech is Wild! Acting & Emotion Performances!

πŸ’Ύ

Today I'm looking at the new AI text-to-speech model, Octave from Hume, and I'm pretty impressed. If you're like me and want AI voices that actually have some feeling behind them, you're going to love this. Plus, the free tier is super generous, and the paid options are incredibly affordable. Let's check it out! In this video, I'm giving you a full walkthrough of Octave, showing you how it balances linguistic accuracy with genuine emotional understanding. We'll listen to some demo voices, create our own custom voices, and even test out the project features for longer content. Get ready to say goodbye to flat, robotic AI voices! LINKS: HUME: https://try.hume.ai/yooyhnds8jtd Key Highlights: Octave by Hume: A new text-to-speech model that focuses on emotional understanding. Generous free tier: 10,000 characters and unlimited custom voices. Affordable paid plans: Perfect for creators of all levels. Customizable voices: Create unique voices with specific tones and accents. Project features: Ideal for longer projects like audiobooks and podcasts. SEO Keywords: AI voice, text to speech, AI speech, voice generation, AI voice generator, Hume Octave, AI audio, voice design, AI podcast, AI audiobook. Here’s what we cover: 00:00 – Introduction: The Problem with AI Voices and Octave's Solution 01:16 – Octave in Action: Demo and First Impressions 02:04 – Exploring the Hume Interface: Quick Start and Speech-to-Speech 03:51 – Diving into the Octave Playground: Preset Voices 04:11 - Preset Voices Continued 04:53 – Custom Voice Design: Creating Your Own Unique Voices 05:36 - Name The Quote! 05:56 - Interesting Trick Quoth The Raven 06:53 – Advanced Text Prompts: Enhancing Voice Performance 07:33 - Acting Results 08:14 – Project Features: Octave For Longer Content (Audiobooks, Podcasts, etc.) 08:32 - Alan Wake The Audio Novel Backstory 08:58 - Alan Wake The Audiobook Sample 09:22 - Additional Trick with Projects 10:19 – Pricing and Plans: Affordable AI Voice Solutions 11:07 – Future Updates and Final Thoughts: What's Next for Octave What do you think of Octave? Have you tried it yet? Let me know in the comments below! Don't forget to like this video and subscribe for more AI content. And if you know a great lip sync model, please tell me about it!
❌