Vista de Lectura

Hay nuevos artículos disponibles. Pincha para refrescar la página.

The BEST AI Video Generator Is OUT & FREE! (with some catches...)

💾

Google's Stunning AI Video Generator, Veo-2 has finally been released and it's FREE! Although, there are a few interesting catches! For one, Veo-2 is available via the Youtube App, and is probably more aimed at Youtube Shorts. That said, I do have a solution capturing outputs so you can edit the videos on your own. Also, is this a Turbo Model? I'll take a shootout between the Web Veo2 model and the Youtube version to see. Spoilers: It likely is. LINK: Star Wars: The Ghost's Apprentice: https://youtu.be/KWlxMC0j498?si=d643alvIrhrQVZI8 Chapters: 0:00 - Intro 0:35 - Veo-2 Has Released! 1:09 - Youtube Shorts? 1:18 - Getting Started With Veo-2 1:52 - Imagen-3 2:22 - First Shot 2:53- Some Limitations 3:20 - The Timeline 3:56 - Adding Music 4:19 - Uploading 4:37 - I don't like this 4:51- Test Short 5:04 - Is this a Turbo Version? 5:20 Testing the same Prompt 5:56 - How Much Faster Is It? 6:05 - Moderation Problems 7:02 - Animated Presets 7:24.- Is it Perfect?

Luma Is The BEST AI Image To Video Now? Plus: Pika & Google Updates!

💾

Today we're diving into Luma Labs Dream Machines and their updated Ray2 model that’s pushing the boundaries of image-to-video technology. In This Video: • AI Image-to-Video Revolution: Discover how Luma’s Ray-2 model is setting the bar for image-to-video conversion—with stunning visuals and unexpected “happy accidents” that spark creativity. • Google’s Imagen 3: See how Google is upping its game with new inpainting capabilities, making it one of the best free AI image generators available today. • VFX Compositing Game Changer: Get a sneak peek at an upcoming model poised to transform VFX compositing, blending video elements seamlessly like never before. • Creative Process Insights: Follow the demo featuring Midjourney-inspired visuals, detailed project boards, and camera control tips to achieve that Michael Bay cinematic style. • Community Showcase: Enjoy a curated selection of impressive community outputs—from cyberpunk anime scenes to dynamic video compositing examples. LINKS: Luma Labs: https://lumalabsai.partnerlinks.io/dd1jzuzx6o87 Imagen3: https://deepmind.google/technologies/imagen-3/ Pika: https://pika.art/ Snap Research: https://x.com/AbermanKfir/status/1888987447763292305 CHAPTERS: 0:00 - Intro 0:37 - Luma Labs Image to Video Update 1:04 - Demo Video 1:27 - Luma Platform Walkthrough 2:00 - Midjourney Bayham 2:28 - First Generation with Luma's Image To Video 3:00 - More Generations 3:15 - Limitations and More Like 3:47 - Wonky Shots 4:04 - Camera Controls 4:52 - Motion and Outputs 5:30 - Community Outputs 6:00 - Animation 6:44 - Lunch on a Skyscraper 7:17- Imagen 3 Updates 8:21 - Inpainting 9:17 - DallE-4 incoming? 10:00 - Pika additions 10:31 - Another Example 11:01 - Jon Finger's impressive tests 12:15 - Snap Research - AI VFX!

AI Video’s Morphing Problem Is Over!

💾

AI Video Just Leveled Up! Learn more about Hostinger at https://hostinger.com/theoretically (Coupon Code: THEORETICALLY) In this video, I dive deep into the latest breakthroughs in AI video technology and creative tools that are reshaping digital storytelling. We start by tackling the age‐old problem of “noodle bone” character movements with Meta’s new Video Jam solution, then move on to incredible VFX demos that augment real-life footage (yes—even paintings!). I also share insider scoops on Midjourney’s upcoming video advances, updates on Omni Human’s dancing avatar tech, Topaz Labs’ next‐gen upscaling in Project Starlight, and even a look at agile robotics training using real-world sports data. Plus, don’t miss the fun reveal of AI Granny—the chatbot fighting phone scammers—and a walkthrough of an innovative AI storytelling platform that lets you craft your very own narrative (complete with comic-style visuals). My Thanks to Hostinger for Sponsoring Today's Video! If you’re passionate about AI, video production, VFX, or futuristic tech trends, this video is for you! LINKS: VideoJam: https://hila-chefer.github.io/videojam-paper.github.io/ TopazLabs: https://topazlabs.com/ref/2518/ ASAP Robots: https://agile.human2humanoid.com/ LoreMachine: https://www.loremachine.world/ 0:00 – Introduction & Overview 0:26 – Video Jam: Enhanced AI Motion 0:58 - How VideoJam Works 1:31 - Examples of VideoJam 2:37 - VideoJam Vs Other Models 4:00 - VideoJam Release? 4:34 - DYNVFX Video VFX Inpainting? 4:53 - DynVFX Examples 5:32 - Taking a closer look at DynVFX 6:29 - Midjourney and Veo2 6:47 - Tracking Upcoming Models 7:29 - Veo2 API? 8:07 - Building a Site With Hostinger 11:58 - OmniHuman Update 12:30 - Topaz Labs Stunning Upscaler 13:19 - ASAP Robots 14:46 - AI Granny 15:12 - LoreMachine: Ai Storytelling 15:55 - Character Image Generation 16:18 - Storyboards and Comic Books? Thanks for watching – and as always, ship it!

Deep Research SHOCKED Me With This Report!

💾

Dive into the latest in AI creativity and innovation in this deep-dive video where I put OpenAI’s new Deep Research to the test! In this episode, I explore cutting-edge AI video generators, cost analyses for cloud subscriptions versus local hardware setups, mind-blowing lip sync avatar technology, and the most impressive AI rotoscoping I’ve seen yet. Plus, I share my firsthand experience from Adobe’s Gen AI Leadership Summit and a Comfy UI event at GitHub HQ! LINKS DEEP RESEARCH PDF Reports Here: https://theoreticallymedia.gumroad.com/l/DEEP In This Video You’ll Discover: OpenAI Deep Research Overview: How OpenAI’s deep research release (powered by the latest model) is reshaping the creative AI landscape. AI Video Generation Showdown: A detailed report comparing popular tools (cling, runway, minimax, Gen three, etc.) on quality, pricing, and performance. Viral AI-Generated Films: Analyzing what made films like Airhead, The Heist, Queens of Avalon, and the AI-driven James Bond trailer go viral. Subscription vs. Local Machine: A breakdown of costs—including hardware specs (RTX 4090, Ryzen 9, and more!) and training time—to determine which is best for creators. Next-Gen AI Avatars & Rotoscoping: Check out the impressive Omni Human One lip sync demo and “Matt. Anyone”—the ultimate stable video matting technology. Behind the Scenes at Adobe HQ: My experiences and honest feedback from Adobe’s Gen AI Leadership Summit and the spirited discussions with industry pros. CHAPTERS: 0:00 - Intro 0:38 - Open AI's New Deep Research 1:49 - What is the Best AI Video Generator? 4:58 - Research Results 5:41 - How to Make a Viral AI Video? 7:05 - Research Results 8:07 - Additional Questions 8:34 - What is more Cost Effective, AI Platforms or Local Models? 9:28 - Research Results 12:30 - OmniHuman Video Lip Sync & Avatar 14:53 - MatAnyone 15:49 - Adobe GenAI Leader Summit 17:42 - Comfy UI Meetup

AI Music's FREE New King?

💾

A suprising new challenger to the AI Music Throne has appeared! In this video, we explore the surprise comeback of Riffusion, an AI music generator that’s challenging big names like Udio and Suno for the AI music throne. From instant full-track creation and advanced composition controls to remixing and extending your own uploaded audio, Riffusion offers a jaw-dropping blend of features—and it’s currently FREE (for now). Join Tim as he tests everything from hip-hop and country to power metal and personalized guitar riffs. If you’re passionate about the future of AI-generated music, you won’t want to miss this one! 00:00 – The King Is Dead! 00:22 – AI Music Evolution: Riffusion’s Return 00:39 – Let’s Hop in the Hot Tub… 00:58 – Riffusion Beta Overview 01:10 – Prompt Bar & Compose Tab 01:14 – First RiffFusion Prompt 01:40 – “July Can’t Come Fast Enough” 02:34 – Modern Country Shots Fired 02:53 – "Same Song at Ten" 03:39 – Hip-Hop Test 04:51 – Power Metal Coffee 05:33 – Ending the Song 05:59 – Instrumentals / Pushing Prompts 06:46 - Compose Tab (Advanced Mode) 07:25 – Blending Genres Test 07:59 – Uploading Audio 08:12 - Original Song 08:44 – Cover Song Prompt Overview 09:25 - Prompting Close To Genre 09:52 –Changing a Cover's Genre 10:31 – Inputting Your Own Music 10:40 - Little Peavey Wonder 11:00 - The Riff 11:20 - Riff To Fusion 11:39 - Riffusion Output 12:15 – Layering & Creative Possibilities 12:51 – Personalized Feature 13:12 – Free Beta & Final Thoughts What I Cover • AI Music Creation: How Riffusion generates entire songs in different styles. • Advanced Composition Tools: Demoing the “Compose” tab, the Weirdness Slider, and multi-genre blends. • Remix & Extend: Transforming public-domain tracks—or my own riffs—into something brand new. • Instrumentals & Vocals: From orchestral and synthwave to power metal anthems. • Practical Tips & Workflow Ideas: Layering AI outputs with original recordings. Why Riffusion Stands Out Riffusion’s unlimited free beta is a rare chance to experiment with AI music at no cost. Whether you’re a casual musician or a pro producer, now’s the perfect time to explore AI-driven music innovation. Try Riffusion (Beta): https://www.riffusion.com/?utm_source=theoreticallymedia&utm_medium=youtube&utm_campaign=launch

PIKA 2.1 is FIRE!!

💾

Today we’re diving into two big updates in the AI Video generation space. First up is Pika’s new 2.1 model which not only offers outputs in 1080p, but touts more realistic movement and better prompt coherence. Does it live up to the claims, and how does it stack with Pika’s “Ingredients” feature? Hailuo’s Minimax also has a cool game changer for Camera Control, and this isn’t motion brushes or sliders, it’s something much more interesting! CHAPTERS 0:00 - Intro 0:26 - Pika Introduces Version 2.1 1:02 - What is in Version 2.1 1:15 - Hopping into Pika 1:42 - Testing Text to Video with 2.1 2:29 - An 80s Sitcom Text to Video 3:39 - 70s Spy Film, Text to Video 4:24 - Testing Image to Video in Pika 2.1 5:08 - Stylized Image to Video in Pika 2.1 5:35 - Prompt Direction in Pika 2.1 6:02 - It might take a re-roll 6:22 - Pirate Woman Example 6:44 - Overall Prompt Understanding 7:04 - How is Pika 2.1 With Ingredients? 7:41- Viking Cop is Awesome 8:01 - Getting Closer with Ingredients 8:28 - Where I think 2.1 and Ingredients Excels 9:04 - Community Outputs 9:43 - Minimax Releases Director Mode 10:01 - Director Mode Overview 10:23 Director Mode Examples 11:14 - Community Outputs 11:55 - Closing Out

Real Time Game Changer & Films From A Prompt!

💾

AI Realtime Just Got Real! Learn more about Hostinger at: https://hostinger.com/theoretically (Coupon Code: THEORETICALLY) AI image generation is getting faster and more creative every day! In this video, we take a look at real-time AI image generation with Krea, an all-in-one AI generation platform. See how it works, what the limitations are, and how you can overcome them with a little ingenuity. Krea's real-time AI image generator has been around for a while, but it's had one glaring problem: consistency. If you make any changes to your prompt, the entire image changes. This can be a problem when you're trying to create a specific image, such as a character or an object. But Krea has solved this problem by allowing you to train up a model and then use it in the real-time module. This means you can now create images with consistent characters and objects, even if you make changes to your prompt. We'll also take a look at how to use Krea's real-time AI image generator to create 3D objects. This is a really cool feature that allows you to create 3D models of your images. Finally, we'll take a look at Film Agent, a multi-agent framework for end-to-end film automation in 3D virtual spaces. This is a really interesting project that could potentially revolutionize the way films are made. Chapters: 0:00 – Introduction Quick preview of today’s topics: real-time AI image generation in Krea, plus an open-source model for prompt-to-movie. 0:22 – Krea’s Real-Time AI Image Generation Overview of Krea as an all-in-one platform with real-time text-to-image functionality. 2:14 – Training Custom Models for Consistency Solving the biggest issue with real-time generation—keeping the subject consistent. 3:38 – Converting 2D Images to 3D Demonstration of how Krea can transform flat images (like a jet fighter) into movable 3D objects. 5:19 - Me in 3D 6:11 - Some Tips on Sliders 6:49 – Building a 3D Character from Concept Art Testing a Midjourney-inspired “Lara Croft”-style character sheet and training it in Krea for 3D manipulation. 8:33 Adding Elements into your Compostion 8:49 – Future Pose Control & Closing Thoughts on Krea Discussing next steps, potential pose/rigging controls, and overall impressions of Krea’s capabilities. 9:39 - Hostinger 12:50 – Film Agent: Multi-Agent Prompt-to-Movie Deep dive into Film Agent, an open-source framework that uses virtual “director,” “screenwriter,” and “actor” agents in Unity. 13:09 - How Film Agent Works 14:54 -Where Unity Comes In 15:29– Sora Remix & Visual Enhancements Running Film Agent’s output through Sora to experiment with AI-driven style and consistency. 15:56 - Sora Remix 16:23 – Pika 2.1 Model Reveal A quick look at Pika’s upcoming 2.1 release and why it has the AI art community excited.

Runway's Stunning Frames & Kling Kills It With Elements!

💾

Diving to the new Frames feature from Runway ML—an AI image generator built entirely by Runway, separate from models like Stable Diffusion or Flux. We'll look at how Frames offers cinematic image outputs, best practices for prompting, and showcases a variety of user-generated examples. On top of that, he explores: • Kling's New “Elements” Feature that allows up to four reference images and yields 10-second AI-generated videos. Tim shows how this boosts continuity in character designs and environments. • A John Wick–Severance Crossover made possible by advanced LoRA (Low-Rank Adaptation) techniques, demonstrating how entire characters (not just faces) can be swapped seamlessly in video footage. • Tribute to David Lynch: Tim takes a moment to acknowledge the legendary filmmaker’s influence and pays homage with a custom “Peak Lynch” style inside Runway’s Frames. Throughout the video, Tim mixes in behind-the-scenes tips—like how to handle glitchy hands, text consistency, and the best ways to write prompts for cohesive results. He also reveals a handy GPT Prompt Builder tool (linked below) designed to help generate descriptive prompts quickly. Finally, Tim closes by teasing a new Minimax feature for reference-to-video that’s just dropped, promising to showcase more examples soon. Timestamps: 00:00 – Intro & Overview of Runway’s Frames 00:58 – Diving into Frames + Prompting Tips 01:47 – Frames First Example 02:29 – Frames To Video 03:16 – Frames Second Example 04:08 - Frames Outputs and Vary 05:00 – Prompting In Frames 05:38 – Frames Prompt Builder 06:19 – Styles In Frames 06:44 - Your Own Styles 07:39 – More Cinematic Examples 08:09 – Community Outputs 09:23 - What I'd like to See In Frames 10:03 - Kling's Elements WOW 11:27 - Community Outputs With Elements 12:24 - John Wick Meets Severance Links & Tools Mentioned: GPT Prompt Builder https://chatgpt.com/g/g-678faa261f888191946af0ba95a374af-runway-frames-prompt-builder If you found this video helpful, consider giving it a thumbs-up and subscribing for more AI art, video, and workflow tips! Feel free to share your experiments in the comments, and stay tuned for more updates on these rapidly evolving creative tools.

Vidu 2.0 is a Sleeper HIT for AI Video!

💾

Vidu AI is a Sleeper Hit! Check out HubSpot's Free ChatGPT resources here: https://clickhubspot.com/7olo In this video, I dive into the newly released Vidu 2.0 model, exploring its image-to-video, reference-to-video features, and overall performance. I compare it to Lumia’s Ray 2 text-to-video approach, as well as Sora’s evolving capabilities, highlighting where each model shines (and where they still struggle). Along the way, I run through fun experiments like blending my own portrait into cyberpunk scenes, creating crowd shots, and even conceptualizing an animated “zombie dogs” project. I also check out how the upcoming Kinetix Tech platform may change character motion control—especially if TikTok ever gets banned! Watch till the end for tips, tricks, and my honest thoughts on these emerging AI video tools. If you find these demos helpful or inspiring, please hit that Like button, consider subscribing, and let me know in the comments what you’d like to see in future videos. Thanks for watching! LINKS: Vidu.AI: https://www.vidu.com/create Video on Point Diffusion: https://youtu.be/DVA8XghGmj4 Kinetix Beta Waitlist: https://www.kinetix.tech/sign-up Today's Sponsor: Hubspot! Thank you! https://clickhubspot.com/7olo MINIMAX GPT Prompter (FREE): https://chatgpt.com/g/g-71Fq47Ec6-minimax-cinemotion-prompter Chapters 0:00 Introduction 0:32 Vidu 2.0 Features & Comparing Versions 1:07 UI Walkthrough & Generating Options 1:43 Amplitude, Duration, & Motion Examples 2:22 First Example 3:01 Man in a Blue Business Suit Update 3:54 Combining Outputs (Luma Ray 2 & Vidu) 4:44 Wizard Orb & Prompting Examples 5:26 Vidu's Scene Understanding 5:49 First Frame Last Frame 6:27 Editing for First and Last Frames 6:55- Limitations of First and Last Frames 7:15 Crowd Scenes & Decoherence Fixes 7:51 Anime/Cartoon Style Tests 8:13 ‘Paws of the Dead’—Animated Zombie Dogs 9:07 - Hubspot's ChatGPT Resouces 10:46 Reference-to-Video is the solution 11:45 Improving Results with Rerolls & Blending 12:08 More experiments with References in Vidu 13:01 Rockstar Tim 13:35 Trying to "break" Reference Video 14:11 Sora Remix Experiments 14:46 Pros & Cons of Longer Clips 15:04 The Best Remix yet 15:29 Kinetix Tech for Motion Capture 16:37 - Pro Use? 17:23 - Closing 17:24 Wrap-Up & Reminders

Luma Labs Goes BEAST Mode with a New Video Model!

💾

In this video, I explore Luma Labs’ latest video model, Ray 2. Let's check out what this new era of Dream Machine will bring us. To note, this is a beta release, but it shows huge potential for text-to-video and image-to-video enthusiasts. I’ll also take a dive into the revamped Dream Machine updates and showcase community creations that highlight the full spectrum of what Ray 2 can do. LINK: Luma Labs: https://lumalabsai.partnerlinks.io/dd1jzuzx6o87 00:00 Intro: Ray 2 Unveiled 00:22 Dream Machine Origins and Evolution 00:49 Early Ray 2 Beta Features (1080p clips, 5-second limitations) 01:09 Test Clips: Deserted Island, Tigers, and… Unwanted Lighthouses?! 01:42 Realistic Animations: Walk Cycles and Water Physics 02:39 Aspect Ratio Options: Exploring Western and Pirate Themes 03:14 When Ray2 Hits, it Hits 03:47 The Best AI Godzilla I've Generated Yet 04:34 - Animated Styles With Ray 2 05:19 The New Image Model: Photon 06:27.- Cyberpunk Looks With Photon 07:04 - Community Outputs with Ray 2 08:14 What’s Next for Ray 2 and Dream Machine 🔧 Tech Highlights: • Improved photorealism • New aspect ratios (16:9, 21:9, and more) • Enhanced animation and camera movement controls

New AI Video Models & Big Updates from Hailuo, Runway, Adobe & Krea!

💾

I’m diving into some of the most exciting breakthroughs in AI video generation. First up is a mind-blowing technique called Diffusion as Shader (DAS), which merges 3D tracking with diffusion so you can seamlessly control subjects, locations, and even camera movements. Then, we explore two new game-changers from Adobe—one of them is open-source!—including TransPixar for text-to-video with built-in transparency and FaceLift for single-image 3D head reconstruction using Gaussian splatting. I’ll also give you a rapid rundown of cool new updates from MiniMax, Runway, and Krea, like subject reference tools, 4K upscaling, and built-in audio generation. You’ll see how these tools push boundaries, from full-body animations to real-time 4D facial capture. Chapters 00:00 – Intro 00:30 - A Breakthrough in AI Video (DAS) 00:51 - A Brief Primer on Diffusion and Shaders 01:59 – How DAS Solves 2D Animation Issues 03:08 – Benefits of 3d for AI Video 03:44 - Using DAS with FLUX Depth 04:22 – Video-to-Video Magic & Mesh Animations 05:26 – Adobe’s Open-Source TransPixar for Transparency 07:22 – FaceLift for 3D Heads & 4D Novel View Synthesis 08:44 –MiniMax Subject Reference Now Available and What's Next 09:15 - Gen-3 4K Upscale 09:46 – Krea's Video and Sound Generation

Hailuo's Wild New Video Model & More!

💾

Ready to level up your AI video creation game? In this video, Tim dives into two major updates—Hailou Minimax’s new SVO1 “subject to video” model and Runway Gen3’s brand-new Middle Frame feature—showing you how to achieve more consistent character references and creative transitions in your AI-generated videos. You’ll see hands-on demos in Midjourney, tips for crafting better character sheets, and even a few “stupid” but surprisingly effective compositing tricks in Minimax. Plus, stick around to see how Runway’s latest functionality can help you keyframe animations and spice up your video intros in record time. LINKS & RESOURCES Minimax Prompt Generator (GPT): https://chatgpt.com/g/g-71Fq47Ec6-minimax-cinemotion-prompter THE INTERVIEW Short AI Film: https://x.com/TheoMediaAI/status/1866160155139715248 CHAPTERS: 0:00 - Intro 0:17 - Minimax's New Subject Reference 1:07 - First Tests 1:31 - One Shot vs LORA 1:53 - AI Characters as References 3:01 - Male Character from The Interview 4:02 - Famous Faces 4:20 - Animated Characters 5:07 - Limitations of The Model 5:45 - Using a Character Sheet 6:09 - How To Make A Character Sheet in Midjourney 6:42 - Character Sheet Results 7:12 - Comping for Stylistic Results 8:23 - A Really Stupid Minimax Trick 9:23 - Runway Gen-3's New Middle Frame 10:05 - First Text 10:30 - Usage for a Title Sequence 11:17 - Is a Middle Frame Useful? 11:41 - Closing

Exactly How "Gamechanging" Was AI Video In 2024?

💾

2024 was a WILD Year for AI Video! Check out HubSpot's Free 1000 Prompts here: https://clickhubspot.com/fy4o What a wild ride 2024 has been for creative AI! From groundbreaking launches to unexpected twists, this year was nothing short of incredible. In this video, I take you month by month through the biggest developments in AI technology. Whether it’s MidJourney’s evolution, OpenAI’s Sora, or the game-changing tools like Runway Gen-3 and Stable Diffusion 3, we’ve got it all covered. Don’t miss out on how AI shaped the creative world this year! Thanks to today's sponsor, Hubspot! Download the 1000 Marketing and Productivity Prompts here: https://clickhubspot.com/fy4o Chapters 0:00 – Intro: A Year of Creative AI 0:29 – January: Multi-Motion Brushes with Runway 0:55 – February: Sora’s Iconic Debut 2:24 – March: AI Music and Talking Heads Revolution 3:51 - Hubspot's 1000 Prompts 5:18 – April: Firefly Video and Dream Machine 7:00 – May: ChatGPT Voice Assistant Controversy 7:52 – June: Gen-3 and Dream Machine Go Head-to-Head 9:04 – July: Talking Heads and Quiet Progress 9:30 – August: Black Forest Labs Flux Dominates 10:29 – September: MiniMax Steals the Spotlight 11:10– October: Meta’s MovieGen and Spooky Updates 11:49 – November: Training Models Everywhere 12:07 – December: Sora’s Disappointing Release and the Future 13:39 – Outro: Thanks for an Incredible 2024

Recraft Your AI Images For A Unique Spin

💾

Let's dive into Recraft, the AI image generation platform making waves with its unique model! In this in-depth review, I'll explore everything Recraft has to offer creators and designers. From generating stunning visuals with strong prompt adherence to advanced features like color adjustments, area modification, vector editing, and powerful image sets, I've got you covered! Key features covered: Red Panda (Recraft Version 3) Model Prompt Adherence Diverse Artistic Styles Color Adjustments & Palettes Area Modification (Lasso Tool) Image Variations & Aspect Ratios Vector Editing (SVG Export) Image Sets (Storyboarding) Creating Custom Styles Check out Recraft and use code MEDIA12 for a $12 discount on all plans – https://go.recraft.ai/theoreticallymedia 0:00 Introduction to Recraft 0:33 Who Is Recraft For? 1:08 What was the Red Panda Model? 1:55 Creating Your First Image 2:57 Initial Results & Basic Adjustments 3:45 Advanced Editing: Adding Elements and Modifying Areas 4:28 Creating Image Variations and Different Aspect Ratios 5:21 Exploring Diverse Artistic Styles 6:27 Color Palettes and Style Customization 7:35 Vector Editing and SVG Export 8:30 Palette Options and Storyboarding 8:50 Using Image Sets for Storytelling 9:31 Mini Short Film 9:43 - Image Sets 10:27 Refining Image Sets and Adding Text 11:51 Creating Your Own Style 12:25 Final Thoughts

Veo 2 Hands On Review - The New King Of AI Video?

💾

Google's Veo 2 Is Stunning! Check out HubSpot's Free ChatGPT resources here: https://clickhubspot.com/wwow In this video, I dive into Google’s newly released Veo 2 AI video generation model, just one week after the arrival of another major player in the AI video scene. How does Veo-2 measure up, and is it really the new king of AI video? I put it to the test with a series of prompts—from photorealistic island crash landings to eerie ’80s horror vibes, gritty sci-fi settings, and beyond. I’ll share insights into the UI, show off early outputs, and offer tips on prompting for better results. As this is an early-access look, the model still has quirks and limitations, but the leaps in video realism, character physics, and scene composition are undeniably impressive. LINKS: Google Veo-2 Waitlist: https://labs.google/experiments Google Labs Discord: https://discord.gg/googlelabs Hubspot: https://clickhubspot.com/wwow Chapters: 0:00 - Intro 0:25 - Setting the Stage 1:09 - Veo Interface 1:28 - First Run 2:16 - Multiple Outputs 2:25 - Video Alternates 3:00 - Abstract Prompts 3:18 - Basic Prompts - 80s Horror Movie 3:44 - Sci-Fi Prompts and Movement 4:27 - Character Design 5:03 - Hubspot ChatGPT resources 6:19 - Veo 2 Prompt Formula 7:00 - Prompt Formula Results 7:27 Other Findings 7:49 Parkour Movments 8:37 - Fight Scenes 9:27 - Image to Video 10:07 - Learning Curve 10:45 - What we Normally Get 11:49 Veo-2 Vs Sora 12:23 Features I'd like to see 13:00 Tips to get access

Sora Has a Secret Super Power! Is It Worth $200 Now?

💾

OpenAI's Sora has a super power and today we're going to dive into it! In this in-depth exploration, we test Sora’s remarkable ability to remix, upscale, and transform videos—turning simple footage or AI-generated clips into cinematic, photorealistic experiences. We’ll dive into prompt engineering, explore the Recut feature, and see how Sora pairs with other tools like Runway a Is it worth the $200/month price tag? You decide, as I reveal insider tips, show real examples, and discuss how to strategically maximize your subscription for next-level content creation. My thanks to LTX Studios for sponsoring this video! Please check them out here: https://bit.ly/LTXVtheo CHAPTERS 00:00 – Introduction: Sora’s Hidden Superpower 00:32 – Understanding Sora’s Turbo Model vs. “Big Daddy” Version 01:18 – Initial Disappointments & Public Expectations 02:02 – Discovering the True Superpower: Video-to-Video Transformation 02:25 – Remixing Videos: From Midjourney & Minimax to Sora 03:24- Kling 1.0 to Sora 03:51 - Treadmill Girl Example 04:24 - Remix Strength Controls 05:10 - Using Your Own Footage 05:53 - Trained Models and Face Swapping 06:27 - Is It Perfect? It Is Not 07:14 - But it is Still Better Than Image To Video 07:30 - Now Lets Get Weird 07:57 - Using Recut Now 08:34 - Now We Have A Fantasy Show 09:00 - Shoutout To LTX Studio 11:05 - Let's Bash Runway's Act-1 11:26 - Revisiting Tuesday 11:38 - First Fails 12:00 - Solving Consistency Problems 12:28 - Bringing it to Runway 12:54 - Tuesday The Sora Version 13:16 - The Current State of Sora 13:40 - Is it Worth $200 a Month?

I Paid $200 for Sora. Is It Worth It?

💾

Sora has officially launched, and I’m diving into this dramatic AI video generator with its unique features like blending, storyboarding, and remixing. In this video, I explore the tool’s capabilities, test its outputs, and share whether it’s living up to its hype. Plus, I answer the big question: Is it worth the hefty $200 monthly subscription? Stick around to see how Sora stacks up against the competition! 00:00 - Introduction: The Saga of Sora 00:26 - First Impressions of the Sora Interface 01:08 - Sora Website 01:26 - First Test: Man in a Blue Suit 02:03 - You Aren't Going to Like This 02:34 - Image-to-Video Results 03:17 - Image Referencing? 03:33 - Exploring the Storyboard Feature 04:21 - Testing Blend & Remix Features 05:33 - Video Recut in Sora 06:07 - More Image To Video Results 07:11- But The Pirate Girl Works? 07:39 - MKBHD can get Image to Video? 08:00 - Blend Results 08:32 - Recut Results 08:58 - Remix Results 09:14 - More Image-to-Video Experiments 09:54 - Testing Copyright Content in Sora 10:13 - More Image to Video Tests 10:45 - 80s Action Movie Image Prompt 11:29 - Is Sora All That? 12:07 - Is it Worth $200?
❌