Just your average tech head teaching you how to use AI and your camera, specialising in Creator tools, Tech and Editing with 1M+ followers
Share
This Week in AI: Veo 3 Ad Prompts, Local Video Models, Live Avatars, and IDE Agents
Published 8 months ago • 3 min read
Your weekly source of AI and tech to help you elevate your creator journey.
Hi Reader,
This week, we saw breakthrough models that think in long sequences, talk in real time, and even decode ancient history.
Let’s get into the updates that are defining what’s next.
In today’s newsletter:
Runway’s Aleph transforms context modeling for video and beyond
Wan 2.2 launches open-source, local text-to-video for free
Hedra releases real-time avatars with sub-100ms latency
DeepMind unveils Aeneas to decode ancient Latin in seconds
GitHub drops Spark, a coding agent inside your IDE
Let’s dive in.
Runway Aleph: The Future of Context-Aware AI Models
Runway just introduced Aleph, its new foundation model designed to understand and generate sequences across video, text, and more. Unlike traditional models that rely on limited windows or chunks, Aleph is context-native, processing entire timelines in one coherent go.
Key details:
Context Native Model Instead of a sliding window or chunked attention, Aleph understands entire sequences from start to finish—ideal for storytelling, editing, and scene continuity.
Built for Video + Beyond Originally optimized for video timelines, but extensible to audio, multimodal tasks, and longform narratives.
Foundation of Runway Gen-3 Powers the next generation of Runway’s Gen-3 video tools used in Hollywood-grade productions.
Temporal + Spatial Awareness Captures movement, emotion, and visual evolution with higher fidelity than ever.
Hedra Live Avatars: Real-Time AI Characters with Sub-100ms Latency
Hedra has launched Live Avatars, its new streaming model that creates hyper-realistic AI characters who respond instantly, ideal for customer service, live streams, and virtual presenters.
Key details:
Ultra-Low Latency Response times under 100ms make conversations feel truly live, not pre-rendered.
Lifelike Expressions + Movement Blinking, talking, smiling, all in sync with audio and intent.
Use in Video Calls, Twitch, Virtual Hosts Just feed text or voice, and the avatar animates in real time across platforms.
Built for Scalability Stream hundreds of avatars at once for global deployments in customer service or media.
DeepMind Aeneas: AI for Decoding Ancient Inscriptions
Google DeepMind just introduced Aeneas, a specialized AI designed to interpret fragmented Latin inscriptions by drawing context from a vast database of ancient texts.
Key details:
Historical Contextualization Matches partial fragments to thousands of known inscriptions to reconstruct missing information.
Trained on Epigraphic Datasets Built in collaboration with historians and archaeologists across Europe.
First of Its Kind in Humanities AI Opens up new possibilities for ancient language recovery and historical understanding.
Fast + Accurate Seconds to decode what once took experts days or weeks.
GitHub Spark: AI Agent That Codes Directly in Your IDE
GitHub has introduced Spark, a new autonomous coding agent that lives directly in your development environment and handles complex programming tasks without manual prompting.
Key details:
Multi-Step Code Execution Spark doesn’t just autocomplete, it plans, edits, and executes multi-line logic flows.
Built into VS Code and GitHub Copilot Seamlessly integrates with your existing dev tools.
Understands Context Across Files Reads, modifies, and manages multiple files at once, perfect for large codebases.
Ideal for Prototyping and Refactoring Can optimize, refactor, or build out entire functions with minimal input.
Your weekly source of AI and tech to help you elevate your creator journey. Hi Reader, This week, we’re not just generating content anymore… we’re directing it, simulating it, and editing reality itself in real time. What used to take full production teams, VFX studios, and expensive hardware can now happen inside a single interface. In today’s newsletter: Beeble VFX No green screen, pure magic . Higgsfield Cinema Studio 2.5 turns your screen into a full AI film set Freepik pushes video...
Your weekly source of AI and tech to help you elevate your creator journey. Hi Reader, The AI updates keep coming fast, and this week’s releases push creation even closer to professional-grade workflows. We’ve got a new talking-video model that acts with emotion and context, an open-source TTS engine simulating natural group conversations, a major leap in video control with start-to-end framing, and Google finally confirming “Nano Banana” as its most advanced image editor yet. In today’s...
Your weekly source of AI and tech to help you elevate your creator journey. Hi Reader, The AI frontier just leaped again and this week’s launches prove we’re right at the edge of imagination meeting execution. In today’s newsletter: Nano Banana breaks the internet and edits with uncanny precision. Runway Game Worlds evolves AI storytelling. ElevenLabs Video‑to‑Music in Studio scores your visuals. Act‑Two Voices empowers expressive AI performance. Kling 2.1 redefines image‑to‑video realism....