
Hi {{firstName|Futurist}},
This is the fiftieth Digital Dips. 50 editions. Crazy! I started a little over two years ago. In that short time, we watched AI move from assistant… to operator. And I’m convinced that in the next two years, it will shift again. From operator to manager. On today’s menu: We now have AI that fakes emotion better than we do. Videos that cost 9 cents to produce. Websites slowly turning into APIs for AI agents. And even bots that hire other bots. If there’s one takeaway from this week, it’s this: we’re building a parallel world. An internet for AI agents. With it’s own economy. Where are ‘we’ in that future?
So grab your favorite snack, settle in, and let's dip into what's cooking. No time to read? Listen to this episode of Digital Dips on Spotify and stay updated while you’re on the move. The link to the podcast is only available to subscribers. If you haven’t subscribe already, I recommend to do so.
🍟 Crispy bites
Fresh tech nuggets. Short, sharp, snackable.
Seedance 2.0 sets a new bar for AI video creation
TL;DR
Seedance 2.0 is now available for beta users on CapCut, marking a significant leap in AI video technology. This model raises the standard for motion, realism, and coherence, transforming what used to be a complex production into something you can achieve with a single prompt. In minutes, it delivers stunning, high-quality video outputs that could easily stand alongside premium content, blurring the lines between amateur and professional filmmaking.
Why this matters
Seedance 2.0 is available for beta users now on CapCut
Achieves high realism and consistency across videos
Compresses the entire video creation process into a single prompt
Supports long-form coherence and multi-media references
CapCut integration streamlines motion design and filmmaking
My Taste
Seedance 2.0 impresses me most in how it handles physics in AI-generated video. I recently watched an AI-rendered football match that looked so real I had to read the comments to realize it was entirely AI. This leap in AI video doesn’t just democratize content creation. Another example I saw online came from the Formula 1 film starring Brad Pitt. The original shot from the film was expensive to produce, but with Seeddance it was recreated for just 9 cents. And instead of an F1 car, the clip showed an SUV with a cat behind the wheel… As more people get their hands on Seedance 2.0, we could see an explosion of high-quality user-generated content that people will genuinely believe is real. Remember that early clip of Will Smith eating spaghetti? Consider that ‘benchmark’ closed.
ElevenLabs' Expressive Mode redefines voice AI communication
TL;DR
ElevenLabs rolls out Expressive Mode, bringing unprecedented emotional nuance to AI conversations. By integrating a conversational model with better turn-taking and advanced text-to-speech, voice agents can de-escalate and guide conversations with empathy in real time. Available in over 70 languages, this evolution narrows the gap between human and AI interactions, enhancing global customer experience.
Why this matters
Expressive Mode enables emotional delivery across 70+ languages.
Integrates a new conversational model for smoother interactions.
Eleven v3 Conversational boosts real-time dialogue capabilities.
Scribe v2 Realtime interprets emotional cues like intonation changes.
My Taste
Emotionally expressive voice agents across 70+ languages represent a significant leap for customer-facing AI. Emotion has always been the missing piece between "technically correct" and "actually helpful." Expressive Mode from ElevenLabs is closing that gap. If AI can truly grasp tone and emotion in real time, we're stepping into a new era of customer experience. The distinction between AI assistants and humans just got even blurrier, potentially making AI better at handling frustrations than human counterparts. We’re officially in the era where AI cares more about your billing issue. Or at least fakes it better than the human customer rep do.
China’s MiniMax M2.5 sets a new standard for AI productivity
TL;DR
MiniMax unveiled M2.5, an open-source AI model geared for real-world productivity. It outperforms heavyweights like Claude Opus and GPT-5.2 on key benchmarks, particularly in tool use and multi-codebase tasks. At just $1 per hour for 100 tasks per second, it's a game-changer for scaling AI efficiently within the enterprise. This makes deploying cost-effective, smarter agents a reality, transforming AI infrastructure into a practical tool for production work.
Why this matters
M2.5 achieves 80.2% on the SWE-Bench, matching Opus 4.6 and surpassing GPT-5.2.
It scores 76.8% on the Berkeley Function Calling Leaderboard, beating Opus by 8.8%.
M2.5 costs $1 per hour at 100 tasks per second, allowing for economic scaling.
It's optimized to execute tasks 37% faster on complex challenges.
Open-sourcing M2.5 democratizes access, making advanced AI development more feasible and accessible to everyone
My Taste
Normally, it was assumed that Chinese open-source models lagged about 4 to 6 months behind American closed-source models. Minimax 2.5? At most two months. That’s unprecedented, and a real shift from how things used to be. No doubt this release raised a few eyebrows in San Francisco. The real story here is MiniMax's ability to compete with giants on critical benchmarks while keeping costs down. This is about making AI scalable and practical for everyday use. The open-source approach combined with real-world productivity focus changes the AI landscape significantly. M2.5 offers a viable path to building expansive AI applications without breaking the bank. No organization has any excuse not to start integrating AI. There simply aren’t any left. With this model, you can continue developing everything in-house, without dependencies on China or the US, and still transform your company.
Gemini 3 Deep Think achieves new heights
TL;DR
In just three months, the Gemini 3 Deep Think upgrade nearly doubled its performance on the challenging ARC-AGI-2 benchmark, scoring 84.6%. This leap was accompanied by an 82% reduction in cost per task, now down to $13.62. Google's fine-tuning of inference-time optimization is reshaping the landscape of AI reasoning, setting a pace others are struggling to match.
Why this matters
Gemini 3 Deep Think scored 84.6% on ARC-AGI-2, up from 45.1%.
Reduced task cost from $77.16 to $13.62.
Codeforces Elo rating hit 3455, placing it among top global programmers.
Surpassed gold medal-level results in the 2025 International Olympiads for Physics and Chemistry.
Enabled to model physical systems and generate 3D-printable files.
My Taste
Let's talk about the real story here: efficiency trumps brute force. Google's focus on inference optimization is a game-changer that shifts the reasoning race. No longer is AI progress about throwing more compute power at the problem; it's about smarter, leaner algorithms. With Deep Think now mainstreaming capabilities once thought exclusive, the gap between AI leaders and followers is widening rapidly. Expect the benchmarks to shift again, Google just redefined the playing field.
Websites become APIs for AI with WebMCP
TL;DR
WebMCP is in early preview, offering a standardized way for AI agents to interact with websites without seeing the UI. Sites register structured tools directly in the browser, allowing agents to perform tasks quickly and accurately. This development is set to revolutionize how AI agents navigate websites, improving reliability and efficiency by replacing traditional methods with structured APIs. As more sites adopt this, we might see a shift in web design priorities.
Why this matters
Early benchmarks show a 67% reduction in computational overhead.
Task accuracy reaches approximately 98% with WebMCP.
AI agents no longer guess which buttons to click, boosting efficiency.
Sites that adopt WebMCP offer a streamlined agent interaction path.
Potential for new discipline: "Agent Experience Optimization."
My Take
Google and Microsoft have turned every website into a potential API for AI agents with WebMCP, creating second-order effects that are hard to ignore. Where agents previously stumbled through guessing interfaces, they now have a clear roadmap for task completion. This has massive implications for site competition and optimization. In the near future, those who don't optimize for AI agents the way they once did for search engines may find themselves at a disadvantage. The landscape of the web is about to change, and it's ushered in by bots that already dominate 51% of web traffic.
🧀 Cheesy pick
A cheesy selection of three tools and one tasty rabbit hole.
Dokie turns your rough notes into polished slide decks instantly.
Nebula turns chats into working agents and workflows.
Orchids let you build any app, slack bot, AI agent, anything
Bonus: this report from Anthropic explores agentic coding in 2026.
🍱 Leftovers
A roundup of updates that are too cheesy to ignore.
Hedra Labs introduces Omnia Alpha. Audio-driven generation with full control over camera, motion, and background.
Hedra Labs’ Avatars crafts seamless, custom talking heads with perfect lip-sync and natural motion.
Firecrawl’s Branding Format v2 sharpens your brand identity extraction, now acing Wix and Framer sites.
Waymo unveils the Waymo World Model to simulate extreme scenarios for autonomous vehicles using DeepMind’s Genie 3.
OpenAI’s Sora's new feature turns photos of people into videos with enhanced moderation.
OpenAI tests ads for ChatGPT free and Go users in the U.S., keeping them visually distinct from its responses.
OpenAI elevates its Responses API with tools for multi-hour agent runs and internet-savvy containers.
OpenAI debuts GPT-5.3-Codex-Spark for lightning-fast coding in ChatGPT Pro's Codex ecosystem.
OpenAI’s ChatGPT now supercharges deep research with the newly released GPT-5.2 engine.
Anthropic’s Claude releases its 2.5x-faster Opus 4.6 version for early experimenters via Claude Code and API.
Anthropic’s Cowork leaps onto Windows with full MacOS feature parity, including file access and multi-step task execution.
Anthropic’s Claude's most-loved features, file creation, connectors, and skills, now free for all users.
Metain brings AI-driven game creation to Roblox Studio, transforming your ideas into playable experiences.
Seedream 5.0 goes live on CapCut, boasting advanced image abilities and global availability.
Google’s YouTube Premium users can now craft instant playlists with AI-powered music curation.
Google’s Veo lands in Google Ads, empowering advertisers to craft studio-quality campaigns in minutes.
Google tests new AI Mode ad format to integrate sponsored shopping results into searches.
Vibecode lets you build Agentic Apps with ease using Claude Code integration.
Alibaba’s Qwen-Image-2.0 turns your paragraphs into pro slides and brings scenes to life with flawless 2K photorealism.
Krea introduces prompt-to-workflow. It transforms text instructions into complete node workflows.
Essential Apps lets you describe your dream app, and AI builds it for you, beta now on Nothing Playground.
Warp’s Oz lets you orchestrate hundreds of agents in the cloud, streamlining your code from terminal to pull request.
Lindy AI’s iMessage AI Assistant launches to streamline your day by integrating with all your apps.
Tavus’ Raven-1 empowers AI to grasp human emotion, intent, and context through audio and video.
Supermemory app launches, letting you own your context and connect seamlessly to tools like Claude, OpenClaw, and Notion.
Stripe introduces machine payments, letting developers charge autonomous agents with just a few lines of code.
PrimeIntellect Lab launches a full-stack platform to train custom agentic models without infrastructure hassles.
Zai unveils GLM-5, powers through 700+ tool calls and 800+ context handoffs, running solo for 24 hours.
Zai’s GLM-5 debuts Agent Mode, automating task execution and delivering ready-to-use files.
Willow for Developers debuts voice dictation for seamless vibe coding without a keyboard.
Autonomous Labs lands on your desk as the first personal AI assistant, ready to tackle tasks via text.
Exa Instant debuts as the first sub-200ms search engine, fueling real-time AI chat and voice.
Higgsfield AI Cinema Studio 2.0 rolls out advanced AI tools for crafting 3D scenes with true emotional depth and total creative control.
Cloudflare enhances agent interactions with real-time Markdown content conversion.
Brave unveils the most powerful Search API yet, outshining ChatGPT and Google AI Mode for AI applications.
AGI Inc unveils the first on-device AI assistant for seamless phone management without lifting a finger.
Meta’s Dear Algo on Threads lets you customize your feed with AI for tailored content control.
Cline CLI 2.0 lets you code with AI agents directly in your terminal, revolutionizing your CI/CD workflows.
How’d this digital dip taste?
This was it. Our fiftieth digital dip together. Forward this to someone who thinks AI is just a smarter search bar.
Maybe you’re reading this and thinking: my website needs to work for AI agents. Not just for humans. But where do you even start? What does that mean for your content, your tools, your data, your team? If you feel the shift but don’t yet have the plan: reply to this email. I’ll help you think it through.
Looking forward to what tomorrow brings! ▽
-Wesley


