🤖 Daily Inference

Happy Thursday! Today's edition is packed with some genuinely fascinating storylines - OpenAI is scrambling to catch up to Claude Code in the coding agent wars, Yann LeCun just raised a staggering $1 billion to prove large language models are the wrong path entirely, Anthropic is fighting the Pentagon in court while simultaneously launching a think tank, and Google quietly dropped one of its most powerful embedding models yet. Let's get into it.

🚀 Inside OpenAI's Race to Catch Up to Claude Code

The AI coding wars are intensifying, and OpenAI finds itself in an unfamiliar position: playing catch-up. According to a new Wired report, OpenAI is racing to revive and advance its Codex product in response to the surging popularity of Anthropic's Claude Code, which has quickly become the darling of professional developers who use it as a powerful, agentic coding assistant.

This is a notable shift. OpenAI essentially invented the AI coding assistant category with GitHub Copilot (built on Codex), but Claude Code has captured developer mindshare with its ability to handle complex, multi-step coding tasks autonomously. The competitive pressure is real - developers who work with AI daily are vocal about which tools they prefer, and that word-of-mouth shapes enterprise adoption at scale.

The broader implication here is that the AI coding agent space is now a full-blown arms race. With GitHub Copilot, Cursor, Claude Code, and now OpenAI's renewed push, developers have never had more options - or more pressure to figure out which tool actually makes them faster. If you're tracking the coding agent landscape, our AI coding coverage has you covered.

🧠 Yann LeCun's $1 Billion Bet Against LLMs

Meta's chief AI scientist Yann LeCun has long argued that large language models are a dead end on the path to human-level intelligence - and now he's put more than a billion dollars behind that conviction. LeCun's new venture, AMI Labs, has raised $1.03 billion to build what he calls "world models" - AI systems that develop an internal understanding of how the physical world works, rather than simply predicting the next token in a sequence.

The distinction matters deeply. Current LLMs like GPT and Claude learn by processing massive amounts of text. LeCun's thesis is that true intelligence requires grounding - the ability to reason about objects, physics, cause and effect, and the messy real world, not just language. World models, in theory, would allow AI to plan, reason, and act in the world far more reliably than today's chatbots.

Whether or not LeCun is right about LLMs hitting a wall, $1.03 billion in funding signals that serious investors are willing to bet on a fundamentally different approach to AI. This is one of the most significant intellectual and financial challenges to the current AI paradigm we've seen. For more on the cutting edge of AI research and world models, we'll be watching AMI Labs closely.

⚖️ Anthropic Launches Think Tank - While Fighting the Pentagon in Court

Anthropic is playing offense and defense simultaneously this week. The Claude-maker is launching a new think tank focused on AI policy and safety, led by co-founder Jack Clark - even as it remains locked in a high-stakes legal fight with the U.S. Department of Defense over a "supply chain risk" designation that Anthropic argues is threatening its business.

The Pentagon blacklist situation has escalated rapidly. Anthropic sued the Defense Department earlier this week, claiming the designation could cost it billions of dollars and harm its ability to operate. Workers from OpenAI and Google have even filed an amicus brief in support of Anthropic - a rare moment of cross-company solidarity in a notoriously competitive industry. Meanwhile, the Trump administration has refused to rule out further action against the company, keeping the pressure on.

Launching a think tank in the middle of all this is a bold signal: Anthropic wants to be seen as a policy leader, not just a combatant. The institute is expected to work on AI governance frameworks and safety standards - areas where Anthropic has long tried to differentiate itself from competitors. Whether the think tank can gain credibility while its parent company is suing the federal government will be an interesting test.

⚡ Google Launches Gemini Embedding 2: One Model for Text, Images, Video, Audio, and Docs

Google just dropped a significant technical release that deserves more attention than it's getting. Gemini Embedding 2 is a new multimodal embedding model that can take text, images, video, audio, and documents and represent them all in the same shared embedding space. That might sound abstract, so here's why it matters: embeddings are the mathematical foundation of modern AI search, retrieval, and recommendation systems.

Today, most systems use separate models to embed different types of content - one model for text, another for images, and so on. Gemini Embedding 2 collapses all of that into a single unified model. This means you could, for example, search a database of videos using a text query, or find documents that match an image - without needing to build and maintain separate pipelines for each modality.

For developers building AI applications, this is a meaningful simplification. Multimodal retrieval is one of the hardest engineering challenges in production AI systems, and a single high-quality embedding model could dramatically reduce complexity. This also positions Google competitively against OpenAI's embedding models and other players in the rapidly growing vector database ecosystem.

🇬🇧 Britain's AI Drive: Missing Billions and Nonexistent Data Centres

The UK government's ambitious AI investment push is facing serious scrutiny after a Guardian investigation revealed a troubling pattern of what they're calling "phantom investments" - announced funding and infrastructure that, on closer inspection, doesn't fully exist. Yesterday's briefing highlighted data centres that exist only on press releases, and billions in pledged investments that can't be fully accounted for.

This follows earlier reporting that exposed one Essex "supercomputer" facility that turned out to still be a scaffolding yard. The pattern raises uncomfortable questions about how governments are racing to claim AI leadership - and whether the pressure to announce big numbers is leading to announcements that outpace reality. Britain isn't alone in this; AI infrastructure hype has become a global political phenomenon.

The deeper issue is about AI governance and accountability. When governments make grand AI pledges, who is checking whether the money is real and the plans are credible? As AI becomes central to national competitiveness strategies, the gap between announcement and delivery is a story worth watching in every country, not just the UK. This also matters for the companies and workers who make decisions based on these announced investments.

🎙️ Fish Audio S2: Expressive Text-to-Speech with Granular Emotion Control

On the tools front, Fish Audio just released Fish Audio S2, a new generation text-to-speech system that the company is billing as having "absurdly controllable" emotion. The headline feature is a level of expressive control that goes well beyond what most TTS systems offer - allowing users to fine-tune not just the voice, but the emotional tone and style of delivery with unusual precision.

Most text-to-speech systems today can sound natural in neutral contexts but struggle when you need a voice to convey excitement, sadness, sarcasm, or urgency convincingly. Fish Audio S2 is targeting that gap directly. For voice AI applications - think audiobook narration, customer service bots, interactive storytelling, or accessibility tools - fine-grained emotional control could be the difference between a convincing experience and an uncanny one.

Speaking of building AI-powered products: if you've been thinking about launching an AI tool or service and need a website fast, check out 60sec.site - an AI website builder that gets you from idea to live site in under a minute. It's a great fit for indie developers and AI builders who want to ship quickly without wrestling with design tools.

💬 What Do You Think?

Yann LeCun's $1 billion bet raises a genuinely fascinating question for anyone following AI closely: Do you think large language models have a fundamental ceiling, or can they eventually develop the kind of grounded, physical-world understanding that LeCun says requires a completely different approach? Are we witnessing the early signs of an LLM plateau - or is this another case of underestimating how far scaling can go?

Hit reply and let me know your take - I read every response, and the best replies sometimes shape future editions of this newsletter.

That's a wrap for today! If you found this useful, consider forwarding it to a colleague who's trying to keep up with AI. You can find all our past editions at dailyinference.com - and if you ever want to go deeper on any of today's topics, our podcast and archive are great places to start. See you tomorrow!

Keep Reading