All Issues
Mar 22 - Mar 28, 2026

AI Weekly: Judge Stops Pentagon's Anthropic Blacklist, Claude Mythos Leaked

Models & Releases

4 stories

Mistral Voxtral TTS: Open-Weight Voice Model Beats ElevenLabs

  • Mistral released Voxtral TTS, a 3B open-weight text-to-speech model that outperforms ElevenLabs Flash v2.5 in human preference tests.
  • It runs on approximately 3GB RAM with 90ms time-to-first-audio and supports 9 languages — making it practical for local and edge deployment.
  • The release accelerates the open-source TTS race and directly challenges commercial voice API providers.

Gemini 3.1 Flash Live: Google's Real-Time Voice and Audio Model

  • Google launched Gemini 3.1 Flash Live, its highest-quality audio model yet — with lower latency, better tone understanding, and a 90.8% score on ComplexFuncBench Audio.
  • Available now via the Gemini Live API in AI Studio and through Gemini Live and Search Live in 200+ countries.
  • All audio output is watermarked at generation — a proactive stance against AI-generated misinformation at scale.

Cohere Transcribe: Open-Source State-of-the-Art ASR

  • Cohere released Transcribe, an open-source automatic speech recognition model trained from scratch to minimise word error rate under real-world production conditions.
  • It has a manageable inference footprint for GPU and local deployment, and is also available via Cohere's Model Vault managed platform.
  • The release positions Cohere in the speech intelligence space alongside Whisper and ElevenLabs as audio becomes a core enterprise AI modality.

People & Business

1 stories

Policy & Ethics

2 stories

Federal Judge Blocks Pentagon's Anthropic Blacklisting

  • Federal judge Rita Lin granted Anthropic a preliminary injunction on March 26, blocking a Pentagon directive that had labelled Anthropic a 'supply chain risk' and 'national security threat.'
  • The judge ruled the government likely violated the law and found evidence the blacklisting was retaliation for Anthropic's public AI safety advocacy.
  • The ruling sets a significant precedent for AI companies resisting politically motivated federal overreach on procurement.

Products & Hardware

3 stories

Claude Code AutoDream: Background Agent Tidies Your Memory Files

  • AutoDream is a new Claude Code background sub-agent that automatically consolidates, prunes, and reorganises memory files across sessions.
  • Triggered automatically or manually via the /dream command, it ensures each session starts with a clean, relevant memory slate.
  • Complements AutoMemory and is especially useful for long-running projects where context accumulates over weeks.

Intel Arc Pro B70: 32GB VRAM GPU for $949

  • Intel's Arc Pro B70 ships March 31 with 32GB VRAM at $949 MSRP and 608 GB/s bandwidth — near NVIDIA RTX 5070 territory at a fraction of the cost.
  • The local AI community is excited about the headroom to run 27B+ models at 4-bit quantisation on a single consumer card.
  • If the drivers hold up, the B70 could meaningfully disrupt NVIDIA's near-monopoly on local LLM GPU hardware.

Research & Resources

4 stories

LeWM: Simple Stable JEPA World Model That Trains from Pixels

  • LeWorldModel (LeWM) is a JEPA-based world model that trains end-to-end from raw pixels using just two loss terms and a Gaussian regularizer (SIGReg) to prevent collapse.
  • Eliminates the need for pre-trained encoders, EMA, or complex multi-term losses that make existing JEPA methods brittle.
  • Achieves competitive control and planning performance at a fraction of the compute cost of prior approaches.

Google TurboQuant: Extreme LLM Compression via Vector Quantization

  • Google Research released TurboQuant, a theoretically-grounded set of quantization algorithms for massively compressing LLMs and vector search engines.
  • Targets the KV-cache bottleneck and high-dimensional vector memory that dominate production inference costs.
  • Enables faster similarity lookups and lower memory footprint at scale without sacrificing output quality.

Andrew Ng's Context Hub: Stop Agents Hallucinating APIs

  • Context Hub (chub) is a CLI tool giving coding agents versioned, curated API docs so they stop hallucinating library interfaces mid-task.
  • Agents can annotate documentation gaps locally and vote docs up or down, feeding improvements back to maintainers over time.
  • Works with Claude Code, Codex, and any agent harness via a simple prompt or SKILL.md integration.