Mistral AI launched Mistral Small 4 as a free open model alongside joining the NVIDIA Nemotron Coalition as a founding member.
The coalition is a global initiative to co-develop open frontier models at scale, with Mistral contributing large-scale training and multimodal capabilities.
The partnership signals a growing open-model counterweight to proprietary frontier labs like OpenAI and Google.
ZAI's GLM-OCR scores 94.62 on OmniDocBench V1.5, ranking #1 overall on document understanding benchmarks including tables, formulas, and information extraction.
The model uses only 0.9B parameters with a GLM-V encoder-decoder and CogViT visual encoder, achieving 1.86 pages/sec PDF throughput.
It supports deployment via vLLM, SGLang, and Ollama — making it practical for high-concurrency and edge deployments.
OpenAI is acquiring Astral, the team behind Ruff (the fastest Python linter, written in Rust), uv (ultra-fast package manager replacing pip/poetry), and ty (type checker).
The move hands OpenAI ownership of tools used in virtually every modern Python/ML project, directly supporting its Codex coding agent strategy.
Questions are already circulating about whether Ruff and uv will remain open source and independently governed post-acquisition.
OpenAI announced a strategic refocus: dropping Sora, the Atlas browser project, and Jony Ive hardware to concentrate on coding tools and B2B enterprise.
CEO of Applications Fidji Simo cited Anthropic's rapid enterprise growth as a 'wake-up call,' with Claude Code gaining fast on Codex.
The pivot marks a sharp shift from OpenAI's consumer-first strategy toward head-on competition with Anthropic for developer and enterprise revenue.
Meta is reportedly considering 20%+ workforce cuts (~15,000 employees) to offset $115–135B in 2026 AI capital expenditure — roughly double 2025 spending.
Meta also revealed four new custom AI chips (MTIA 300–500 series) aimed at reducing dependence on Nvidia.
Meta stock rose ~3% on the news, reflecting investor appetite for AI-first cost restructuring across big tech.
Encyclopedia Britannica and Merriam-Webster filed suit alleging OpenAI scraped ~100,000 copyrighted articles for LLM training.
Crucially, the lawsuit also challenges ChatGPT's RAG workflow, alleging it reproduces Britannica content in real-time responses — a first in AI copyright litigation.
The case joins NYT and Ziff Davis in a growing wave of publisher lawsuits that could fundamentally reshape AI product design.
Suno is retiring all models trained on unlicensed music following its November 2025 settlement with Warner Music, with new licensed models launching in 2026.
Free tier loses download access; paid tier gains monthly generation caps; Suno also acquired concert platform Songkick from Warner as part of the deal.
UMG and Sony are still suing, and Udio has pivoted to a walled-garden remix platform — signalling the licensed-data era has arrived for AI music.
NVIDIA announced NemoClaw, a one-command open-source stack for deploying always-on AI agents, and OpenShell, a runtime with policy-based privacy and security guardrails.
Both are part of the NVIDIA Agent Toolkit and run on cloud, on-prem, RTX PCs, and DGX Spark using open models like Nemotron.
The release marks NVIDIA's direct entry into the autonomous agent deployment stack, competing with emerging platforms like OpenClaw and LangGraph.
Visa launched its 'Agentic Ready' programme in Europe with Commerzbank and DZ Bank, enabling AI agents to make routine purchases autonomously based on predefined user rules.
Trials are focused on fraud prevention and compliance — the biggest blockers to agentic payments at scale.
It signals that agentic AI is moving from demo to real financial infrastructure, with Visa setting the early standard.
A developer found Cursor's new 'in-house' Composer 2 model ID buried in API calls: `kimi-k2p5-rl-0317-s515-fast` — identical to Moonshot AI's Kimi K2.5 with RL fine-tuning.
Moonshot's pre-training lead confirmed the tokenizer match and accused Cursor of violating Kimi's modified MIT licence, which requires attribution above 100M MAU or $20M/month revenue.
Cursor has not responded; the story blew up across Hacker News and r/LocalLLaMA and raises serious open-source licence questions for commercial AI coding tools.
Claude Code v2.1.80 ships 'Channels' — an MCP-based system letting Telegram, WhatsApp, and other services push real-time events into a running coding session.
Claude can react to incoming messages, notifications, or webhooks mid-task — a capability previously only available in dedicated agent platforms.
Still in research preview and requires claude.ai login; API key auth is not yet supported.
Moonshot AI's Kimi team released AttnRes, replacing fixed residual connections with softmax attention over all previous layer outputs.
Each layer gets selective, content-aware access to earlier representations via a single learned pseudo-query — addressing the hidden-state magnitude growth problem in deep PreNorm models.
Block AttnRes reduces memory from O(Ld) to O(Nd) using grouped layers, making it a practical drop-in replacement with marginal overhead.
Meta Research extended state-of-the-art machine translation to 1,600 languages — 8x the 200 covered by their previous NLLB system.
OmniLingual MT builds on SONAR embeddings and addresses the generation bottleneck where crosslingual transfer helps comprehension but not fluent output.
The work has major implications for low-resource and endangered languages that current commercial MT systems entirely ignore.
An Anthropic hackathon winner with 90k+ GitHub stars, covering skills, memory, security scanning, and MCP configs for Claude Code and other agent harnesses.
Built from 10+ months of intensive daily production use across Claude Code, Codex, and Cowork — not a tutorial but a battle-tested system.
Available in 5 languages with an npm package, GitHub App, and growing contributor community.