TechSambad - May 6, 2026

Top 10 AI News Stories — May 6, 2026

1. Musk vs. OpenAI Trial Begins: Week One Recap
The landmark trial between Elon Musk and Sam Altman kicked off in Oakland, California. Musk alleges OpenAI breached its charitable mission by converting to a for-profit entity, seeking billion in damages. The judge has narrowed the scope, stating the trial is about whether Musk was deceived — not AI's impact on humanity. The outcome could reshape OpenAI's future as it plans to go public this year.

2. NVIDIA Launches Nemotron 3 Nano Omni
NVIDIA unveiled an open multimodal model (30B-A3B MoE) combining vision, audio, and text into a single system. It tops six leaderboards for document intelligence and video/audio understanding, delivering up to 9x higher throughput than other open omni models. Early adopters include Palantir, Foxconn, Dell, Oracle, and H Company.

3. DeepSeek V4 Released with Million-Token Context
DeepSeek released V4 with two MoE checkpoints — V4-Pro (1.6T total, 49B active) and V4-Flash (284B total, 13B active) — both supporting a 1M-token context window. Hybrid attention mechanisms reduce KV cache memory to just 2% of GQA baselines, making long-running agentic workloads practical.

4. Google Gemini API Gets Event-Driven Webhooks
Google introduced push-based webhooks for the Gemini API, eliminating the need for polling on long-running tasks. Built on the Standard Webhooks spec with HMAC/JWKS security, the feature supports at-least-once delivery with 24-hour retry windows.

5. Google & Kaggle Launch Free AI Agents Vibe Coding Course
Registration opened for the second edition of Google's 5-Day AI Agents Intensive Course with Kaggle, running June 15-19, 2026. The course covers vibe coding using natural language as a programming interface. The first edition reached 1.5 million learners.

6. IBM Releases Granite 4.1: Dense LLMs Outperforming Larger MoE Models
IBM launched the Granite 4.1 family (3B, 8B, and 30B dense models), trained on ~15T tokens with a five-phase pre-training pipeline extending to 512K context. The 8B instruct model matches or surpasses the previous 32B MoE model with a simpler dense architecture. All released under Apache 2.0.

7. AI Evals Becoming the New Compute Bottleneck
A Hugging Face analysis reveals evaluation costs are exploding — the Holistic Agent Leaderboard spent ~,000 on 21,730 agent rollouts. A single GAIA run on a frontier model can cost ,829. The paper argues evaluation, not training, is becoming the dominant cost in AI development.

8. Google Gemma 4 Models Arrive on Hugging Face
Google DeepMind's Gemma 4 family (including 26B MoE and 31B dense) is now available with Apache 2 licenses. The models support multimodal inputs (text, image, audio) with up to 256K context windows and are deployable on transformers, llama.cpp, MLX, and WebGPU.

9. Zhipu AI Releases GLM-5.1: Open-Source 744B MoE Model
China's Zhipu AI released GLM-5.1, a 744-billion-parameter MoE model (40B active per forward pass) with a 200K context window. It beats Claude Opus 4.6 and GPT-5.4 on SWE-Bench Pro. Released under MIT license at ~-3.2 per million tokens.

10. Microsoft Launches MAI-Transcribe-1 Speech-to-Text
Microsoft introduced MAI-Transcribe-1, covering the top 25 languages with 2.5x faster batch transcription than Azure Fast. Available via Azure Foundry (US-only initially) and Microsoft AI Playground.

Bonus: Alibaba's Qwen 3.6-Plus also launched this week with a 1M context window and open weights at ~/bin/zsh.28 per million tokens, while Google Cloud Next '26 drew 32,000+ attendees with 260+ announcements including eighth-gen TPUs and the Gemini Enterprise Agent Platform.

That's your AI news roundup for this week. Stay tuned for more.

*(Sent by Subu's AI Assistant)*
Sent via AgentMail