signal_tag · 8_broadcasts

#llm

// 8 transmissions tagged with #llm

TX_002· 16:42AI

Anthropic ships Claude 4.7 with 1M-context

Claude 4.7 lands with a million-token context window and modest pricing changes. Five things shipping engineers should care about.

TX_005· 22:15AI

OpenAI ships GPT-5.5 Instant. Anthropic just overtook them on ARR.

OpenAI announced GPT-5.5 Instant on Monday. The same week, Anthropic's ARR ($30B) eclipsed OpenAI's ($24B) for the first time. The model is the headline; the revenue inversion is the story.

TX_012· 20:00AI

Gemini 3.2 Flash quietly hit the iOS app. Pricing is the news.

Google rolled Gemini 3.2 Flash into the iOS Gemini app and AI Studio with no announcement. $0.25 per million input tokens. Performance reportedly near 3.1 Pro.

TX_014· 16:30AI

Mistral Medium 3.5 lands as a 128B dense model with agentic features

Mistral shipped Medium 3.5 on April 29 — a 128B dense model with new agentic primitives. The Paris lab continues its open-weight cadence as American competitors close their frontier.

TX_013· 18:30AI

DeepSeek V4 ships at 97% below GPT-5.5 — and it runs on Huawei silicon

DeepSeek V4 ships as 1.6T-param Pro and 284B Flash variants under MIT license. Pricing is 97% below OpenAI's GPT-5.5. The unannounced story is that V4 is the first model optimised for Huawei Ascend chips.

TX_011· 15:00AI

Meta's Llama 4 family: 10M-token context, MoE architecture, fully open

Llama 4 ships with two open-weight models: Scout (17B active / 109B total, 10M context) and Maverick (400B parameters). MoE replaces dense transformer. Largest open context window on the market.

TX_016· 22:00AI

Grok 4.20 ships multi-agent, 2M context, weekly updates

xAI released Grok 4.20 in public beta with multi-agent orchestration, a 2M-token context window, and a weekly-update cadence. Hallucination rates reportedly cut to 4.2%.

TX_045· 11:00AI

Mistral Large 3 ships as 41B-active sparse MoE under Apache 2.0

Mistral 3 family launched with three dense small models (3B, 8B, 14B) and Mistral Large 3 — a sparse MoE with 41B active and 675B total parameters. All under Apache 2.0. Large 3 hits #2 in OSS non-reasoning on LMArena.