#llm
// 8 transmissions tagged with #llm
Anthropic ships Claude 4.7 with 1M-context
Claude 4.7 lands with a million-token context window and modest pricing changes. Five things shipping engineers should care about.
OpenAI ships GPT-5.5 Instant. Anthropic just overtook them on ARR.
OpenAI announced GPT-5.5 Instant on Monday. The same week, Anthropic's ARR ($30B) eclipsed OpenAI's ($24B) for the first time. The model is the headline; the revenue inversion is the story.
Gemini 3.2 Flash quietly hit the iOS app. Pricing is the news.
Google rolled Gemini 3.2 Flash into the iOS Gemini app and AI Studio with no announcement. $0.25 per million input tokens. Performance reportedly near 3.1 Pro.
Mistral Medium 3.5 lands as a 128B dense model with agentic features
Mistral shipped Medium 3.5 on April 29 — a 128B dense model with new agentic primitives. The Paris lab continues its open-weight cadence as American competitors close their frontier.
DeepSeek V4 ships at 97% below GPT-5.5 — and it runs on Huawei silicon
DeepSeek V4 ships as 1.6T-param Pro and 284B Flash variants under MIT license. Pricing is 97% below OpenAI's GPT-5.5. The unannounced story is that V4 is the first model optimised for Huawei Ascend chips.
Meta's Llama 4 family: 10M-token context, MoE architecture, fully open
Llama 4 ships with two open-weight models: Scout (17B active / 109B total, 10M context) and Maverick (400B parameters). MoE replaces dense transformer. Largest open context window on the market.
Grok 4.20 ships multi-agent, 2M context, weekly updates
xAI released Grok 4.20 in public beta with multi-agent orchestration, a 2M-token context window, and a weekly-update cadence. Hallucination rates reportedly cut to 4.2%.
Mistral Large 3 ships as 41B-active sparse MoE under Apache 2.0
Mistral 3 family launched with three dense small models (3B, 8B, 14B) and Mistral Large 3 — a sparse MoE with 41B active and 675B total parameters. All under Apache 2.0. Large 3 hits #2 in OSS non-reasoning on LMArena.