๐ค AI & LLMs Hugging Face
Llama 4 Scout outperforms GPT-4o on multi-step reasoning
Meta's compact Llama 4 variant hits SOTA on reasoning tasks using only 17B active parameters via MoE with 16 experts.
Open-source model at this quality means local inference without API costs โ a real alternative for production.
๐ค AI & LLMs OpenAI Blog
OpenAI o3 mini API goes GA with structured output support
The reasoning model now supports JSON schema enforcement, making it practical for production pipelines that need deterministic output shapes.
Reasoning + structured output finally makes LLMs reliable enough to drive business logic without a validation layer.
๐ค AI & LLMs The Gradient
Mixture-of-Experts is now the default LLM architecture
Analysis of the top 20 new models released in Q1 2025 shows 17 use MoE. Dense transformers are becoming legacy.
Understanding MoE routing is now table stakes for anyone deploying or fine-tuning foundation models.