Meta Says KernelEvolve Improved AI Inference Throughput by 60%
Meta says its KernelEvolve system improved ads-model inference throughput by more than 60%, highlighting AI infra as a product lever.
Meta has released new details on KernelEvolve, an agentic kernel-optimization system used within its Ranking Engineer Agent workflow. Meta says the system improved inference throughput for its Andromeda ads model by more than 60% on NVIDIA GPUs and improved training throughput for another ads model by more than 25% on MTIA.
The broader point is that Meta is treating kernel tuning as a search problem rather than a manual, expert-only task. According to the company, KernelEvolve explores hundreds of candidate implementations, evaluates them with a dedicated job harness, and uses runtime feedback to keep improving performance across different hardware targets.
Why this matters for PMs: AI product performance is increasingly constrained by infrastructure adaptation, not just by model capability. Throughput gains at this layer can materially affect latency, cost, capacity planning, and how quickly new model architectures become viable in production.
For product teams, that is the real signal. The teams that operationalize infra optimization fastest may unlock better AI experiences before competitors with similar model access.
Original source: Meta Engineering announcement: https://engineering.fb.com/2026/04/02/developer-tools/kernelevolve-how-metas-ranking-engineer-agent-optimizes-ai-infrastructure/