Inside NVIDIA Groq 3 LPX: The Low-Latency Inference Accelerator for the NVIDIA Vera Rubin Platform
By Jakub Antkiewicz
•2026-03-20T08:42:15Z
NVIDIA has introduced the Groq 3 LPX, a rack-scale inference accelerator designed to work alongside its Vera Rubin platform. The system is engineered to address the growing demand for low-latency performance required by interactive and agentic AI systems. Rather than replacing the general-purpose Vera Rubin architecture, the LPX serves as a specialized component within the data center, handling the most time-sensitive parts of AI inference to enable more responsive, real-time user experiences like multi-agent collaboration and high-speed coding assistants.
The LPX architecture is built around 256 interconnected NVIDIA Groq 3 LPU accelerators, which prioritize deterministic, compiler-orchestrated execution over raw arithmetic throughput. This design relies on a flat, SRAM-first memory system (totaling 40 PB/s of on-chip bandwidth) and explicit data movement to minimize the unpredictable delays and jitter common in hardware-managed cache systems. In this heterogeneous model, NVIDIA's Dynamo software routes latency-sensitive decode operations, such as FFN and MoE expert execution, to the LPX, while the more flexible Vera Rubin NVL72 GPUs handle high-throughput tasks like prefill and decode attention. This division of labor aims to optimize both overall data center throughput and the per-token latency for interactive services.
The introduction of the LPX signals a strategic shift toward more specialized, heterogeneous compute architectures within the AI data center. By providing a dedicated path for low-latency inference, NVIDIA is enabling operators to support both cost-effective, high-throughput background AI processes and premium, interactive services within a common infrastructure. This approach directly addresses a key operational challenge: delivering stable, predictable performance for next-generation AI applications without compromising the efficiency of large-scale, general-purpose AI workloads. The move suggests the market is maturing beyond a one-size-fits-all hardware model toward workload-specific optimizations.
NVIDIA's integration of the Groq 3 LPX into its Vera Rubin platform is a pragmatic acknowledgment that the monolithic GPU-centric data center is evolving. By creating a specialized, co-designed accelerator for low-latency inference, NVIDIA is building a more defensible, heterogeneous ecosystem that can absorb niche innovations to address specific bottlenecks in the AI workflow, particularly the growing demand for real-time agentic systems.