Report: NVIDIA will launch a "new inference chip" incorporating Groq LPU design at next month's GTC conference

Wallstreetcn
2026.02.28 03:58
portai
I'm PortAI, I can summarize articles.

NVIDIA's upcoming inference chip system integrates Groq's "Language Processing Unit" (LPU) technology, utilizing an architecture that is fundamentally different from traditional GPUs. It is optimized specifically for latency and memory bandwidth bottlenecks in large model inference through broader SRAM integration and 3D stacking technology. This new product may be based on the next-generation Feynman architecture design, significantly reducing the energy consumption and costs of AI agents. OpenAI has committed to purchasing and investing $30 billion