#KI-Inferenz

News stories tagged with #KI-Inferenz

Intel Advances AI and Gaming Technology with Partnerships, New Compression Tech, and Quad-GPU System

Intel has announced a strategic partnership with Google to advance AI infrastructure by combining Xeon CPUs with Google’s custom IPUs. The company also introduced Texture Set Neural Compression (TSNC) to reduce memory demands in modern games and compete with Nvidia’s neural texture compression efforts. Additionally, Intel unveiled the Arc Pro B70 in a quad-GPU configuration called Battlematrix, consuming up to 720 watts, though AI inference scaling is non-linear due to PCIe-based communication.

Widely Covered

Nvidia Integrates Groq 3 LPU into Vera-Rubin Platform: A New Era of Low-Latency AI Inference Begins

At GTC 2026, Nvidia announced the integration of Groq’s 3rd-generation Language Processing Unit (LPU) into its new Vera-Rubin-NVL72 platform to dramatically boost AI inference throughput with ultra-low latency. Designed specifically for inference workloads, the LPU leverages high SRAM and internal bandwidth for rapid token processing. This technology complements Nvidia’s existing GPU ecosystem and is deployed in new LPX racks. Partners such as HPE and Giga Computing showcased next-generation AI factories and high-performance computing infrastructure at the event, built around these advancements.