MLCommons released its latest MLPerf Inference v5.1 benchmarks, showcasing results across 6 key benchmarks for Intel’s GPU Systems featuring Intel® Xeon® with P-cores and Intel® Arc™ Pro B60 graphics, inference workstations code-named Project Battlematrix. In Llama 8B, Intel Arc Pro B60 performance per dollar advantages of up to 1.25x and up to 4x compared to NVIDIA RTX Pro 6000 and L40S respectively.1 The results underscore the performance and accessibility of an all-Intel platform that addresses emerging AI inference workloads across high-end workstations and edge applications.

Why It Matters: Until now, limited options existed for professionals who prioritized platforms capable of delivering high inference performance without compromising data privacy or incurring heavy subscription costs tied to proprietary AI models, but required capabilities to deploy large language models (LLMs).

Momax introduces S.Pass, Revolutionary Baumm Collection, and Artisan Collaborations at IFA 2025

These new Intel GPU Systems, code-named Project Battlematrix, are designed to meet the needs of modern AI inference and provide an all-in-one inference platform combining full-stack validated hardware and software.

Intel GPU systems aim to simplify the adoption and ease of use with a new containerized solution built for Linux environments, optimized to deliver incredible inference performance with multi-GPU scaling and PCIe P2P data transfers, and designed to include enterprise-class reliability and manageability features such as ECC, SRIOV, telemetry and remote firmware updates.

CPUs continue to play a vital role in AI systems. As the orchestration hub, the CPU handles preprocessing, transmission, and overall system coordination. Intel sustained improvements in CPU-based AI performance over the past four years have established Intel Xeon as the preferred CPU for hosting and managing AI workloads in GPU-powered systems.

Intel also remains the only vendor submitting server CPU results to MLPerf, demonstrating leadership and a deep commitment to accelerating AI inference capabilities across both compute and accelerator architectures. Notably, Intel Xeon 6 with P-cores achieved 1.9x performance improvement gen-over-gen in MLPerf Inference v5.1

 
Top