Intel Delivers Scalable AI Performance In MLPerf Inference V6.0
Intel is demonstrating scalable artificial intelligence performance with results from the newly released MLPerf Inference v6.0 benchmarks, showcasing its Xeon 6 processors and Arc Pro B-Series GPUs for workstations, datacenters, and edge systems. The benchmarks reveal a four GPU system utilizing Intel Arc Pro B70 and B65 graphics delivers 128GB of VRAM, capable of running 120 billion parameter models, with the Arc Pro B70 achieving up to 1.8 times higher inference performance than the Arc Pro B60. Software optimizations within an open, containerized stack also improve performance, yielding up to 1.18 times higher gains on existing Intel Arc Pro B60 hardware compared to MLPerf v5.1. Anil Nanduri, Intel vice president, AI Products and GTM, Intel Data Center Group, said, “The combination of Intel Xeon 6 and Intel’s Arc Pro B-Series GPUs represent our investment to expand customer choice and value, offering real-world solutions that address both LLM models as well as traditional machine learning workloads, with leading performance and incredible value for graphics professionals and AI developers worldwide.” Intel Xeon 6 and Arc Pro B-Series GPU Performance in MLPerf v6.0 These systems are designed to cater to high-end workstations, data centers, and edge computing applications, offering a versatile solution for AI workloads. The architecture of the Intel Arc Pro B70 allows it to handle substantially larger models and context windows in multi-GPU configurations, providing up to...