Intel leaders focus on the company's Data Center and Artificial Intelligence business unit with roadmap and progress updates.
All eyes - especially investors' eyes - are on Intel's data center business today.
Intel's
Executives dished out updates on Intel's data center business for investors. This included disclosures about future generations of Intel Xeon chips, progress updates on 4th Gen Intel Xeon Scalable processors (code-named Sapphire Rapids) and demos of Intel hardware tackling the competition, heavy AI workloads and more.
More: Taking on the Compute and Sustainability Challenges of Generative AI
If you don't have time to watch the replay, here's the top four takeaways:
Xeon Roadmap Roll Call
Among Sapphire Rapids, Emerald Rapids,
Today: A 4th Gen Intel Xeon Scalable update
All major original equipment manufacturers (OEMs) and original design manufacturers (ODMs) are shipping 4th Gen Xeon systems, and the top 10 global cloud service providers are working on deploying services on 4th Gen Xeon. It's the highest quality data center CPU Intel has ever delivered, and the ramp is continuing aggressively.
Q4 2023: 5th Gen Intel Xeon Scalable (code-named Emerald Rapids)
Emerald Rapids is Intel's next Performance-core (P-core) product. Starting today, it is officially known as 5th Generation Intel Xeon Scalable. As it sheds its code name and dons formal branding in preparation for its launch, the CPU is already sampling to customers. Volume validation is underway and
First-half 2024: Intel Xeon Processor code-named
Intel Xeon Processor code-named
A Fast Follow: Intel Xeon Processor code-named Granite Rapids
Intel Xeon Processor code-named Granite Rapids will arrive hot on the heels of
'We are building the fastest memory interface in the world for Granite Rapids,' Spelman said as she fired up a command prompt to show off its memory configuration. 'Intel invented and led the ecosystem in developing a new type of DIMM called Multiplexer Combined Rank (MCR) that lets us achieve speeds of 8,800 mega transfers per second, based on DDR5.'
The MCR DIMM innovation achieves an 80% peak bandwidth increase over current-gen server memory technology, and Lisa was able to demonstrate Granite Rapids' stability while saturating a healthy memory subsystem with read/writes.
Further into the Future
Speaking to it publicly for the first time, Intel will continue to execute on its E-core roadmap with the follow-on to
Bonus Bits
The Habana Gaudi3 AI accelerator has taped in. Performance details haven't been shared yet, but Habana Gaudi2 is in the environment and offering 1.8x advantage in throughput-per-watt over a comparable A100 server when running a popular computer vision workload.1
There are 15 new FPGA products scheduled to go through the production release qualification (PRQ) process this year - that's more new product introductions than ever before in Intel's FPGA business.
Intel Investor Webinar: Data Center and AI (Replay)
On
Looking Beyond the CPU for
This packed roadmap is on track to deliver against an even bigger total available market (TAM) than previously considered. The
'When we talk about compute demand, we often look at the TAM through the lens of CPU units,' Rivera explained. 'However, counting sockets does not fully reflect how silicon innovations deliver value to the market. Today, innovations are delivered in several ways, including increased CPU core density, the use of accelerators built into the silicon and the use of discrete accelerators.'
Based on the above, Rivera cited a couple of specific reasons for the increased TAM: With the integration of accelerator compute and advanced GPU offerings into its data center business, Intel is better positioned to serve a wider swath of customers. What's more, Xeon's performance in high-profile workloads (like AI, analytics, security, networking and HPC) is driving demand for mainstream compute as well as discrete accelerators tailored for the task at hand.
AI is Everywhere, It's for Everyone and It's Running on Intel
Silicon updates aside, AI was a persistent theme in today's webinar.
Intel is already foundational in AI hardware, from data preparation and management to small- and medium-scale training and inference on Xeon, and increasingly large model training and inference using Intel GPUs and accelerators.
Today, Rivera defined Intel's commitment to the true democratization of AI from the cloud to the network and to the edge by enabling broader access to solutions and more cost-effective deployments through an open ecosystem approach.
'Customers want portability in their AI workloads. They want to build once and deploy anywhere,' Rivera said. 'As we continue to deliver heterogenous architectures for AI workloads, deploying them at scale will require software that makes it easy for developers to program and a vibrant open and secure ecosystem to flourish.'
'The desire for an open, multivendor multiarchitecture alternative to (Nvidia's) CUDA is not waning,' Lavender said. 'We believe that the industry will benefit from a standardized programming language that everyone can contribute to, collaborate on, is not locked into a particular vendor and can evolve organically based on its members and public requirements.'
Intel has made contributions to SYCL, an open C++-based programming model, and acquired
Powering Progress in Generative AI
Intel's open philosophy extends to the field of generative AI, where powerful tools like ChatGPT and text-to-image model DALL*E 2 are leveraging AI algorithms and large data sets to produce new content based on increasingly complex human prompts.
Generative AI is evolving at warp speed (it feels like there are near-daily headlines about breakthroughs), and Intel is actively working to support an open generative AI ecosystem while also addressing increasing performance demands.
Earlier this week, Hugging Face, a major player in machine learning application development, announced that it had enabled the 176 billion parameter BLOOMZ model on Habana's Gaudi 2. The BLOOM model is an open source large language AI model analogous to the 175B parameter GPT-3 model employed by ChatGPT. The company also confirmed it is running deep-learning text-to-image model Stable Diffusion on 4th Gen Xeon with built-in Intel AMX for AI workload acceleration (Stable Diffusion is an open access alternative to DALL*E).
Beyond Intel's efforts to bolster an open and ethical ecosystem, Xeon is present and accounted for elsewhere in generative AI landscape, too. Rivera reminded investors today that Nvidia is using 4th Gen Xeon as the head-node to run alongside its H100 GPUs to power the virtual machines accelerating generative AI models in Microsoft Azure, including ChatGPT.
1 Supermicro L12 Validation Report of Gaudi2 HL-225H SYS-820GH-THR2,
(C) 2023 Electronic News Publishing, source