Oracle and NVIDIA announced an expanded collaboration to deliver sovereign AI solutions to customers around the world. Oracle's distributed cloud, AI infrastructure, and generative AI services, combined with NVIDIA's accelerated computing and generative AI software, are enabling governments and enterprises to deploy AI factories. These AI factories can run cloud services locally, and within a country's or organization's secure premises with a range of operational controls, supporting sovereign goals of diversifying and boosting economic growth.

The combination of NVIDIA's full-stack AI platform with Oracle's Enterprise AI ? deployable across OCI Dedicated Region, Oracle Alloy, Oracle EU Sovereign Cloud, and Oracle Government Cloud ? offers customers a AI solution that provides greater control over operations, location, and security to help support digital sovereignty.

The companies' combined offerings can be deployed via the public cloud or in a customer's data center in specific locations, with flexible operational controls. Oracle is the only hyperscaler capable of delivering AI and full cloud services locally, anywhere. OCI services and pricing are consistent across deployment types to simplify planning, portability, and management.

Oracle's cloud services leverage a range of NVIDIA's stack, including NVIDIA accelerated computing infrastructure and the NVIDIA AI Enterprise software platform, including newly announced NVIDIA NIM? inference microservices, which are built on the foundation of NVIDIA inference software such as NVIDIA TensorRT?, NVIDIA TensorRT-LLM, and NVIDIA Triton Inference Server?. To help customers address the ever-increasing needs of AI models, Oracle plans to take advantage of the latest NVIDIA Grace Blackwell computing platform, announced today at GTC, across OCI Supercluster and OCI Compute.

OCI Supercluster will become significantly faster with new OCI Compute bare metal instances, ultra-low-latency RDMA networking, and high-performance storage. OCI Compute will adopt both the NVIDIA GB200 Grace Blackwell Superchip and the NVIDIA Blackwell B200 Tensor Core GPU. The NVIDIA GB200 Grace?

Blackwell Superchip will power a new era of computing. GB200 delivers up to 30X faster real-time large language model (LLM) inference, 25X lower TCO, and requires 25X less energy compared to the previous generation of GPUs, supercharging AI training, data processing, and engineering design and simulation. NVIDIA Blackwell B200 Tensor Core GPUs are designed for the most demanding AI, data analytics, and high-performance computing (HPC) workloads.

NVIDIA NIM and CUDA-X? microservices, including NVIDIA NeMo Retriever for retrieval- augmented generation (RAG) inference deployments, will also help OCI customers bring more insight and accuracy to their generative AI copilots and other productivity tools using their own data. To meet escalating customer demand for increasingly complex AI models, the companies are adding NVIDIA Grace Blackwell to NVIDIA DGX? Cloud on OCI.

Customers will be able to access new GB200 NVL72 based instances through this co-engineered supercomputing service designed for energy-efficient training and inference in an era of trillion-parameter LLMs. The full DGX Cloud cluster buildout will include more than 20,000 GB200 accelerators and NVIDIA CX8 InfiniBand networking, providing a highly scalable and performant cloud infrastructure. The cluster will consist of 72 Blackwell GPUs NVL72 and 36 Grace CPUs with fifth-generation NVLink?. Availability: Oracle and NVIDIA's sovereign AI solutions are available immediately.