Oracle Cloud Infrastructure (OCI) Compute provides industry-leading scalability and cost-performance for bare metal and virtual machine (VM) instances powered by NVIDIA GPUs for mainstream graphics, AI inference, AI training, digital twins, and HPC.
OCI Supercluster with NVIDIA H200 Tensor Core and AMD MI300X can support tens of thousands of GPUs, with added benefits such as hardware acceleration, bare metal instances with no hypervisor overhead, and much more.
Scalability
131,072
Maximum number of GPUs in an OCI Supercluster1
Performance
3,200
Up to 3,200 Gb/sec of RDMA cluster network bandwidth2
Value
220%
GPUs for other CSPs can be up to 220% more expensive3
Choice
VM/BM
Rightsizing with VM and performance with bare metal instances
1. OCI Supercluster scales up to 131,072 NVIDIA B200 GPUs (planned); more than 100,000 NVIDIA B200 GPUs in NVIDIA GB200 Superchips (planned); 65,536 H200 GPUs; 32,768 NVIDIA A100 GPUs; 16,384 NVIDIA H100 GPUs; and 16,384 AMD MI300X GPUs.
2. For bare metal instances with NVIDIA H100 GPUs and AMD MI300X GPUs.
3. Based on on-demand pricing as of June 5, 2024.
OCI is the only major cloud provider to offer bare metal instances with NVIDIA and AMD GPUs for high performance that’s free of virtualization overhead. For checkpointing during AI training, our instances provide the most local storage per node (61.4 TB with H100 GPUs). For a balance of performance and price, OCI VMs with NVIDIA GPUs are consistently cheaper than AWS and Azure.
OCI offers the highest value and performance for bare metal and virtual machine compute instances powered by NVIDIA H100 Tensor Core GPUs, L40S GPUs, A100 Tensor Core GPUs, A10 Tensor Core GPU, and older-generation NVIDIA GPUs. OCI plans to offer instances with NVIDIA H200 and Blackwell GPUs.
OCI offers the NVIDIA GH200 Grace Hopper Superchip and plans to offer the GB200 Grace Blackwell Superchip for LLM inference.
OCI offers AMD Instinct MI300X GPUs with 192 GB of memory at a competitive price.
Oracle’s ultralow-latency cluster networking, based on remote direct memory access (RDMA), provides microsecond-level latency.
For VMs, choose from NVIDIA’s Hopper, Ampere, and older GPU architectures with one to four cores, 16 to 64 GB of GPU memory per VM, and up to 48 Gb/sec of network bandwidth.
Use OCI Supercluster with bare metal instances that include AMD Instinct GPUs, NVIDIA Blackwell GPUs or Superchips, NVIDIA Hopper GPUs or Superchips, and NVIDIA Ampere GPUs.
Take advantage of managed Kubernetes, service mesh, and container registry to orchestrate AI and machine learning (ML) training and inference with containers.
Competing GPU instances from AWS and Azure can be consistently more expensive.
AWS, Azure, and Google Cloud Platform can be up to 6X more expensive.
AWS, Azure, and Google Cloud Platform can be up to 2X more expensive.
Public bandwidth transferred out on OCI can be up to an order of magnitude cheaper than AWS, Azure, and Google Cloud Platform.
Oracle Cloud Marketplace provides software and disk images for data science, analytics, artificial intelligence (AI), and machine learning (ML) models so customers can quickly gain insight from their data.
Get access to NVIDIA AI Enterprise, an end-to-end software platform for data science and production AI, including generative AI, computer vision, and speech AI.
NVIDIA DGX Cloud on OCI is an AI-training-as-a-service platform, offering a serverless experience for developers that’s optimized for generative AI.
Use NVIDIA GPU Cloud Machine Image for hundreds of GPU-optimized applications for machine learning, deep learning, and high performance computing covering a wide range of industries and workloads.
Deliver powerful workstation performance wherever employees need it by running NVIDIA RTX Virtual Workstation on Oracle Cloud.
When combined with GPU compute, OCI’s distributed cloud helps organizations run AI and cloud services where and how they’re needed.
Support data residency within a region or country, including the EU, the US, the UK, and Australia.
Deploy a complete cloud region in your data center with OCI Dedicated Region to retain full control of your data and applications.
Become a partner for Oracle Alloy and deliver your cloud services to address specific market needs.
Developers building applications using containers leverage a highly available, Oracle-managed private container registry service for storing and sharing container images. Push or pull Docker images to and from the registry using the Docker V2 API and the standard Docker command line interface (CLI). Images can be pulled directly into a Kubernetes deployment.
Functions as a service (FaaS) lets developers run serverless applications that integrate with Oracle Cloud Infrastructure, Oracle Cloud Applications, and third-party services. Gain developer efficiency along with the community of the open source Fn Project.
Train AI models using OCI Data Science, bare metal instances, cluster networking based on RDMA, and NVIDIA GPUs.
OCI Compute powered by NVIDIA GPUs provide consistent high performance for VDI.
OCI enables computer-aided engineering and computational fluid dynamics for fast predictions of the aerodynamic properties of objects.
We’re excited to announce the general availability of Oracle Cloud Infrastructure (OCI) Supercluster with NVIDIA H200 Tensor Core GPUs. The largest AI supercomputer available in the cloud, our latest Supercluster scales up to an industry-leading 65,536 GPUs.
Read the complete postOracle offers a free pricing tier for most AI services as well as a free trial account with US$300 in credits to try additional cloud services. AI services are a collection of offerings, including generative AI, with prebuilt machine learning models that make it easier for developers to apply AI to applications and business operations.
You also only have to pay compute and storage charges for OCI Data Science.
Learn more about AI infrastructure, AI services and generative AI, and compute.
Oracle Cloud pricing is simple, with consistent low pricing worldwide, supporting a wide range of use cases. To estimate your low rate, check out the cost estimator and configure the services to suit your needs.
Get help building your next GPU solution or deploying your AI workload on OCI AI infrastructure.