DigitalOcean Elevates its Agentic Inference Cloud with GPU Droplets powered by AMD Instinct™ MI350X GPUs
Key Terms
gpu technical
generative ai technical
high-performance computing technical
architecture technical
latency technical
throughput technical
token technical
DigitalOcean and AMD collaborate to deliver new GPUs with lower latency and higher throughput for complex inference workloads
Optimizing production inference with AMD Instinct™ MI350X GPUs
AMD Instinct™ MI350X Series GPUs set a new standard for generative AI and high-performance computing (HPC). Built on the AMD CDNA™ 4 architecture, these GPUs deliver cutting-edge efficiency and performance for training massive AI models, high-speed inference, and complex HPC workloads including scientific simulations, data processing, and computational modeling. The capabilities of the GPUs allow DigitalOcean to optimize for compute bound prefill phase, while enabling high-performance inference at low latency and high token generation throughput. This provides the ability to load large models and larger context windows, leading to supporting a higher inference request density per GPU. Paired with DigitalOcean’s optimized inference platform, these feature enhancements of AMD Instinct™ MI350X GPUs offer lower latency and higher throughput.
“These results demonstrate that the DigitalOcean Agentic Inference Cloud isn't just about providing raw compute, but about delivering the operational efficiency, inference optimizations, and scale required for demanding AI builders,” said Vinay Kumar, Chief Product and Technology Officer at DigitalOcean. “The availability of the AMD Instinct™ MI350X GPUs, combined with DigitalOcean’s inference optimized platform offers our customers a boost in performance and the massive memory capacity needed to run the world’s most complex AI workloads while delivering compelling unit economics.”
Earlier this year, DigitalOcean announced that by optimizing AMD Instinct™ GPUs, they were able to deliver 2X production request throughput and a
"Our collaboration with DigitalOcean is rooted in a shared commitment to pairing leadership AI infrastructure with a platform designed to make large-scale AI applications more accessible to the world’s most ambitious developers and enterprises,” said Negin Oliver, Corporate Vice President of Business Development, Data Center GPU Business at AMD. “By bringing the AMD Instinct™ MI350 Series GPUs to DigitalOcean’s Agentic Inference Cloud, we are empowering startups and enterprises alike to deploy and scale next-generation AI workloads with confidence.”
This initiative builds upon previous collaboration between DigitalOcean and AMD, including the launch of the AMD Developer Cloud and the release of AMD Instinct™ MI300X and MI325X GPUs on DigitalOcean last year.
Enterprise performance with predictable cost-efficiency and simple operations at the forefront
In addition to offering the latest AMD GPUs, DigitalOcean maintains its commitment to transparency and simplicity, ensuring this powerful technology is easy to adopt for developers and emerging businesses:
- Cost-effective, predictable pricing: DigitalOcean offers transparent, usage based pricing with flexible contracts and no hidden fees.
- Simple setup: GPU Droplets can be provisioned and configured with security, storage, and networking requirements in just a few clicks, drastically simplifying deployment compared to complex cloud environments.
- Access to enterprise features: GPU Droplets offer enterprise-grade SLAs, observability features, and are HIPAA-eligible and SOC 2 compliant.
The new GPU Droplets powered by AMD Instinct™ MI350X are available in the Atlanta Region Datacenter.
To learn more about using AMD Instinct™ MI350X GPUs on DigitalOcean, please visit the DigitalOcean website.
About DigitalOcean
DigitalOcean is the Agentic Inference Cloud built for AI-native and Digital-native enterprises scaling production workloads. The platform combines production-ready GPU infrastructure with a full-stack cloud to deliver operational simplicity and predictable economics at scale. By integrating inference capabilities with core cloud services, DigitalOcean’s Agentic Inference Cloud enables customers to expand as they grow — driving durable, compounding usage over time. More than 640,000 customers trust DigitalOcean to power their cloud and AI infrastructure. To learn more, visit www.digitalocean.com
View source version on businesswire.com: https://www.businesswire.com/news/home/20260219844245/en/
Media Relations
Julie Wolf: press@digitalocean.com
Investor Relations
Melanie Strate: investors@digitalocean.com
Source: DigitalOcean