Supermicro Delivers Performance and Efficiency Optimized Liquid-Cooled and Air-Cooled AI Solutions with AMD Instinct™ MI350 Series GPUs and Platforms
Rhea-AI Summary
Positive
- 40% more tokens-per-dollar efficiency compared to competitors
- Up to 40% power consumption reduction with new Direct Liquid Cooling architecture
- 1.5x memory capacity increase with 288GB HBM3e per GPU
- 1.8x performance improvement in FP16/FP8 compared to previous generation
Negative
- None.
News Market Reaction
On the day this news was published, SMCI gained 0.35%, reflecting a mild positive market reaction.
Data tracked by StockTitan Argus on the day of publication.
- Supermicro launches highly optimized AI solutions, based on AMD Instinct MI350 Series GPUs and AMD ROCm™ Software, delivering breakthrough inference performance and power efficiency.
- The new Supermicro H14 GPU solutions are powered by the newest 4th generation AMD CDNA™ architecture, delivering optimized performance and efficiency for large AI training models and high-speed inference workloads
- Large memory capacity with 2.304TB total HBM3e per 8-GPU server which delivers faster computation and more efficient scaling for AI, inferencing, and training
"Supermicro continues to lead the industry with the most experience in delivering high-performance systems designed for AI and HPC applications," said Charles Liang, president and CEO of Supermicro. "Our Data Center Building Block Solutions® enable us to quickly deploy end-to-end data center solutions to market, bringing the latest technologies for the most demanding applications. The addition of the new AMD Instinct MI350 series GPUs to our GPU server lineup strengthens and expands our industry-leading AI solutions and gives customers greater choice and better performance as they design and build the next generation of data centers."
Learn more about Supermicro Servers with AMD Instinct MI350 Series GPUs here.
Supermicro's H14 generation data center solutions offer unparalleled performance and flexibility for a wide range of AI, HPC, Cloud, and Enterprise workloads. These systems are built on Supermicro's proven building-block architecture to help enterprise customers efficiently upgrade and scale their workloads. The new Supermicro GPU servers feature both the latest AMD EPYC CPU and AMD Instinct GPUs which together offer a powerhouse family of AI-enabled solutions, adding to Supermicro's data center building block solutions (DCBBS) ensuring customer success from consultation to on-site deployment and support.
"Our new AMD Instinct MI350 series delivers up to 40 percent more tokens-per-dollar relative to the competition, while retaining the industry standard form factor for OEM infrastructure compatibility," said Dr. Lisa Su, CEO and Chair, AMD. "By combining these GPUs with Supermicro's proven platforms, their customers can deploy fully integrated, air- or liquid-cooled racks built entirely on AMD technology, giving them the flexibility and efficiency they need to deploy robust AI solutions at any scale."
Supermicro is expanding its proven liquid-cooled and air-cooled high-performance fabric 8-GPU system, leveraging the open multi-architecture industry standard OCP Accelerator Module (OAM) to support the latest AMD Instinct MI350 series GPUs. For higher-density environments, the 4U liquid-cooled system with AMD Instinct MI355X GPUs features Supermicro's new improved Direct Liquid Cooling (DLC) architecture incorporating new technologies for cooling various server components and reducing power consumption by up to
Designed to maximize computational throughput, memory bandwidth utilization, and energy efficiency to enable more power-efficient AI inference. these accelerated GPU servers with AMD Instinct MI350 series will offer 288GB HBM3e per GPU, boosting 1.5x memory capacity compared to previous generations of AMD Instinct accelerator, 8TB/s bandwidth, and 1.8x petaflops of FP16 / FP8 compared to the previous generation Instinct MI325X, so customers can process more data faster for their AI workloads.
"AI models aren't just increasing in size; they're demanding faster, more efficient infrastructure that can stay consistently deployable in real-world environments," said Paul Schell, Industry Analyst at ABI Research. "By adding AMD MI350 series GPUs into its latest H14 generation, Supermicro demonstrates a commitment to providing scalable, high-performance and memory solutions that are optimized for both AI training and inference. With support for both liquid and air cooling, these systems offer the flexibility and efficiency that CSPs, neo-clouds and enterprises alike need to power the next wave of AI growth."
These new GPU solutions are designed for powering AI at scale, across cloud service providers and enterprise partners, built on the 4th Gen AMD CDNA architecture, the AMD Instinct MI350 Series delivers exceptional performance and energy efficiency for a wide range of workloads, from AI training and inference to complex scientific simulations. The new AMD MI350 series will also feature new FP6 and FP4 data types to provide exceptional AI capabilities, handling larger models for AI deployments.
These Supermicro GPU systems supporting the AMD Instinct GPUs are available to order:
Liquid-Cooled GPU, 4U https://www.supermicro.com/en/products/system/gpu/4u/as%20-4126gs-nmr-lcc
Air-Cooled GPU, 8U
https://www.supermicro.com/en/products/system/gpu/8u/as%20-8126gs-tnmr
About Super Micro Computer, Inc.
Supermicro (NASDAQ: SMCI) is a global leader in Application-Optimized Total IT Solutions. Founded and operating in
Supermicro, Server Building Block Solutions, and We Keep IT Green are trademarks and/or registered trademarks of Super Micro Computer, Inc.
All other brands, names, and trademarks are the property of their respective owners.
View original content to download multimedia:https://www.prnewswire.com/news-releases/supermicro-delivers-performance-and-efficiency-optimized-liquid-cooled-and-air-cooled-ai-solutions-with-amd-instinct-mi350-series-gpus-and-platforms-302480186.html
SOURCE Super Micro Computer, Inc.