Industry News | 6/17/2025

AMD Introduces MI350 Series to Compete with Nvidia in AI Hardware

AMD is set to launch its Instinct MI350 series of AI accelerators, boasting significant memory capacity advantages over Nvidia's offerings. However, challenges remain in networking capabilities and software ecosystem as AMD aims to capture a larger share of the AI hardware market.

AMD Introduces MI350 Series to Compete with Nvidia in AI Hardware

Advanced Micro Devices (AMD) is preparing to launch its new Instinct MI350 series of accelerators in the third quarter of this year, aiming to challenge Nvidia's dominance in the artificial intelligence hardware market. The MI350 series is designed with a focus on enhanced memory capacity, a crucial aspect for training and deploying large AI models.

Key Features of the MI350 Series

The MI350X and MI355X models will feature 288GB of HBM3E memory, representing a 60% increase compared to Nvidia's B200 GPU. This substantial memory capacity allows a single AMD chip to support AI models with up to 520 billion parameters. Additionally, both MI350 models will offer a memory bandwidth of 8 TBps, matching Nvidia's offerings, which is intended to improve throughput for AI training and inference tasks.

AMD claims that the MI355X can deliver up to 40% more tokens per dollar during inference compared to Nvidia's B200, providing a compelling value proposition for potential customers. In terms of raw compute power, AMD asserts that the MI350 series is competitive with Nvidia's Blackwell B200, particularly in newer floating-point formats used in AI applications.

Networking Challenges

Despite the advantages in memory capacity, AMD faces challenges in networking capabilities. Nvidia has established a strong position with its NVLink and InfiniBand technologies, which facilitate large-scale GPU clusters. For instance, the Blackwell GB200 NVL72 can connect 72 GPUs within a single rack. In contrast, AMD's MI350 series currently supports the connection of only eight GPUs using its Infinity Fabric technology.

While AMD promotes open standards like Ultra Ethernet for networking, it lacks a direct competitor to Nvidia's integrated rack-scale systems, which is critical for large enterprises and hyperscalers that require high-speed inter-GPU communication. Analysts note that while the MI350 series supports 400 Gbit/s per GPU for scale-out, upcoming Nvidia products are expected to offer 800 Gbit/s per GPU networking.

Software Ecosystem Considerations

Another significant hurdle for AMD is its software ecosystem, ROCm (Radeon Open Compute). Although ROCm is an open-source alternative, it is still working to gain traction against Nvidia's well-established CUDA platform, which has benefited from over a decade of development. CUDA's mature ecosystem includes extensive libraries and broad framework support, giving Nvidia a considerable advantage in the AI and high-performance computing sectors.

AMD is actively working to improve its software ecosystem by collaborating with the open-source community and partners like PyTorch and Hugging Face. The recent announcement of ROCm 7, which includes support for Windows and major Linux distributions, is a step towards enhancing accessibility for developers. Additionally, tools like HIP are being promoted to help port CUDA code to a more hardware-agnostic platform.

Conclusion

The launch of AMD's Instinct MI350 series represents a significant advancement in its efforts to provide a competitive alternative to Nvidia in the AI accelerator market. The notable memory capacity of the MI350 chips is likely to attract customers focused on large language models and memory-intensive applications. However, AMD must continue to address its limitations in high-speed networking and work to strengthen its software ecosystem to effectively compete in this rapidly evolving market.