AMD’s approach resonates with a growing community of cloud providers, enterprises, and academic institutions.
On 12 June 2025, Advanced Micro Devices (AMD) hosted its highly anticipated “Advancing AI” conference in San Jose, California, marking a defining moment in the company’s strategic trajectory within the artificial intelligence (AI) sector. This landmark event served not only as a showcase of AMD’s most ambitious technological advancements to date but also as a bold declaration of its intent to lead in the rapidly evolving world of AI infrastructure.
At the forefront of the conference was Dr Lisa Su, AMD’s President and CEO, who delivered a keynote speech that combined both technical expertise and visionary leadership. Her address underscored AMD’s commitment to developing open, collaborative, and energy-efficient AI systems — a deliberate contrast to the more proprietary and closed ecosystems favoured by some of its major competitors, most notably Nvidia. Dr Su emphasised the critical importance of interoperability, scalability, and sustainability, presenting AMD’s roadmap for achieving transformative AI performance across cloud, enterprise, and edge deployments.
The event marked a pivotal chapter in AMD’s intensifying rivalry with Nvidia, showcasing a series of groundbreaking announcements that spanned next-generation hardware accelerators, robust software platforms, cutting-edge cloud infrastructure, and a rapidly expanding network of ecosystem partnerships. From unveiling the high-performance Instinct MI350X GPUs to teasing the upcoming MI400 Series and its revolutionary Helios rack-scale systems, AMD signalled that it is not only closing the gap with Nvidia but potentially redefining the future landscape of AI compute.
By championing an open and inclusive approach, AMD is making a strategic play for leadership in AI — one rooted not only in performance metrics but in the values of accessibility, developer empowerment, and global impact. The “Advancing AI” conference served as a compelling demonstration of that mission in action.
Dr Lisa Su began her keynote by outlining AMD’s ambitious mission to power the “full spectrum of AI”, spanning everything from training colossal foundational models in cloud-scale data centres to delivering low-latency inference at the edge. This broad scope, she explained, is essential for unlocking AI’s full potential across industries — from enterprise applications and scientific computing to consumer devices and intelligent automation. At the heart of her message was a defining theme that echoed throughout the event: openness.
“We believe open ecosystems drive faster innovation, better performance, and more choice. The next decade of AI must be built on collaboration, not closed walls.” -Lisa Su, Advancing AI 2025 Keynote
This powerful statement was not only a clear expression of AMD’s guiding philosophy but also a pointed critique of the current status quo, dominated by Nvidia’s proprietary software and hardware stack. Through its widely adopted CUDA platform and NVLink interconnect technology, Nvidia has created a highly integrated and closed ecosystem which, while technically advanced, restricts flexibility, limits vendor diversity, and slows broader innovation. This approach has increasingly raised concerns among developers and enterprises regarding vendor lock-in and the long-term sustainability of closed infrastructures.
By contrast, AMD’s vision champions open standards, cross-platform compatibility, and community-driven development, especially via its ROCm software platform and its leadership in the UA-Link Consortium, which promotes open GPU-to-GPU interconnect solutions. Dr Su’s remarks underscored a fundamental strategic divergence: that future AI workloads — which will demand scalability, heterogeneity, and power efficiency — will be better supported by open innovation frameworks rather than closed, vertically integrated systems.
In centring openness as a core pillar of its AI roadmap, AMD not only appeals to developers and researchers seeking flexibility and transparency but also positions itself as a key driver of a more democratic, energy-efficient, and forward-thinking AI ecosystem.
The standout moment of AMD’s hardware announcements at the “Advancing AI” conference was undoubtedly the unveiling of the Instinct MI350 Series — the company’s next-generation line of AI accelerators, purpose-built for both training and inference at scale. Representing a major leap forward from its predecessor, the MI300X, this new series is engineered to meet the demands of the most advanced AI workloads, including large language models (LLMs), computer vision systems, and generative AI frameworks.
Key Specifications of the Instinct MI350 Series:
The MI355X, the flagship variant of the series, is designed to maximise both performance and efficiency, positioning it as a direct rival to Nvidia’s Blackwell B200. According to AMD, the MI355X achieves a significantly better tokens-per-dollar ratio — a crucial metric for companies deploying large-scale language models such as OpenAI’s GPT-4, Meta’s LLaMA 3, and Anthropic’s Claude 3. This metric essentially determines how cost-effectively a chip can process AI-generated tokens, making it especially relevant for hyperscale data centres and AI start-ups focused on delivering competitive inference services at scale.
AMD’s decision to support the FP4 format — a lower-precision but highly efficient numerical representation — is especially noteworthy. It allows for more computations per watt and better memory bandwidth utilisation, without significantly compromising model accuracy. This demonstrates AMD’s focus not only on raw compute but also on energy-efficient AI acceleration, a growing priority in data centre operations globally.
By offering performance parity — and in some cases, superiority — to Nvidia’s offerings while promoting open software ecosystems through ROCm, the MI350 Series strengthens AMD’s position as a formidable competitor in the AI silicon market.
Read more: Barron’s – AMD Unveils New AI Chips to Take On Nvidia
Perhaps even more significant than the unveiling of the MI350 Series was the first official glimpse of AMD’s upcoming MI400 Series, due to launch in 2026. Designed to power the next generation of artificial intelligence models, these accelerators are specifically optimised for Mixture-of-Experts (MoE) architectures — a fast-growing technique used in advanced models to dynamically activate only parts of a neural network, allowing for greater scalability and efficiency. AMD claims the MI400 chips will deliver up to a 10× increase in inference performance, setting a new bar for high-throughput AI deployments.
Complementing this breakthrough was the announcement of Helios, AMD’s first rack-scale AI supercomputer, purpose-built to house and exploit the full capabilities of the MI400 GPUs. Helios represents a major expansion in AMD’s strategy — no longer focusing solely on chips but entering the system-level AI infrastructure space.
Helios AI Rack Features:
In contrast to Nvidia’s tightly integrated and proprietary NVLink-based NVL72 “Vera Rubin” system, which has faced criticism for vendor lock-in and limited interoperability, AMD’s Helios prioritises open networking and cross-vendor compatibility. This positions Helios as a compelling alternative for cloud providers and enterprises seeking flexible, scalable, and cost-efficient AI infrastructure, one not confined by the restrictions of a closed ecosystem.
By offering Helios as an open, collaborative platform, AMD is not just launching a supercomputer — it is making a strategic challenge to the very foundations of Nvidia’s dominance in AI systems.
Q1: How do Mixture-of-Experts (MoE) architectures benefit from AMD’s MI400 chips?
MoE architectures work by activating only a subset of model parameters during inference or training, making them far more efficient for large-scale models. AMD’s MI400 chips are tailored for this approach, offering massive parallelism and memory bandwidth while supporting precision formats like FP4 and FP8, crucial for MoE tasks. This allows more dynamic and cost-effective scaling, which is vital for models like GPT-4 and LLaMA 3 that rely on selective expert routing.
Q2: What advantages does Helios offer over Nvidia’s NVL72 Vera Rubin system?
Helios is fundamentally built on principles of openness and interoperability. Unlike Nvidia’s Vera Rubin, which relies on proprietary NVLink connections, Helios supports open interconnects via UA-Link™, enabling mixed-vendor GPU environments. This not only lowers costs but also gives enterprises greater flexibility to integrate Helios into diverse data centre ecosystems without being locked into a single vendor’s stack.
Q3: Why is AMD’s move into rack-scale AI infrastructure significant for the broader industry?
AMD’s entry into rack-scale AI infrastructure signals a shift in industry dynamics. By offering end-to-end solutions — from silicon to systems — AMD can better optimise performance across layers while appealing to clients seeking holistic, open, and energy-efficient alternatives. This move could spur greater competition, encourage standardisation in AI infrastructure, and reduce Nvidia’s control over the high-performance AI market.
Read more: Investopedia – AMD Unveils Its Latest Chips, With ChatGPT Maker OpenAI Among Its Customers
While hardware captured headlines, the software story was just as important. AMD announced the latest version of its open-source AI software platform: ROCm 7.
Higlights:
Dr Anush Elangovan, Corporate VP of AI Software Development, said that AMD will shift to a bi-weekly release cycle, ensuring faster iteration and community engagement. ROCm’s goal is to democratise GPU programming, making it accessible beyond CUDA’s ecosystem.
Read more: Reuters – AMD unveils AI server as OpenAI taps its newest chips
AMD took a strong stance on AI’s growing carbon footprint, citing that current data centre models often consume power on the scale of small cities. Su committed to an ambitious 2030 target: reduce energy use per model by 20x relative to 2024 systems.
Already, the MI350 series exceeds AMD’s 30x energy-efficiency target (first announced in 2021), with a 38x improvement. The Helios system, thanks to its FP4/FP8 support and advanced memory fabric, is also expected to lower energy use per token dramatically.
This positions AMD not just as a performance leader, but a climate-conscious innovator.
Read more: Investor’s Business Daily – AMD Bolsters AI Data Center Pitch With Full-Rack Systems
To streamline experimentation and development, AMD also launched the AMD Developer Cloud, providing direct access to:
This move significantly lowers the barrier to entry for smaller AI startups and individual researchers, removing the need for costly on-premises systems.
One of the most powerful validations of AMD’s AI trajectory came from its partners. Several major players took the stage in support:
OpenAI – Sam Altman
Confirmed that OpenAI is actively using MI300X chips for inference on Microsoft Azure. He also noted that OpenAI is helping design the MI400 series, a strong signal of AMD’s growing influence in foundational AI infrastructure.
Meta – Yee Jiun Song (VP, Infrastructure)
Said that Meta’s engineering teams are already transitioning from MI300X to MI350 GPUs for LLaMA 3 inference — and will adopt MI400 for LLaMA 4 training next year.
Microsoft Azure & Oracle Cloud Infrastructure
Both confirmed deployment of rack-scale AMD GPU clusters for cloud AI services. OCI revealed they are building massive AI clusters with over 130,000 MI355X GPUs.
Cohere, Groq, and Hugging Face
Showcased early benchmarks and LLM inference demos using ROCm and MI350X hardware.
Red Hat & OpenShift AI
Confirmed full ROCm integration into hybrid cloud AI environments, giving enterprises flexible deployment models across public and private clouds.
While AMD’s ambitions were unmistakably evident at the Advancing AI 2025 conference, the road ahead remains intensely competitive. Nvidia, the current market leader in AI compute, recently unveiled its Blackwell B200 and GB200 superchips, which continue to dominate the field of large-scale model training. These chips, coupled with Nvidia’s mature and deeply entrenched software stack, including CUDA, cuDNN, and TensorRT, provide developers with a seamless, highly optimised environment that remains difficult to displace in the short term.
Nonetheless, AMD is steadily making inroads, particularly in areas that matter most to enterprise buyers and cloud service providers. The company’s clear tokens-per-dollar advantage, bolstered by innovations in the Instinct MI355X, represents a significant shift in value proposition. For AI workloads that require large-scale inference, particularly for generative models such as GPT-4, LLaMA 3, or Claude 3, this metric translates directly to lower operational costs — a critical factor in commercial deployment.
AMD’s advantage also extends to energy efficiency and rack-scale availability, thanks to its innovations in the Helios system and the EPYC processor integration. These strengths are increasingly important in a world where AI workloads are putting immense pressure on data centre energy budgets and environmental sustainability goals.
Another compelling aspect of AMD’s approach lies in its commitment to open standards. The company’s leadership in the UA-Link™ Consortium, alongside partners like Astera Labs, Broadcom, and Marvell, signals a serious effort to create a new standard for GPU-to-GPU interconnects that are vendor-neutral and openly accessible. This is in stark contrast to Nvidia’s NVLink, which only supports its own hardware. If successful, UA-Link could drive greater hardware interoperability across the ecosystem, enabling cloud providers, OEMs, and AI researchers to build mixed-vendor systems without performance penalties or proprietary lock-in.
In short, while Nvidia remains a formidable leader, AMD’s multi-pronged strategy — balancing performance, openness, cost-efficiency, and collaboration — could prove increasingly disruptive as the AI infrastructure market matures.
Read more: DataCenterKnowledge – AMD Bolsters AI Chips and Software to Further Challenge Nvidia
The “Advancing AI” event firmly established AMD’s position not only as a serious contender in AI hardware but also as a visionary force advocating for open, sustainable, and scalable AI ecosystems. Rather than merely competing on performance benchmarks, AMD is attempting to redefine the rules of engagement in the AI space by prioritising interoperability, energy efficiency, and developer empowerment.
By launching a powerful new GPU line-up, extending compatibility with its open-source ROCm software stack, and forging strategic alliances with top-tier partners such as OpenAI, Meta, and Microsoft, AMD is building a credible and increasingly attractive alternative to Nvidia’s vertically integrated AI dominance. These partnerships not only validate AMD’s hardware capabilities but also demonstrate a growing industry appetite for vendor diversity and open innovation.
It’s still early days in what is likely to be a decade-long race, but if AMD can maintain its executional momentum and meet its ambitious performance and efficiency targets, it could fundamentally reshape the AI compute landscape, and do so on its own open, inclusive terms. Importantly, AMD’s approach resonates with a growing community of cloud providers, enterprises, and academic institutions seeking to break free from vendor lock-in and co-create the next generation of AI infrastructure.
In championing open standards and broad ecosystem collaboration, AMD is not just playing catch-up; it is setting the stage for a more democratised and accessible future of AI, one where innovation is not confined to a single platform, but is distributed across a rich and diverse technological landscape.