IBM and AMD have teamed up to fuel Zyphra’s open-source AI ambitions, delivering one of the largest generative AI training environments to date. The San Francisco-based startup, recently valued at $1 billion, will use the collaboration to train frontier multimodal foundation models on a dedicated IBM Cloud cluster powered by AMD Instinct MI300X GPUs, AMD Pensando™ Pollara 400 AI NICs, and AMD Pensando Ortano DPUs.
This marks the first large-scale deployment of AMD’s full-stack AI training platform on IBM Cloud, with Zyphra already running the initial cluster since September and expansion planned for 2026. The infrastructure will underpin Maia, a general-purpose superagent designed to enhance productivity for knowledge workers across enterprises by integrating language, vision, and audio capabilities.
Open-Source Superintelligence Gets a Lift
Zyphra’s lab focuses on advancing neural network architectures, long-term memory, and continual learning. By leveraging IBM Cloud’s scalable infrastructure and AMD’s high-performance accelerators, the company aims to push the boundaries of open-science AI research.
Krithik Puthalath, Zyphra CEO, said, “This collaboration marks the first time AMD’s full-stack training platform has been successfully integrated and scaled on IBM Cloud, and Zyphra is honored to lead the way in developing frontier models.”
Why IBM and AMD?
The choice wasn’t arbitrary. Zyphra selected IBM and AMD for their speed, scalability, and proven AI infrastructure roadmaps. IBM Cloud offers security, reliability, and enterprise-grade support, while AMD contributes power-efficient GPUs and AI-optimized networking for high-performance workloads.
Alan Peacock, GM of IBM Cloud, emphasized the value: “Scaling AI workloads faster and more efficiently is a key differentiator in achieving ROI… we are delighted to support Zyphra’s strategic roadmap.” Philip Guido, AMD EVP, added that the partnership sets “a new standard in AI infrastructure,” supporting advanced AI workloads for both enterprises and startups.
Implications for AI Infrastructure
This collaboration highlights a trend toward full-stack, cloud-based AI infrastructure for open-source and enterprise users alike. By integrating compute, networking, and storage at scale, IBM and AMD are setting a template for high-performance, inference-efficient AI model training.
The partnership also dovetails with broader efforts from IBM and AMD to develop quantum-centric supercomputing architectures, blending IBM’s quantum expertise with AMD’s high-performance computing capabilities—a move that could redefine performance and efficiency for next-generation AI workloads.
Why It Matters
As AI models grow more complex, enterprises and research labs face mounting pressure to secure scalable, efficient training environments. Zyphra’s deployment demonstrates how strategic cloud partnerships and specialized hardware can accelerate breakthroughs in multimodal AI, enabling startups and enterprises alike to experiment, iterate, and deploy AI models faster than ever.
Power Tomorrow’s Intelligence — Build It with TechEdgeAI