Skip to main content

Red Hat and Google Cloud Extend Alliance to Drive Open and Agentic AI for the Enterprise

  • Enhanced vLLM integration on Google Cloud TPUs and availability of Red Hat AI Inference Server on Google Cloud enable more cost-effective AI inference and hardware choice for gen AI at scale
  • Google Cloud joins Red Hat’s new llm-d open source project as founding contributor
  • Red Hat to become early tester for Google’s open Gemma 3 model, providing Day 0 support

Red Hat, the world's leading provider of open source solutions, and Google Cloud today announced an expanded collaboration to advance AI for enterprise applications by uniting Red Hat’s open source technologies with Google Cloud’s purpose-built infrastructure and Google’s family of open models, Gemma.

Together, the companies will advance enterprise-grade use cases for scaling AI by:

  • Launching the llm-d open source project with Google as a founding contributor
  • Enabling support for vLLM on Google Cloud TPUs and GPU-based virtual machines (VMs) to enhance AI inference
  • Delivering Day 0 support for vLLM on Gemma 3 model distributions
  • Supporting Red Hat AI Inference Server on Google Cloud
  • Propelling agentic AI with Red Hat as a community contributor for Google’s Agent2Agent (A2A) protocol

Bolstering AI inference with vLLM

Demonstrating its commitment to Day 0 readiness, Red Hat is now an early tester for Google’s family of open models, Gemma, starting with Gemma 3, delivering immediate support for vLLM. vLLM is an open source inference server that speeds the output of generative AI (gen AI) applications. As the leading commercial contributor to vLLM, Red Hat is driving a more cost-efficient and responsive platform for gen AI applications.

Additionally, Google Cloud TPUs, the high-performance AI accelerators powering Google’s AI portfolio, are now fully enabled on vLLM. This integration empowers developers to maximize resources while achieving the performance and efficiency crucial for fast and accurate inference.

Recognizing the shift from AI research to real-world deployment, many organizations face the complexities of a diverse AI ecosystem and the need to shift to more distributed compute strategies. To address this, Red Hat has launched the llm-d open source project, with Google as a founding contributor. Building on the momentum of the vLLM community, this initiative pioneers a new era of gen AI inference. The goal is to enable greater scalability across heterogeneous resources, optimize costs and enhance workload efficiency – all while fostering continued innovation.

Driving enterprise AI with community-powered innovation

Bringing the latest upstream community advancements to the enterprise, Red Hat AI Inference Server is now available on Google Cloud. As Red Hat’s enterprise distribution of vLLM, Red Hat AI Inference Server helps enterprises optimize model inference across their entire hybrid cloud environment. By leveraging the robust and trusted infrastructure of Google Cloud, enterprises can deploy production-ready gen AI models that are both highly responsive and cost-efficient at scale.

Underscoring their joint commitment to open AI, Red Hat is also now contributing to Google’s Agent2Agent (A2A) protocol – an application-level protocol facilitating more seamless communication between end-users or agents across diverse platforms and cloud environments. By actively participating in the A2A ecosystem, Red Hat aims to help users unlock new avenues for rapid innovation, ensuring AI workflows remain dynamic and highly effective through the power of agentic AI.

Red Hat Summit

Join the Red Hat Summit keynotes to hear the latest from Red Hat executives, customers and partners:

Supporting Quotes

Brian Stevens, senior vice president and Chief Technology Officer – AI, Red Hat

"With this extended collaboration, Red Hat and Google Cloud are committed to driving groundbreaking AI innovations with our combined expertise and platforms. Bringing the power of vLLM and Red Hat open source technologies to Google Cloud and Google’s Gemma equips developers with the resources they need to build more accurate, high-performing AI solutions, powered by optimized inference capabilities.”

Mark Lohmeyer, vice president and general manager, AI and Computing Infrastructure, Google Cloud

“The deepening of our collaboration with Red Hat is driven by our shared commitment to foster open innovation and bring the full potential of AI to our customers. As we enter a new age of AI inference, together we are paving the way for organizations to more effectively scale AI inference and enable agentic AI with the necessary cost-efficiency and high performance.”

Additional Resources

Connect with Red Hat

About Red Hat

Red Hat is the open hybrid cloud technology leader, delivering a trusted, consistent and comprehensive foundation for transformative IT innovation and AI applications. Its portfolio of cloud, developer, AI, Linux, automation and application platform technologies enables any application, anywhere—from the datacenter to the edge. As the world's leading provider of enterprise open source software solutions, Red Hat invests in open ecosystems and communities to solve tomorrow's IT challenges. Collaborating with partners and customers, Red Hat helps them build, connect, automate, secure and manage their IT environments, supported by consulting services and award-winning training and certification offerings.

About Google Cloud

Google Cloud is the new way to the cloud, providing AI, infrastructure, developer, data, security, and collaboration tools built for today and tomorrow. Google Cloud offers a powerful, fully integrated and optimized AI stack with its own planet-scale infrastructure, custom-built chips, generative AI models and development platform, as well as AI-powered applications, to help organizations transform. Customers in more than 200 countries and territories turn to Google Cloud as their trusted technology partner.

Forward-Looking Statements

Except for the historical information and discussions contained herein, statements contained in this press release may constitute forward-looking statements within the meaning of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are based on the company’s current assumptions regarding future business and financial performance. These statements involve a number of risks, uncertainties and other factors that could cause actual results to differ materially. Any forward-looking statement in this press release speaks only as of the date on which it is made. Except as required by law, the company assumes no obligation to update or revise any forward-looking statements.

Red Hat and the Red Hat logo are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in the U.S. and other countries.

Contacts

Stock Quote API & Stock News API supplied by www.cloudquote.io
Quotes delayed at least 20 minutes.
By accessing this page, you agree to the following
Privacy Policy and Terms Of Service.