Red Hat has launched its AI Factory with NVIDIA, a co-engineered software platform developed with NVIDIA to help enterprises build, deploy and scale AI-enabled applications in production environments.
The platform combines Red Hat AI Enterprise and NVIDIA AI Enterprise, delivering an integrated stack designed for AI inference, model tuning, customisation, and agent deployment across on-premises, cloud and edge environments.
Red Hat said the offering provides Day 0 support for NVIDIA hardware architectures and is aimed at organisations shifting from AI experimentation to production-scale, agentic AI workloads. Enterprise AI spending is projected to exceed $1 trillion by 2029, driven largely by agentic AI applications, the company noted.
The platform is supported on infrastructure from Cisco, Dell Technologies, Lenovo and Supermicro. It is built to help IT teams manage AI workloads alongside traditional enterprise systems with enterprise-grade security and compliance, based on Red Hat Enterprise Linux.
Red Hat AI Factory includes pre-configured models such as the IBM Granite family, NVIDIA Nemotron and NVIDIA Cosmos open models, delivered as NVIDIA NIM microservices. It also integrates NVIDIA NeMo to align models with enterprise data. The serving stack leverages vLLM, NVIDIA TensorRT-LLM and NVIDIA Dynamo for inference optimisation and observability.
Chris Wright, CTO and SVP, Global Engineering, Red Hat, said, “The shift from AI experimentation to industrial-scale, enterprise-wide production requires a fundamental change in how we manage the AI computing stack. We’re accelerating the path to deploy AI and move quickly to production using Red Hat AI Factory with NVIDIA.”
“Enterprises are building AI factories that turn data into intelligence at scale during inference, requiring production-grade infrastructure and software that span the hybrid cloud,” said Justin Boitano, VP of Enterprise AI at NVIDIA.
“Red Hat AI Factory with NVIDIA provides the software foundation that helps organisations keep pace with rapid infrastructure innovation while reliably building and deploying the next generation of agentic AI applications.”
ALSO READ: AMD, Meta Sign Multi-Year Deal to Deploy Up to 6 Gigawatts of AI GPUs