Red Hat today announced the Red Hat AI Factory with NVIDIA, a co-engineered software platform that integrates Red Hat AI Enterprise and NVIDIA AI Enterprise to deliver a unified, enterprise-grade foundation for building, deploying, and scaling AI-enabled applications. This solution accelerates the journey to production AI by providing optimized performance on accelerated computing infrastructure, Day 0 support for NVIDIA hardware architectures, and support from leading system manufacturers including Cisco, Dell Technologies, Lenovo, and Supermicro.
With enterprise AI spending projected to exceed $1 trillion by 2029—driven significantly by agentic AI applications—organizations require high-density, inference-optimized workflows that maintain operational predictability. Red Hat AI Factory with NVIDIA addresses these demands by streamlining management of both traditional infrastructure and evolving AI stacks, enabling IT operations teams to scale deployments with the same rigor applied to any enterprise workload.
The platform combines open source collaboration, engineering expertise, and enterprise support from Red Hat and NVIDIA to offer a highly scalable foundation across on-premises, cloud, and edge environments. It emphasizes high-performance AI inference, model tuning, customization, agent deployment, and built-in security, allowing organizations to retain architectural control from the datacenter to public clouds.
Organizations gain immediate access to ready-to-deploy models via NVIDIA NIM microservices, including the indemnified IBM Granite family, NVIDIA Nemotron, and NVIDIA Cosmos open models. Further customization aligns models to enterprise-specific data using NVIDIA NeMo, significantly reducing tuning time and associated costs while advancing production readiness.
The unified serving stack maximizes infrastructure utilization and inference performance to meet demanding service level objectives. Built-in observability and advanced inference engines powered by vLLM, NVIDIA TensorRT-LLM, and NVIDIA Dynamo optimize the connection between models and NVIDIA GPUs, helping organizations lower total cost of ownership while delivering consistent, high-throughput AI services.
Red Hat AI Factory with NVIDIA enables on-demand GPU resource allocation through intelligent orchestration and pooled infrastructure. Automatic checkpointing safeguards long-running jobs, ensuring greater predictability and cost control even in fluctuating compute demands.
Built on the stable foundation of Red Hat Enterprise Linux, the platform incorporates advanced security features from the outset to reduce risk, minimize downtime, and support compliance requirements. NVIDIA DOCA microservices extend this with zero-trust architecture and runtime security across the AI infrastructure, providing isolation and continuous verification for mission-critical workloads.
This collaboration marks a significant step in delivering trusted, scalable AI solutions that empower enterprises to deploy agentic and high-density AI applications with confidence, performance, and operational consistency.
About Red Hat
Red Hat is the world’s leading provider of open source solutions, using a community-powered approach to deliver high-performing Linux, hybrid cloud, container, and Kubernetes technologies. Red Hat helps customers integrate new and existing IT applications, develop cloud-native applications, and standardize across environments—physical, virtual, containerized, and across public cloud providers. A subsidiary of IBM, Red Hat is headquartered in Raleigh, North Carolina.