New co-engineered offering combines Red Hat AI Enterprise and NVIDIA’s accelerated computing software to provide a unified foundation for building, deploying, and scaling AI-enabled applications
Red Hat, the world’s leading provider of open source solutions, announced the Red Hat AI Factory with NVIDIA,a co-engineered software platform that combines Red Hat AI Enterprise and NVIDIA AI Enterprise to provide an end-to-end AI solution optimized for organizations deploying AI at scale. Red Hat AI Factory with NVIDIA is the latest milestone in the companies’ deep collaboration, accelerating the delivery of the newest AI innovations to enterprise customers today while also delivering Day 0 support for NVIDIA hardware architectures.
With enterprise AI spending expected to reach over $1 trillion by 20291, driven in large part by agentic AI applications, organizations are looking to shift their strategies toward high-density, agentic workflows and address the resulting demands on AI inference and infrastructure. To help organizations keep pace, Red Hat AI Factory with NVIDIA empowers IT operations teams to streamline management of both traditional infrastructure and the evolving demands of the AI stack.
Red Hat AI Factory with NVIDIA accelerates the path to production AI and delivers the software platform for AI factories, running on accelerated computing infrastructure that fuels higher performance for the models and NVIDIA GPUs driving the inference stack. The platform is supported on AI factory infrastructure from leading systems manufacturers, including Cisco, Dell Technologies, Lenovo and Supermicro. This empowers IT administrators and operations teams to scale and maintain AI deployments with the same operational rigor and predictability as any enterprise workload.
Also Read: Databricks Advances Real-Time Data Pipelines with Zerobus Ingest in Lakeflow Connect
This co-engineered software platform integrates the open source collaboration, engineering and support expertise of both Red Hat and NVIDIA to deliver a trusted, enterprise-grade solution. The Red Hat AI Factory with NVIDIA provides a highly scalable foundation for AI deployments across any environment, whether on-premises, in the cloud or at the edge. It includes core capabilities for high-performance AI inference, model tuning, customization and agent deployment and management, with a focus on security.
This allows organizations to maintain architectural control from the datacenter to the public cloud, delivering:
- Accelerated time-to-value: Advance to production AI with streamlined workflows and instant access to pre-configured models, including the indemnified IBM Granite family, NVIDIA Nemotron, and NVIDIA Cosmos open models, delivered as NVIDIA NIM microservices. Additionally, organizations can further align models to enterprise data using NVIDIA NeMo, reducing tuning time and cost.
- Optimized performance and cost: Maximize infrastructure usage and bolster inference performance with a unified, high-performance serving stack. Red Hat AI Factory with NVIDIA delivers built-in observability capabilities and taps Red Hat AI inference capabilities powered by vLLM, NVIDIA TensorRT-LLM, and NVIDIA Dynamo to meet strict AI service level objectives. This helps organizations reduce the total cost of ownership (TCO) for AI by optimizing the connection between models and NVIDIA GPUs.
- Intelligent GPU orchestration: Enable on-demand access to GPU resources through intelligent orchestration and pooled infrastructure, with automatic checkpointing to protect long-running jobs and maintain more predictable compute costs in dynamic environments.
- Strengthened enterprise posture: Leveraging the flexible and stable foundation of Red Hat Enterprise Linux, organizations benefit from advanced security and compliance capabilities built-in from the start that help to lower risk, save time and mitigate downtime. This delivers a security-hardened foundation for mission-critical AI workloads that require isolation and continuous verification. NVIDIA DOCA microservices build on this foundation, creating a zero-trust architecture and delivering AI runtime security across the infrastructure.
Source: Businesswire


