Page 17 - Red Hat PR REPORT - MAY-JUNE 2025
P. 17
Press Release
Red Hat Optimizes Red Hat AI to Speed Enterprise AI
Deployments Across Models, AI Accelerators and Clouds
Red Hat AI Inference Server, validated models and integration of Llama Stack and
Model Context Protocol help users deliver higher-performing, more consistent AI
applications and agents
MAY 22, 2025 — Red Hat, the world's leading provider of open source solutions, today continues
to deliver customer choice in enterprise AI with the introduction of Red Hat AI Inference Server,
Red Hat AI third-party validated models and the integration of Llama Stack and Model Context
Protocol (MCP) APIs, along with significant updates across the Red Hat AI portfolio. With these
developments, Red Hat intends to further advance the capabilities organizations need to accelerate
AI adoption while providing greater customer choice and confidence in generative AI (gen AI)
production deployments across the hybrid cloud.
According to Forrester, open source software will be the spark for accelerating enterprise AI
1
efforts. As the AI landscape grows more complex and dynamic, Red Hat AI Inference Server and
third party validated models provide efficient model inference and a tested collection of AI models
optimized for performance on the Red Hat AI platform. Coupled with the integration of new APIs
for gen AI agent development, including Llama Stack and MCP, Red Hat is working to tackle
deployment complexity, empowering IT leaders, data scientists and developers to accelerate AI
initiatives with greater control and efficiency.
Efficient inference across the hybrid cloud with Red Hat AI Inference Server
The Red Hat AI portfolio now includes the new Red Hat AI Inference Server, providing faster,
more consistent and cost-effective inference at scale across hybrid cloud environments. This key
addition is integrated into the latest releases of Red Hat OpenShift AI and Red Hat Enterprise
Linux AI, and is also available as a standalone offering, enabling organizations to deploy
intelligent applications with greater efficiency, flexibility and performance.
Tested and optimized models with Red Hat AI third party validated models
Red Hat AI third party validated models, available on Hugging Face, make it easier for
enterprises to find the right models for their specific needs. Red Hat AI offers a collection of
validated models, as well as deployment guidance to enhance customer confidence in model
performance and outcome reproducibility. Select models are also optimized by Red Hat,
leveraging model compression techniques to reduce size and increase inference speed, helping to
minimize resource consumption and operating costs. Additionally, the ongoing model validation
process helps Red Hat AI customers continue to stay at the forefront of optimized gen AI
innovation.
Standardized APIs for AI application and agent development with Llama Stack and MCP
Red Hat AI is integrating Llama Stack, initially developed by Meta, along with Anthropic’s MCP,
to provide users with standardized APIs for building and deploying AI applications and agents.
Currently available in developer preview in Red Hat AI, Llama Stack provides a unified API to