Page 17 - Red Hat PR REPORT - MAY-JUNE 2025
P. 17

Press Release




                       Red Hat Optimizes Red Hat AI to Speed Enterprise AI
                     Deployments Across Models, AI Accelerators and Clouds



               Red Hat AI Inference Server, validated models and integration of Llama Stack and
               Model Context Protocol help users deliver higher-performing, more consistent AI
               applications and agents

               MAY 22, 2025 — Red Hat, the world's leading provider of open source solutions, today continues
               to deliver customer choice in enterprise AI with the introduction of Red Hat AI Inference Server,
               Red Hat AI third-party validated models and the integration of Llama Stack and Model Context
               Protocol (MCP) APIs, along with significant updates across the Red Hat AI portfolio. With these
               developments, Red Hat intends to further advance the capabilities organizations need to accelerate
               AI adoption while providing greater customer choice and confidence in generative AI (gen AI)
               production deployments across the hybrid cloud.

               According to  Forrester,  open source  software will be the  spark for accelerating  enterprise  AI
                      1
               efforts.  As the AI landscape grows more complex and dynamic, Red Hat AI Inference Server and
               third party validated models provide efficient model inference and a tested collection of AI models
               optimized for performance on the Red Hat AI platform. Coupled with the integration of new APIs
               for gen AI agent development, including Llama Stack and MCP, Red Hat is working to tackle
               deployment complexity, empowering IT leaders, data scientists and developers to accelerate AI
               initiatives with greater control and efficiency.

               Efficient inference across the hybrid cloud with Red Hat AI Inference Server
               The Red Hat AI portfolio now includes the new Red Hat AI Inference Server, providing faster,
               more consistent and cost-effective inference at scale across hybrid cloud environments. This key
               addition is integrated into the latest releases of Red Hat OpenShift AI and Red Hat Enterprise
               Linux AI,  and is also available as a standalone offering, enabling organizations to deploy
               intelligent applications with greater efficiency, flexibility and performance.

               Tested and optimized models with Red Hat AI third party validated models
               Red Hat AI third party validated models, available on  Hugging Face, make it easier for
               enterprises to find the right models for their specific needs. Red Hat AI offers a collection of
               validated models,  as well as deployment guidance to enhance customer confidence in model
               performance  and outcome reproducibility. Select models are also optimized by Red Hat,
               leveraging model compression techniques to reduce size and increase inference speed, helping to
               minimize resource consumption and operating costs. Additionally, the ongoing model validation
               process helps Red  Hat  AI customers  continue  to stay at the forefront of optimized gen AI
               innovation.

               Standardized APIs for AI application and agent development with Llama Stack and MCP
               Red Hat AI is integrating Llama Stack, initially developed by Meta, along with Anthropic’s MCP,
               to provide users with standardized APIs for building and deploying AI applications and agents.
               Currently available in developer preview in Red Hat AI, Llama Stack provides a unified API to
   12   13   14   15   16   17   18   19   20   21   22