Page 132 - Red Hat PR REPORT - OCTOBER 2025
P. 132

10/16/25, 11:30 AM         Red Hat Brings Distributed AI Inference to Production AI Workloads with Red Hat AI 3 - Middle East News 247













       To accelerate agent creation and deployment, Red Hat has introduced a Unified API layer based on Llama Stack, which helps align development with industry standards like OpenAI-compatible LLM interface protocols.
       Additionally, to champion a more open and interoperable ecosystem, Red Hat is an early adopter of the Model Context Protocol (MCP), a powerful, emerging standard that streamlines how AI models interact with
       external tools—a fundamental feature for modern AI agents.
       Red Hat AI 3 introduces a new modular and extensible toolkit for model customization, built on existing InstructLab functionality. It provides specialized Python libraries that give developers greater flexibility and control.
       The toolkit is powered by open source projects like Docling for data processing, which streamlines the ingestion of unstructured documents into an AI-readable format. It also includes a flexible framework for synthetic data
       generation and a training hub for LLM fine tuning. The integrated evaluation hub helps AI engineers monitor and validate results, empowering them to confidently leverage their proprietary data for more accurate and
       relevant AI outcomes.
       Supporting Quotes
       Joe Fernandes, vice president and general manager, AI Business Unit, Red Hat
       “As enterprises scale AI from experimentation to production, they face a new wave of complexity, cost and control challenges. With Red Hat AI 3, we are providing an enterprise-grade, open source platform that minimizes
       these hurdles. By bringing new capabilities like distributed inference with llm-d and a foundation for agentic AI, we are enabling IT teams to more confidently operationalize next-generation AI, on their own terms, across
       any infrastructure.”
















       Dan McNamara, senior vice president and general manager, Server and Enterprise AI, AMD
       “As Red Hat brings distributed AI inference into production, AMD is proud to provide the high-performance foundation behind it. Together, we’ve integrated the efficiency of AMD EPYC™ processors, the scalability of AMD
       Instinct™ GPUs, and the openness of the AMD ROCm™ software stack to help enterprises move beyond experimentation and operationalize next-generation AI — turning performance and scalability into real business
       impact across on-prem, cloud, and edge environments.”
       Mariano Greco, chief executive officer, ARSAT
       “As a provider of connectivity infrastructure for Argentina, ARSAT handles massive volumes of customer interactions and sensitive data. We needed a solution that would move us beyond simple automation to ‘Augmented
       Intelligence’ while delivering absolute data sovereignty for our customers. By building our agentic AI platform on Red Hat OpenShift AI, we went from identifying the need to live production in just 45 days. Red Hat
       OpenShift AI has not only helped us improve our service and reduce the time engineers spend on support issues but also freed them up to focus on innovation and new developments.”
       Rick Villars, group vice president, Worldwide Research, IDC
       “2026 will mark an inflection point as enterprises shift from starting their AI pivot to demanding more measurable and repeatable business outcomes from investments. While initial projects focused on training and testing
       models, the real value – and the real challenge – is to operationalize model-derived insights with efficient, secure and cost-effective inference. This shift requires more modern infrastructure, data, and app deployment
       environments with ready to use production-grade inference capabilities that can handle real-world scale and complexity, especially as agentic AI supercharges inference loads. Companies that succeed in becoming AI-
       fueled businesses will be those who establish a unified platform to orchestrate these ever more sophisticated workloads in hybrid cloud environments, not just in silo domains.”













       Ujval Kapasi, vice president, Engineering AI Frameworks, NVIDIA
       “Scalable, high-performance inference is key to the next wave of generative and agentic AI. With built-in support for accelerated inference with open-source NVIDIA Dynamo and NIXL technologies, Red Hat AI 3 provides a
       unified platform that empowers teams to move swiftly from experimentation to running advanced AI workloads and agents at scale.”








      https://menews247.com/red-hat-brings-distributed-ai-inference-to-production-ai-workloads-with-red-hat-ai-3/   2/2
   127   128   129   130   131   132   133   134   135   136   137