Red Hat AI

Accelerate the development and deployment of enterprise AI solutions across the hybrid cloud.

red-hat-ai-reverse-logo-transparent-background

Deliver AI solutions with Red Hat AI

Red Hat AI provides the flexibility and consistency you need to deploy and manage predictive and generative AI models for your organization’s workload strategy. The Red Hat AI portfolio includes Red Hat Enterprise Linux AI for individual Linux server environments and Red Hat OpenShift AI for distributed Kubernetes platform environments.

Red Hat AI provides access to small, fit-for-purpose models based on the open source Granite model family that are efficient, cost-effective, and fully supported by Red Hat. It also enables simple yet powerful model tuning with InstructLab, making it easy to align and customize models with your organization’s private data.

Red Hat AI Standard Large icon

Red Hat Enterprise Linux AI

Red Hat Enterprise Linux AI is a foundation model platform that helps simplify and accelerate generative AI model development, testing, and deployment within enterprise environments.

Features and benefits include:

  • IBM’s Granite family large language models (LLMs)
  • Local model fine-tuning with InstructLab
  • Cost-efficient GPU access

Explore Red Hat Enterprise Linux AI

Red Hat OpenShift AI

Red Hat OpenShift AI is an MLOps platform that lets you quickly build, train, and deploy AI models and applications across hybrid cloud environments.

Features and benefits include:

  • Enterprise MLOps capabilities
  • IBM Granite LLMs and InstructLab tooling
  • Hardware accelerators and hybrid cloud support for building and delivering AI at scale

Explore Red Hat OpenShift AI

Introducing Red Hat AI Inference Server

Deploy your preferred models faster and more cost-effectively across the hybrid cloud with Red Hat AI Inference Server. Its vLLM runtime maximizes inference throughput and minimizes latency. A pre-optimized model repository ensures rapid model serving, while the LLM compressor reduces compute costs without sacrificing accuracy. Experience fast, accurate inference for a wide range of applications.

Red Hat AI Inference Server is included in Red Hat OpenShift AI and Red Hat Enterprise Linux AI and supported on Red Hat OpenShift and Red Hat Enterprise Linux.

Bearded person working at a computer

Red Hat AI use cases

Discover what’s possible with Red Hat AI.

Build, migrate, and run machine learning and predictive AI models

Build machine learning models on your own and leverage advanced AI tooling for delivering predictive models. Red Hat AI also provides support for ITOps teams that want to manage and run models from a Kubernetes-based platform.

tooling_icon

Build, deliver, and run generative AI applications

Get access to Granite models, InstructLab, and development tools for delivering generative AI applications.

Gain access to trusted content and security documentation

Safeguard your data with private AI

Red Hat AI provides support for both predictive and generative AI model development and delivery, whether in on-premise data centers or in your own private cloud. Red Hat AI helps reduce the risk of exposing sensitive data by providing support for on-premise and air-gapped deployments.

Private Cloud icon

Operationalize your AI models

Accelerate your move from experimentation to production with the tools you need to automate the model life cycle. Red Hat AI streamlines model training, validation, storing, and serving by combining MLOps and DevOps capabilities in a single platform.

Multi-cluster operations

Multi-architecture AI deployments

Red Hat AI provides support for multi-cloud, hybrid cloud, and hardware acceleration architectures to ensure high-performance stability and scalability across various infrastructures.

_hybrid-cloud_architecture_icon

Red Hat AI: Powered by open source

Red Hat’s product development is rooted in open source and community innovation. Explore the upstream communities that build Red Hat AI.

vllm-logo

vLLM

vLLM, which stands for virtual large language model, is a library of open source code that helps LLMs perform calculations more efficiently and at scale. Specifically, vLLM is an inference server that speeds up the output of gen AI applications by making better use of GPU memory.

Read about vLLM inference

instructlab logo

InstructLab

InstructLab is an open source project for enhancing LLMs used in gen AI applications. Created by IBM and Red Hat, InstructLab provides a cost-effective solution for improving LLM alignment  The project enables anyone to contribute, even those with minimal machine learning experience.

Read about InstructLab and gen AI

granite logo

Granite

Granite is IBM’s third generation of AI language models. Fit for purpose and open source, these enterprise-ready, multimodal models deliver exceptional performance against safety benchmarks and across a wide range of enterprise tasks, from cybersecurity to retrieval-augmented generation (RAG).

Read about Granite models

Open Data Hub and Kubeflow installation customization

Open Data Hub

Red Hat OpenShift AI is based on the upstream project Open Data Hub, which is a blueprint for building an AI-as-a-Service platform on Red Hat's Kubernetes-based OpenShift Container Platform. Open Data Hub is a meta-project that integrates over 20 open source AI/ML projects into a practical solution.

Contribute to Open Data Hub

Jupyter logomark

Jupyter

Project Jupyter, which spun off from the IPython Project in 2014, supports interactive data science and scientific computing across all programming languages. Jupyter is supported by a community of data enthusiasts who believe in the power of open tools and standards for education, research, and data analytics.

Watch a Jupyter notebook demo

Tensorflow logomark

TensorFlow

TensorFlow is an end-to-end, open source platform for machine learning (ML). Its comprehensive, flexible ecosystem of tools, libraries, and community resources helps developers easily build and deploy ML-powered applications.

Learn about TensorFlow and Quarkus

Pytorch logomark

PyTorch

PyTorch is an open source machine learning framework that fast-tracks the path from research prototyping to production deployment. It is used for applications such as computer vision and natural language processing.

Build, train, and run a PyTorch model

SciKit logomark

scikit-learn

scikit-learn is a machine learning library for Python. Built on NumPy, SciPy, and Matplotlib, it offers simple and efficient tools for predictive data analysis.

Explore ML with scikit-learn

Kubeflow

Kubeflow

Kubeflow is an open source framework aimed at simplifying AI/ML workflow deployment at scale. Red Hat OpenShift AI integrates the Kubeflow notebook controller, model serving, and data science pipeline components into the core product.

Fine-tune LLMs with Kubeflow

Featured Red Hat AI blogs & articles

Article Featured image for InstructLab.

Introducing InstructLab, an open source project for enhancing large language...

Article Featured image for Red Hat OpenShift AI.

Learn how to build a ModelCar container image and deploy it with OpenShift AI.

Article Featured image for Red Hat OpenShift AI.
May 01, 2024
Diego Alvarez Ponce +1

Learn how to install the Red Hat OpenShift AI operator and its components in...

Article Featured image for Red Hat OpenShift AI.

Learn how to integrate NVIDIA NIM with OpenShift AI to build, deploy, and...

Article Featured image for Meta vLLM

Discover the new Llama 4 Scout and Llama 4 Maverick models from Meta, with...

Article Featured blog image with the following text: vLLM and DeepSeek
Mar 19, 2025
Michael Goin +4

Explore inference performance improvements that help vLLM serve DeepSeek AI...

Article Featured image showing scaffolding that forms the word "V1".

Explore how vLLM's new multimodal AI inference capabilities enhance...

Blog Featured image for InstructLab.

Learn how to fine-tune large language models with specific skills and...

Ready to use Red Hat AI in production?

Take your deployment to the next level. Transitioning to production with Red Hat AI offers you enhanced stability, security, and support. Our dedicated team is here to ensure a smooth migration and to help with any questions you may have.

Core dimensions of developer productivity