Artificial intelligence

Featured image for vLLM interference article.
Article

Run Gemma 4 with Red Hat AI on Day 0: A step-by-step guide

Saša Zelenović +4

Learn how to deploy and experiment with Gemma 4, the latest open model family from Google DeepMind. This guide covers text, image, and video input, Mixture-of-Experts architecture, and more. Get started with Red Hat AI Inference Server today.

Featured image for Red Hat Enterprise Linux.
Article

How to plan your RHEL lifecycle with AI

Samiksha Saxena +1

Discover how the Model Context Protocol server for Red Hat Lightspeed transforms the manual process of managing a RHEL fleet lifecycle into an AI-driven strategy.

Red Hat AI
Article

Unsloth and Training Hub: Lightning-fast LoRA and QLoRA fine-tuning

Aditi Saluja +2

Learn how to fine-tune large language models in enterprise environments with Training Hub, an open source library for LLM post-training. Discover the benefits of LoRA and QLoRA using Unsloth, including reduced VRAM requirements and faster training times.

ai-ml
Article

Vibes, specs, skills, and agents: The four pillars of AI coding

Rich Naszcyniec

Explore the four pillars of AI coding: vibes, secs, skills, and agents, and learn how they can improve the coding quality and reduce the encoding/decoding gap. Discover the benefits of a spec-driven approach and the importance of modular specs and skills in achieving harmony.

Featured image for vLLM interference article.
Article

Integrate Claude Code with Red Hat AI Inference Server on OpenShift

Alexander Barbosa Ayala

Learn how to integrate Anthropic's Claude Code, an agentic coding tool, using Red Hat AI Inference Server on OpenShift. Keep the inference process private on your own infrastructure while retaining the full Claude Code workflow.

Featured image for Red Hat OpenShift AI.
Article

Run Model-as-a-Service for multiple LLMs on OpenShift

Vladimir Belousov

Learn how to deploy multiple large language models (LLMs) behind a single OpenAI-compatible endpoint on OpenShift using a Model-as-a-Service (MaaS) approach. This guide demonstrates how to build an intelligent routing infrastructure that dynamically inspects the request payload and directs traffic based on the specified model field, reducing GPU waste and simplifying application logic.

Featured image for Red Hat OpenShift AI.
Article

Hybrid loan-decisioning with OpenShift AI and Vertex AI

Harshil Sabhnani

Discover a practical solution pattern for building a modern financial application that makes loan decisions using multiple machine learning systems deployed across hybrid environments.

Event

Red Hat at Devoxx UK 2026

Headed to Devoxx UK 2026? Visit the Red Hat Developer booth on-site to speak to our expert technologists.

Event

Red Hat at Devoxx France 2026

Headed to Devoxx France 2026? Visit the Red Hat Developer booth on-site to speak to our expert technologists.

LLM Compressor v0.10.0 is here
Article

LLM Compressor v0.10: Faster compression with distributed GPTQ

Kyle Sayers +2

LLM Compressor v0.10 introduces Distributed Data Parallel (DDP) for faster compression, memory management, and advanced quantization formats. Make model compression workflows more efficient for large language models.

Red Hat AI
Article

Configure NVIDIA Blackwell GPUs for Red Hat AI workloads

Erwan Gallen +4

Learn how to enable the NVIDIA RTX PRO 4500 Blackwell Server Edition on Red Hat AI for compact, power-efficient AI deployments. This hardware offers inference performance without adding unnecessary operational complexity for Red Hat AI users.

Jupyter Notebooks on Red Hat OpenShift AI share/feature image
Article

Accelerated expert-parallel distributed tuning in Red Hat OpenShift AI

Karel Suta +4

Discover how to optimize training of MoE models with fms-hf-tuning, an open source tuning library for PyTorch FSDP and Hugging Face libraries. Learn about preprocessing data, throughput and memory efficiency features, distributed training, and expert parallelism. Improve your AI and agentic applications on domain-specific enterprise tasks.