Artificial intelligence

Featured image for machine learning
Article

From 200 lines to 15: How Helion is rewriting the rules of GPU programming

Sumantro Mukherjee +1

Discover how Helion, a Python embedded domain-specific language, abstracts low-level parallelism details to allow developers to write GPU operations using simple, intuitive PyTorch-like syntax. Automatically generate hundreds or even thousands of Triton variants for optimal performance.

OpenShift Dev Spaces
Article

OpenCode: A model-neutral AI coding assistant for OpenShift Dev Spaces

Rohan Kumar

Discover OpenCode, a model-neutral AI coding assistant that supports over 75 providers, including OpenAI, Anthropic Claude, Google Gemini, and local large language models (LLMs) via Ollama. Switch models on demand, compare outputs, avoid vendor lock-in, and even run fully offline with local models. Learn how to set up your environment in Red Hat OpenShift Dev Spaces.

A stylized illustration representing an artificial neural network, set against a dark purple background within a slightly rounded, darker purple square icon shape. The neural network consists of multiple layers of interconnected nodes, depicted as glossy, spherical red orbs. Lines connect these red orbs, forming a complex web. White arrow shapes extend horizontally from the left side, pointing towards the network, suggesting input or data flowing into the system.
Article

Combining KServe and llm-d for optimized generative AI inference

Ran Pollak +1

Learn how to combine KServe and llm-d to optimize generative AI inference, improve performance, and reduce infrastructure costs. This article demonstrates the integration architecture and provides practical guidance for AI platform teams.

Featured image for vLLM interference article.
Article

AI-powered documentation updates: From code diff to docs PR in one comment

Carmel Soceanu

Learn how to automate documentation updates for code changes using Code-to-Docs, an open source GitHub Action. This tool uses AI to analyze your code changes, identify affected documentation files, and generate updated content. Get started with this guide.

Featured image for agentic AI
Article

3 lessons for building reliable ServiceNow AI integrations

Tomer Golan

Learn about critical lessons from building an MCP-powered AI agent for ServiceNow, including how to structure testing environments, best practices for implementing safeguards, and a phased approach to deploying enterprise AI integrations.

Kubernetes + OpenShift featured image
Article

Red Hat build of Kueue 1.3: Enhanced batch workload management on Kubernetes

Kevin Hannon

Explore new features in Red Hat build of Kueue 1.3, including integration with JobSet for efficient batch job scheduling, support for LeaderWorkerSet for distributed ML workloads, and the introduction of v1beta2 APIs. Learn how to get started with the updated Kueue operator.

Featured image for vLLM interference article.
Article

Performance improvements with speculative decoding in vLLM for gpt-oss

Harshith Umesh

Learn how speculative decoding in vLLM can significantly increase throughput without altering a model's output quality, resulting in 19% cost savings at scale for enterprise AI. This post benchmarks gpt-oss-120B with Eagle3 speculative decoding on vLLM and demonstrates consistent throughput and latency improvements across varying concurrency levels, datasets, tensor-parallelism settings, and draft-token budgets.

RHEL CentOS image
Article

Building Red Hat MCP-ready images with image mode for Red Hat Enterprise Linux

Louis Imershein

Learn how to leverage the Model Context Protocol (MCP) to connect VS Code or Cursor to two specialized intelligence streams: one for local system telemetry and one for global proactive security. Discover the benefits of using MCP servers for Red Hat Enterprise Linux and Red Hat Lightspeed for Red Hat Enterprise Linux to diagnose issues with your image mode for Red Hat Enterprise Linux servers.

Featured image for agentic AI
Article

Deploying agents with Red Hat AI: The curious case of OpenClaw

Nati Fridman +2

Explore how Red Hat AI simplifies agent deployment with OpenClaw, showcasing model inference, safety guardrails, agent identity, and persistent state. Learn about vLLM, Llama Stack, and Models-as-a-Service (MaaS) options, and discover the benefits of agent identity and zero trust with Kagenti and AuthBridge.

Featured image for: micropipenv: Installing Python dependencies in containerized applications.
Article

Build more secure, optimized AI supply chains with Fromager

Lalatendu Mohanty

Learn how Fromager, an open source project, helps protect Python dependencies by rebuilding entire dependency trees from source, providing network-isolated builds, and managing dependencies as a verifiable map. Discover how Fromager ensures supply chain verifiability, ABI compatibility, and customization.

Video Thumbnail
Video

Deploying open source AI agents on OpenShift using OpenClaw

Grace Ableidinger +1

Learn how to run OpenClaw on Red Hat OpenShift with production-grade security and observability. We cover default-deny network policies for blast radius containment, container-level sandboxing with OpenShift, Kubernetes Secrets for credential management, and end-to-end OpenTelemetry tracing with MLflow, so every decision your AI agent makes is isolated, auditable, and safe by default. Whether you're a developer exploring AI agents for the first time or a platform engineer thinking about running agentic workloads at scale, this is the infrastructure story that makes it production-ready.

Red Hat AI
Article

Build resilient guardrails for OpenClaw AI agents on Kubernetes

Cedric Clyburn +2

Learn how to build security hygiene into OpenClaw by using containers for isolation, role-based access control (RBAC) for user access permissions, and secrets for sensitive information. This article explores how to use infrastructure powered by open source technology to help protect these workflows.

ai-ml
Article

Manage AI context with the Lola package manager

Daniele Martinoli +2

Learn how to use Lola, a unified package manager for AI context. Treat your AI context as versioned, auditable code with Lola modules and marketplaces. Improve your AI assistant workflow with this open source tool.

Featured image for agentic AI
Article

Distributed tracing for agentic workflows with OpenTelemetry

Fabio Massimo Ercoli

Learn how to set up distributed tracing for an agentic workflow based on lessons learned while developing the it-self-service-agent AI quickstart. This post covers configuring OpenTelemetry to track requests end-to-end across application workloads, MCP servers, and Llama Stack.

Featured image for vLLM interference article.
Article

Run Gemma 4 with Red Hat AI on Day 0: A step-by-step guide

Saša Zelenović +4

Learn how to deploy and experiment with Gemma 4, the latest open model family from Google DeepMind. This guide covers text, image, and video input, Mixture-of-Experts architecture, and more. Get started with Red Hat AI Inference Server today.

Featured image for Red Hat Enterprise Linux.
Article

How to plan your RHEL lifecycle with AI

Samiksha Saxena +1

Discover how the Model Context Protocol server for Red Hat Lightspeed transforms the manual process of managing a RHEL fleet lifecycle into an AI-driven strategy.