Artificial intelligence
The benefits of dynamic GPU slicing in OpenShift
Learn how the dynamic accelerator slicer operator improves GPU resource management in OpenShift by dynamically adjusting allocation based on workload needs.
Node.js function calling with LangGraph.js in Podman AI Lab
Get an introduction to AI function calling using Node.js and the LangGraph.js framework, now available in the Podman AI Lab extension.
Retrieval-augmented generation with Llama Stack and Node.js
This tutorial shows you how to use the Llama Stack API to implement retrieval-augmented generation for an AI application built with Node.js.
How to run performance and scale validation for OpenShift AI
Learn about the Red Hat OpenShift AI model fine-tuning stack and how to run performance and scale validation.
Accelerate model training on OpenShift AI with NVIDIA GPUDirect RDMA
Learn how NVIDIA GPUDirect RDMA over Ethernet enhances distributed model training performance and reduces communication bottlenecks in Red Hat OpenShift AI.
How reinforcement learning improves DeepSeek performance
Learn how the DeepSeek training process used reinforcement learning algorithms to generate human-like text and improve overall performance.
Performance boosts in vLLM 0.8.1: Switching to the V1 engine
Explore performance and usability improvements in vLLM 0.8.1 on OpenShift, including crucial architectural overhauls and multimodal inference optimizations.
Bobbycar, a Red Hat Connected Vehicle Architecture Solution Pattern - Part 1: Automotive Use Cases
This Red Hat solution pattern implements key aspects of a modern IoT/edge architecture in an exemplary manner. It uses Red Hat OpenShift Container Platform and various middleware components optimized for cloud-native use. This enterprise architecture can serve as a foundation for an IoT/edge hybrid cloud environment supporting various use cases like over-the-air (OTA) deployments, driver monitoring, AI/ML, and others. Bobbycar aims to showcase an end-to-end workflow, from connecting in-vehicle components to a cloud back-end, processing telemetry data in batch or as stream, and training AI/ML models, to deploying containers through a DevSecOps pipeline and by leveraging GitOps to the edge.
Processing IoT data and serving AI/ML models with OpenShift Serverless
Explore Knative Serving, Eventing, and Functions through an example use case. You’ll see how to collect telemetry data from simulated vehicles, process the data with OpenShift Serverless, and use the data to train a machine learning model with Red Hat OpenShift AI, Red Hat's MLOps platform. The model will then be deployed as a Knative Service, providing the inference endpoint for our business application.
Red Hat Advanced Developer Suite
A comprehensive offering for developers that includes a range of tools to
Cracking the code: How neural networks might actually “think”
Discover a new combinatorial approach to decoding AI’s hidden logic, exploring how neural networks truly compute and reason."
Fine-tune LLMs with Kubeflow Trainer on OpenShift AI
Discover how to fine-tune large language models (LLMs) with Kubeflow Training, PyTorch FSDP, and Hugging Face SFTTrainer in OpenShift AI.
Camel integration quarterly digest: Q1 2025
Dive into the Q1’25 edition of Camel integration quarterly digest, covering the
Incident detection for OpenShift tech preview is here
The technology preview of incident detection is now available in the Red Hat OpenShift web console monitoring UI plug-in.
How Developer Hub and OpenShift AI work with OpenShift
Explore how Red Hat Developer Hub and OpenShift AI work together with OpenShift to build workbenches and accelerate AI/ML development.
The road to AI: A guide to understanding AI/ML models
This article demystifies AI/ML models by explaining how they transform raw data into actionable business insights.
How building workbenches accelerates AI/ML development
Learn how to build AI applications with OpenShift AI by integrating workbenches in Red Hat Developer Hub for training models (part 1 of 2).
Essential Node.js Observability Resources
A listing of Essential Node.js Observability Posts from Red Hat Developer and
How to build AI-ready applications with Quarkus
Develop AI-integrated Java applications more efficiently using Quarkus. This article covers implementing chatbots, real-time interaction, and RAG functionality.
Llama 4 herd is here with Day 0 inference support in vLLM
Discover the new Llama 4 Scout and Llama 4 Maverick models from Meta, with mixture of experts architecture, early fusion multimodality, and Day 0 model support.
Async-GRPO: Open, fast, and performant
Discover Async-GRPO, a new library for reinforcement learning tasks that efficiently handles large models, eliminates bottlenecks, and accelerates experiments.
Sculpting subspaces: How we solved continual learning in LLMs
Discover how the adaptive SVD approach enables LLMs to continually learn and adapt without forgetting previously acquired knowledge.
Simplify AI data integration with RamaLama and RAG
Explore how RamaLama makes it easier to share data with AI models using retrieval-augmented generation (RAG), a technique for enhancing large language models.
How to navigate LLM model names
Learning the naming conventions of large language models (LLMs) helps users select the right model for their needs.