Artificial intelligence

Featured image for vLLM interference article.
Article

5 steps to triage vLLM performance

David Whyte-Gray +3

Learn how to improve the performance of your vLLM deployments with a diagnostic workflow that isolates latency issues and server saturation. Discover the key metrics to monitor and techniques to alleviate memory pressure.

Featured image for agentic AI
Article

Automate AI agents with the Responses API in Llama Stack

Michael Dawson

Learn how the Responses API in Llama Stack automates complex tool calling while maintaining granular control over conversation flow for AI agents. Discover the benefits and implementation details.

camel rag
Article

Boring RAG: When similarity is just a SQL query

Ivo Bek

Learn how to create a baseline RAG system using Apache Camel, PostgreSQL, and pgvector. This implementation demonstrates a 'boring' approach to RAG, making it easy to understand and debug. Discover the anatomy of a RAG pipeline, including indexing, retrieval, and answering.

Red Hat AI
Article

Estimate GPU memory for LLM fine-tuning with Red Hat AI

Mohib Azam

Learn how to estimate memory requirements for your LLM fine-tuning experiments using Red Hat Training Hub's memory_estimator.py API. This guide covers the memory components, adjusting training setups for specific GPU specifications, and using the memory estimator in your code. Streamline your model fine-tuning process with runtime estimates and automated hyperparameter suggestions.

Featured image for Red Hat OpenShift AI.
Article

Serve and benchmark Prithvi models with vLLM on OpenShift

Michele Gazzetti +3

Learn how to deploy and test an Earth and space model inference service on Red Hat AI Inference Server and Red Hat OpenShift AI. This article includes two self-contained activities, one deploying Prithvi using a traditional Deployment object and another serving the model using KServe and observing Knative scaling.

ai-ml
Article

Optimize PyTorch training with the autograd engine

Vishal Goyal

Understand the PyTorch autograd engine internals to debug gradient flows. Learn about computational graphs, saved tensors, and performance optimization techniques.

Featured image for vLLM interference article.
Article

Practical strategies for vLLM performance tuning

Trevor Royer

Optimize vLLM performance with practical tuning tips. Learn how to use GuideLLM for benchmarking, adjust GPU ratios, and maximize KV cache to improve throughput.

Featured image for Red Hat OpenShift AI.
Article

Fine-tune AI pipelines in Red Hat OpenShift AI 3.3

Ana Biazetti +2

Learn how to fine-tune AI pipelines in Red Hat OpenShift AI 3.3. Use Kubeflow Trainer and modular components for reproducible, production-grade model tuning.

ai-ml
Article

Understanding ATen: PyTorch's tensor library

Vishal Goyal

Learn how ATen serves as PyTorch's C++ engine, handling tensor operations across CPU, GPU, and accelerators via a high-performance dispatch system and kernels.

Colorful abstract coding elements displayed over a dark gray background.
Article

The uncomfortable truth about vibe coding

Todd Wardzinski

Learn how vibe coding and spec-driven development are shaping the future of software development. Discover the benefits and challenges of each approach, and how to combine them for sustainable software development.

Linuxc ommands fo rDev Sandbox feature image
Article

Leverage AI for root-cause analysis with MCP servers in VS Code and Cursor

Louis Imershein

Learn how to integrate model context protocol (MCP) servers for Red Hat Enterprise Linux and Red Hat Lightspeed into your IDE for data-driven troubleshooting and proactive analytics. Improve your development workflow with actionable intelligence from natural language queries.

Event

Red Hat at DevNexus 2026

Headed to DevNexus? Visit the Red Hat Developer booth on-site to speak to our expert technologists.

IDE
Article

How to connect OpenShift Lightspeed MCP to your IDE

Diego Alvarez Ponce

Learn how to integrate OpenShift Lightspeed into an IDE using the MCP server to generate configurations and query cluster resources without leaving your IDE.