Red Hat Developer Blog
Here's our most recent blog content. Explore our featured monthly resource as well as our most recently published items. Don't miss the chance to learn more about our contributors.
View all blogs & articles
An update on reproducing R1-like reasoning in small LLMs: Granite models show...
Open-sourced on Hugging Face, deployment-ready with vLLM, and extensible...
Learn about the alpha release of vLLM V1, a major upgrade to vLLM’s core...
Learn how to integrate Model Context Protocol (MCP) with LLMs using Node.js....
Advancing AI efficiency is more critical than ever, and sparsity has proven...
Quantized LLMs achieve near-full accuracy with minimal trade-offs after 500K+...
Machete, Neural Magic’s optimized kernel for NVIDIA Hopper GPUs, achieves...
Discover LLM Compressor, a unified library for creating accurate compressed...
Explore the integration of FP8 in vLLM. Learn how to receive up to a 2x...