Exploring Llama Stack with Python: Tool calling and agents
Harness Llama Stack with Python for LLM development. Explore tool calling, agents, and Model Context Protocol (MCP) for versatile integrations.
Harness Llama Stack with Python for LLM development. Explore tool calling, agents, and Model Context Protocol (MCP) for versatile integrations.
Learn how to build a Model-as-a-Service platform with this simple demo. (Part 3 of 4)
Explore the architecture of a Models-as-a-Service (MaaS) platform and how enterprises can create a secure and scalable environment for AI models. (Part 2 of 4)
This article introduces Models-as-a-Service (MaaS) for enterprises, outlining the challenges, benefits, key technologies, and workflows. (Part 1 of 4)
Integrate Red Hat AI Inference Server with LangChain to build agentic document processing workflows. This article presents a use case and Python code.
Enhance your Node.js AI applications with distributed tracing. Discover how to use Jaeger and OpenTelemetry for insights into Llama Stack interactions.
Deploy AI at the edge with Red Hat OpenShift AI. Learn to set up OpenShift AI, configure storage, train models, and serve using KServe's RawDeployment.
In this recording, we demonstrate how to compose model compression experiments, highlighting the benefits of advanced algorithms requiring custom data sets and how evaluation results and model artifacts can be shared with stakeholders.
Podman enables developers to run Linux containers on MacOS within virtual machines, including GPU acceleration for improved AI inference performance.
Explore how to utilize guardrails for safety mechanisms in large language models (LLMs) with Node.js and Llama Stack, focusing on LlamaGuard and PromptGuard.
Members from the Red Hat Node.js team were recently at PowerUp 2025. It was held
Discover how IBM used OpenShift AI to maximize GPU efficiency on its internal AI supercomputer, using open source tools like Kueue for efficient AI workloads.
Gain detailed insights into vLLM deployments on OpenShift AI. Learn to build dashboards with Dynatrace and OpenTelemetry to enable reliable LLM performance.
Learn how to use Red Hat OpenShift AI to quickly develop, train, and deploy
Explore the complete machine learning operations (MLOps) pipeline utilizing Red
Optimize model inference and reduce costs with model compression techniques like quantization and pruning with LLM Compressor on Red Hat OpenShift AI.
Learn how to use synthetic data generation (SDG) and fine-tuning in Red Hat AI to customize reasoning models for your enterprise workflows.
Learn how to deploy a trained model with Red Hat OpenShift AI and use its
Explore how to use large language models (LLMs) with Node.js by observing Ollama
More Essential AI tutorials for Node.js Developers
vLLM empowers macOS and iOS developers to build powerful AI-driven applications by providing a robust and optimized engine for running large language models.
PowerUP 2025 is the week of May 19th. It's held in Anaheim, California this year
Learn how to use pipelines in OpenShift AI to automate the full AI/ML lifecycle on a single-node OpenShift instance.
Jupyter Notebook works with OpenShift AI to interactively classify images. In
Learn how to set up NVIDIA NIM on Red Hat OpenShift AI and how this benefits AI and data science workloads.