
TrustyAI Detoxify: Guardrailing LLMs during training
Learn how to prevent large language models (LLMs) from generating toxic content during training using TrustyAI Detoxify and Hugging Face SFTTrainer.
Learn how to prevent large language models (LLMs) from generating toxic content during training using TrustyAI Detoxify and Hugging Face SFTTrainer.
Learn how to deploy and use the Multi-Cloud Object Gateway (MCG) from Red Hat OpenShift Data Foundation to support development and testing of applications and Artificial Intelligence (AI) models which require S3 object storage.
Train and deploy an AI model using OpenShift AI, then integrate it into an application running on OpenShift.
BERT, which stands for Bidirectional Encoder Representations from Transformers
Develop, deploy, and run large language models (LLMs) in individual server environments. The solution includes Red Hat AI Inference Server, delivering fast, cost-effective hybrid cloud inference by maximizing throughput, minimizing latency, and reducing compute costs.
This article explains how to use Red Hat OpenShift AI in the Developer Sandbox for Red Hat OpenShift to create and deploy models.
Explore large language models (LLMs) by trying out the Granite model on Podman AI Lab.
This article demonstrates how to register the SKlearn runtime as a Custom ServingRuntime, deploy the iris model on KServe with OpenDataHub, and apply authentication using Authorino to protect the model endpoints.
Develop, deploy, and run large language models (LLMs) in individual server
Explore the integration of FP8 in vLLM. Learn how to receive up to a 2x reduction in latency on NVIDIA GPUs with minimal accuracy degradation.
Explore how to use OpenVINO Model Server (OVMS) built on Intel's OpenVINO toolkit to streamline the deployment and management of deep learning models.
Event-driven Sentiment Analysis using Kafka, Knative and AI/ML
End-to-end AI-enabled applications and data pipelines across the hybrid cloud
Learn a simplified method for installing KServe, a highly scalable and standards-based model inference platform on Kubernetes for scalable AI.
Learn how to generate complete Ansible Playbooks using natural language prompts and boost automation productivity with Red Hat's new Ansible VS Code extension.
Llama 3's advancements, particularly at 8 billion parameters, make AI more accessible and efficient.
Podman AI Lab provides a containerized environment for exploring, testing, and integrating open source AI models locally using Podman Desktop.
A practical example to deploy machine learning model using data science...
Learn how to fine-tune large language models with specific skills and knowledge
Dive into the end-to-end process of building and managing machine learning (ML)
Are you curious about the power of artificial intelligence (AI) but not sure
This blog post explores the integration of Large Language Models (LLMs) with
The Edge to Core Pipeline Pattern automates a continuous cycle for releasing and deploying new AI/ML models using Red Hat build of Apache Camel and more.
Explore the fundamental concepts of artificial intelligence (AI), including machine learning and deep learning, and learn how to integrate AI into your platforms and applications.
Introducing InstructLab, an open source project for enhancing large language models (LLMs) used in generative AI applications through a community approach.