Skip to main content
Redhat Developers  Logo
  • AI

    Get started with AI

    • Red Hat AI
      Accelerate the development and deployment of enterprise AI solutions.
    • AI learning hub
      Explore learning materials and tools, organized by task.
    • AI interactive demos
      Click through scenarios with Red Hat AI, including training LLMs and more.
    • AI/ML learning paths
      Expand your OpenShift AI knowledge using these learning resources.
    • AI quickstarts
      Focused AI use cases designed for fast deployment on Red Hat AI platforms.
    • No-cost AI training
      Foundational Red Hat AI training.

    Featured resources

    • OpenShift AI learning
    • Open source AI for developers
    • AI product application development
    • Open source-powered AI/ML for hybrid cloud
    • AI and Node.js cheat sheet

    Red Hat AI Factory with NVIDIA

    • Red Hat AI Factory with NVIDIA is a co-engineered, enterprise-grade AI solution for building, deploying, and managing AI at scale across hybrid cloud environments.
    • Explore the solution
  • Learn

    Self-guided

    • Documentation
      Find answers, get step-by-step guidance, and learn how to use Red Hat products.
    • Learning paths
      Explore curated walkthroughs for common development tasks.
    • Guided learning
      Receive custom learning paths powered by our AI assistant.
    • See all learning

    Hands-on

    • Developer Sandbox
      Spin up Red Hat's products and technologies without setup or configuration.
    • Interactive labs
      Learn by doing in these hands-on, browser-based experiences.
    • Interactive demos
      Click through product features in these guided tours.

    Browse by topic

    • AI/ML
    • Automation
    • Java
    • Kubernetes
    • Linux
    • See all topics

    Training & certifications

    • Courses and exams
    • Certifications
    • Skills assessments
    • Red Hat Academy
    • Learning subscription
    • Explore training
  • Build

    Get started

    • Red Hat build of Podman Desktop
      A downloadable, local development hub to experiment with our products and builds.
    • Developer Sandbox
      Spin up Red Hat's products and technologies without setup or configuration.

    Download products

    • Access product downloads to start building and testing right away.
    • Red Hat Enterprise Linux
    • Red Hat AI
    • Red Hat OpenShift
    • Red Hat Ansible Automation Platform
    • See all products

    Featured

    • Red Hat build of OpenJDK
    • Red Hat JBoss Enterprise Application Platform
    • Red Hat OpenShift Dev Spaces
    • Red Hat Developer Toolset

    References

    • E-books
    • Documentation
    • Cheat sheets
    • Architecture center
  • Community

    Get involved

    • Events
    • Live AI events
    • Red Hat Summit
    • Red Hat Accelerators
    • Community discussions

    Follow along

    • Articles & blogs
    • Developer newsletter
    • Videos
    • Github

    Get help

    • Customer service
    • Customer support
    • Regional contacts
    • Find a partner

    Join the Red Hat Developer program

    • Download Red Hat products and project builds, access support documentation, learning content, and more.
    • Explore the benefits

Introducing Podman AI Lab: Developer tooling for working with LLMs

An open source extension for Podman Desktop to work with large language models (LLMs) on a local environment

May 7, 2024
Stevan Le Meur Jeff Maury
Related topics:
Artificial intelligenceContainersDeveloper toolsKubernetes
Related products:
Podman DesktopRed Hat Enterprise Linux

    The rapid proliferation of generative AI (gen AI) and open source large language models (LLMs) is revolutionizing computing, and AI-enabled applications are now becoming the norm. As a result, application developers now have to learn best practices and how to build applications taking advantage of AI in their solutions. Hence, the need for tools and technologies that make AI more accessible and approachable, enabling developers to tap into its vast potential. 

    However, this shift also underscores the importance of ensuring data privacy and security, as sensitive information is being processed and stored in these new AI-powered systems. To address this growing demand, it is essential to develop and deploy solutions that not only simplify AI development, but also safeguard user data, fostering trust and innovation in this emerging era of AI-driven computing.

    Last year, we announced the general availability of Podman Desktop 1.0, which provides a user-friendly interface for managing containers and working with Kubernetes on your desktop. Today, we are excited to announce Podman AI Lab, a dedicated extension for Podman Desktop that gives developers the ability to build, test and run gen AI-powered applications using an intuitive graphical interface running on their local workstation (Figure 1).

    ai-lab-hero
    Figure 1: Podman AI Lab.

    From getting started with AI to experimenting with models and prompts, Podman AI Lab enables developers to easily bring AI into their applications without depending on infrastructure beyond a laptop. Let’s explore what Podman AI is and why it can be advantageous for the future of your applications.

    Why use Podman AI Lab?

    As developers, we are always using a heterogeneous set of tools to get our developer environment up and properly configured. The convenience of our desktop workstations is real, enabling us fast turnarounds to quickly iterate on our applications. In this emerging era of gen AI-infused applications, getting access to LLMs and running them is a necessity. 

    More importantly, our motivation is to democratize the journey to gen AI for application developers. A lot of us believe that AI/ML is primarily in the hands of data scientists. But in fact, with gen AI, AI/ML expertise is not a significant factor for integrating models into existing applications. By simplifying LLMs and gen AI for all developers, we believe everyone can seamlessly enhance existing applications with AI and create new experiences.

    Running LLMs locally for development and debugging

    The main advantage of running LLMs locally is that it provides an easier environment for debugging and developing AI-infused applications. You can quickly test and iterate on your code, you continue to benefit from the convenience of the different tools you are already using, and you can run the LLMs even when offline. All of these reasons make it easier to incorporate LLMs and leverage their capabilities in your applications. 

    Cost efficiency

    When you are experimenting and just playing around with the LLMs, it doesn’t need to light up your cloud billing accounts with GPU charges. Small but efficient models can allow developers to go a long way. While there is an initial investment in hardware and setup, running models on your own machine reduces ongoing costs of cloud computing services and alleviates vendor-locking. 

    Education and training

    Podman Desktop simplifies the onboarding process for developers new to containerization, providing an intuitive interface and streamlined management process that accelerates the adoption of containers and the transition to Kubernetes. Meanwhile, Podman AI Lab's recipe catalog offers sample applications showcasing common use cases for LLMs, serving as a great jump start for developers to learn from and build upon. By reviewing the source code and seeing how these applications are built, developers can gain valuable insights into best practices for wrapping their own code around AI models. 

    Data security and ownership

    By running LLMs locally, you maintain ownership of your data and have full control over how it's processed and stored. This is particularly important when working with sensitive, proprietary information or when you are concerned by regulatory compliance restrictions on how data is processed. This is something you can only achieve in the safe zone of your on-premise deployment and local environment, helping to ensuring your data remains secure and confidential.

    Easier transition to production

    Podman AI Lab also simplifies the transition to production. With everything running on Podman, containers and pods are portable and can easily be run anywhere across the hybrid cloud using Kubernetes, Red Hat OpenShift, and Red Hat OpenShift AI. 

    In addition, with the bootc extension also available for Podman Desktop, it provides a tight integration with image mode for Red Hat Enterprise Linux, a new deployment method for the world’s leading enterprise Linux platform that delivers the operating system as a container image. This integration enables developers to more easily go from prototyping and working with models on their laptop, to then turning the new AI-infused application into a bootable container.

    Key features

    Podman AI Lab includes many features that streamline workflows of working with LLMs to easily get started bringing AI into your applications.

    • Discover generative AI
      • Built-in catalog of recipes (sample applications) to discover and learn best practices
      • Experiment with different compatible models
    • Catalog of curated models
      • Out-of-the-box list of ready-to-use open source models
      • Import your own model files (GGUF)
    • Local model serving
      • OpenAI-compatible API
      • Code snippets generation for instant integration of developers’ applications
      • Pre-built container image, based on UBI and kept up-to-date
    • Playground environments
      • Experiment and test models
      • Configure model settings and parameters

    To learn more about those features, we invite you to read the getting started guide, which gives a complete overview of each of the capabilities provided. 

    How to get started and install the extension

    To install Podman AI Lab, you’ll need Podman Desktop 1.10 or later. You’ll then find the catalog of extensions (Figure 2).

    extension catalog
    Figure 2: Podman AI Lab extensions catalog.

    Alternatively, you can install the extension from the dashboard of Podman Desktop (Figure 3).

    dashboard-ai-panel
    Figure 3: install the extension from the dashboard of Podman Desktop.

    What’s next?

    We plan to improve Podman AI Lab and add more capabilities to help you infuse AI into your own application, enabling fine-tuning with your own data and more easily integrating LLMs into your applications. We’ll soon enable running the models with GPU acceleration as well.

    For developers targeting Red Hat OpenShift AI, you'll see more tools integrated with Podman Desktop and Podman AI Lab to make it smoother and more efficient to build your application locally and transition it to production. 

    More information

    You can get a complete tour of Podman AI Lab by following the getting started guide as well as an deep dive into the Recipes provided by Podman AI Lab.

    Since last year's announcement, Podman Desktop has crossed 1 million total downloads and continues its momentum. We are now expanding to help developers with AI and we are excited about the road ahead. Visit podman-desktop.io to download the tool and learn more about the project.

    • Press release
    • Podman AI Lab on Red Hat Developer
    • Read more about Podman Desktop on Red Hat Developer
    • Visit the Podman website: podman.io
    • Support us by giving a ⭐️! 

    Related Posts

    • What is Podman Desktop? A developer's introduction

    • Image mode for Red Hat Enterprise Linux quick start: AI inference

    • Working with Kubernetes in Podman Desktop

    • How to install and use Podman Desktop on Windows

    • Podman Desktop 1.0: Local container development made easy

    • AI Lab Recipes

    Recent Posts

    • Every layer counts: Defense in depth for AI agents with Red Hat AI

    • Fun in the RUN instruction: Why container builds with distroless images can surprise you

    • Trusted software factory: Building trust in the agentic AI era

    • Build a zero trust AI pipeline with OpenShift and RHEL CVMs

    • Red Hat Hardened Images: Top 5 benefits for software developers

    What’s up next?

    Read Podman in Action for an introduction to Podman’s features and capabilities, including how to work with containers, build container images, and convert containerized applications into either single-node services to run on edge devices or Kubernetes-based microservices.

    Get the e-book
    Red Hat Developers logo LinkedIn YouTube Twitter Facebook

    Platforms

    • Red Hat AI
    • Red Hat Enterprise Linux
    • Red Hat OpenShift
    • Red Hat Ansible Automation Platform
    • See all products

    Build

    • Developer Sandbox
    • Developer tools
    • Interactive tutorials
    • API catalog

    Quicklinks

    • Learning resources
    • E-books
    • Cheat sheets
    • Blog
    • Events
    • Newsletter

    Communicate

    • About us
    • Contact sales
    • Find a partner
    • Report a website issue
    • Site status dashboard
    • Report a security problem

    RED HAT DEVELOPER

    Build here. Go anywhere.

    We serve the builders. The problem solvers who create careers with code.

    Join us if you’re a developer, software engineer, web designer, front-end designer, UX designer, computer scientist, architect, tester, product manager, project manager or team lead.

    Sign me up

    Red Hat legal and privacy links

    • About Red Hat
    • Jobs
    • Events
    • Locations
    • Contact Red Hat
    • Red Hat Blog
    • Inclusion at Red Hat
    • Cool Stuff Store
    • Red Hat Summit
    © 2026 Red Hat

    Red Hat legal and privacy links

    • Privacy statement
    • Terms of use
    • All policies and guidelines
    • Digital accessibility

    Chat Support

    Please log in with your Red Hat account to access chat support.