Skip to main content
Redhat Developers  Logo
  • Products

    Featured

    • Red Hat Enterprise Linux
      Red Hat Enterprise Linux Icon
    • Red Hat OpenShift AI
      Red Hat OpenShift AI
    • Red Hat Enterprise Linux AI
      Linux icon inside of a brain
    • Image mode for Red Hat Enterprise Linux
      RHEL image mode
    • Red Hat OpenShift
      Openshift icon
    • Red Hat Ansible Automation Platform
      Ansible icon
    • Red Hat Developer Hub
      Developer Hub
    • View All Red Hat Products
    • Linux

      • Red Hat Enterprise Linux
      • Image mode for Red Hat Enterprise Linux
      • Red Hat Universal Base Images (UBI)
    • Java runtimes & frameworks

      • JBoss Enterprise Application Platform
      • Red Hat build of OpenJDK
    • Kubernetes

      • Red Hat OpenShift
      • Microsoft Azure Red Hat OpenShift
      • Red Hat OpenShift Virtualization
      • Red Hat OpenShift Lightspeed
    • Integration & App Connectivity

      • Red Hat Build of Apache Camel
      • Red Hat Service Interconnect
      • Red Hat Connectivity Link
    • AI/ML

      • Red Hat OpenShift AI
      • Red Hat Enterprise Linux AI
    • Automation

      • Red Hat Ansible Automation Platform
      • Red Hat Ansible Lightspeed
    • Developer tools

      • Red Hat Trusted Software Supply Chain
      • Podman Desktop
      • Red Hat OpenShift Dev Spaces
    • Developer Sandbox

      Developer Sandbox
      Try Red Hat products and technologies without setup or configuration fees for 30 days with this shared Openshift and Kubernetes cluster.
    • Try at no cost
  • Technologies

    Featured

    • AI/ML
      AI/ML Icon
    • Linux
      Linux Icon
    • Kubernetes
      Cloud icon
    • Automation
      Automation Icon showing arrows moving in a circle around a gear
    • View All Technologies
    • Programming Languages & Frameworks

      • Java
      • Python
      • JavaScript
    • System Design & Architecture

      • Red Hat architecture and design patterns
      • Microservices
      • Event-Driven Architecture
      • Databases
    • Developer Productivity

      • Developer productivity
      • Developer Tools
      • GitOps
    • Secure Development & Architectures

      • Security
      • Secure coding
    • Platform Engineering

      • DevOps
      • DevSecOps
      • Ansible automation for applications and services
    • Automated Data Processing

      • AI/ML
      • Data Science
      • Apache Kafka on Kubernetes
      • View All Technologies
    • Start exploring in the Developer Sandbox for free

      sandbox graphic
      Try Red Hat's products and technologies without setup or configuration.
    • Try at no cost
  • Learn

    Featured

    • Kubernetes & Cloud Native
      Openshift icon
    • Linux
      Rhel icon
    • Automation
      Ansible cloud icon
    • Java
      Java icon
    • AI/ML
      AI/ML Icon
    • View All Learning Resources

    E-Books

    • GitOps Cookbook
    • Podman in Action
    • Kubernetes Operators
    • The Path to GitOps
    • View All E-books

    Cheat Sheets

    • Linux Commands
    • Bash Commands
    • Git
    • systemd Commands
    • View All Cheat Sheets

    Documentation

    • API Catalog
    • Product Documentation
    • Legacy Documentation
    • Red Hat Learning

      Learning image
      Boost your technical skills to expert-level with the help of interactive lessons offered by various Red Hat Learning programs.
    • Explore Red Hat Learning
  • Developer Sandbox

    Developer Sandbox

    • Access Red Hat’s products and technologies without setup or configuration, and start developing quicker than ever before with our new, no-cost sandbox environments.
    • Explore Developer Sandbox

    Featured Developer Sandbox activities

    • Get started with your Developer Sandbox
    • OpenShift virtualization and application modernization using the Developer Sandbox
    • Explore all Developer Sandbox activities

    Ready to start developing apps?

    • Try at no cost
  • Blog
  • Events
  • Videos

Unleashing multimodal magic with RamaLama

Introducing RamaLama's new multimodal feature

June 20, 2025
Eric Curtin Daniel Walsh Stef Walter Kevin Pouget
Related topics:
Artificial intelligenceContainersOpen source
Related products:
Red Hat AI

Share:

    The world of AI is rapidly evolving, and with it, the need for flexible, powerful, and easily deployable models. At Red Hat, we're always looking for ways to empower developers to build the next generation of intelligent applications. That's why we're thrilled to highlight RamaLama's new multimodal feature, bringing cutting-edge vision-language models (VLMs) directly to your fingertips, seamlessly integrated with the power of containers.

    Beyond text: Embracing the multimodal revolution

    While large language models (LLMs) have taken the world by storm with their text generation capabilities, the real power of AI lies in its ability to understand and interact with the world in a more holistic way. This is where multimodal models come in, bridging the gap between different data types–think images, audio, and text–to create a richer, more nuanced understanding.

    Multimodal

    Multimodal models bridge the gap between different data types, such as images, audio, and text, allowing AI to process and generate information across these diverse modalities. Unlike traditional LLMs that primarily focus on text-in and text-out, multimodal models can, for example, take an image as input and generate a descriptive text, or process spoken language to control a visual output. This capability enables a richer, more nuanced understanding and interaction with the world.

    RamaLama now allows you to easily download and serve multimodal models, opening up a world of possibilities for applications that can see, understand, and respond to visual information alongside text.

    Getting started: Serving your VLM with RamaLama

    The process is incredibly straightforward. With RamaLama, you can get a multimodal model up and running with a single command:

    ramalama serve smolvlm

    This command handles everything from downloading the smolvlm model to setting up the necessary infrastructure to serve it. Behind the scenes, RamaLama leverages the power of containers to ensure a consistent and isolated environment for your model.

    Connecting your web application: A camera demo

    Once your smolvlm model is served, you can easily connect to it using an application. Imagine building an interactive application that can analyze images from a user's camera in real time and provide intelligent responses. RamaLama makes this a reality.

    You can explore a practical example of this in action with the camera-demo.html in the RamaLama repository. This demo showcases how a simple web page can send image data to your running smolvlm instance and receive insights back, all thanks to the robust back end provided by RamaLama. See Figure 1.

    Animation of a web page shows a person in a blue and white shirt putting on Yoda ears. Below the camera are three fields: Base API (set to localhost:8080), Instruction (Who is Stef dressed up as today?), and Response, which provides a changing set of responses such as "Stef is dressed up as a bunny today."
    Figure 1: Demo showing an interactive application that analyzes images from a user's camera in real time.

    The containerization magic: How RamaLama elevates llama-server

    One of RamaLama’s core strengths lies in its intelligent containerization of llama-server. By default, RamaLama packages llama-server within a container, providing several key benefits:

    • Portability: Your llama-server instance, along with all its dependencies, is self-contained. This means you can run it consistently across different environments, from your local development machine to a production server, without worrying about dependency conflicts.
    • Isolation: The containerized environment ensures that llama-server operates in its own isolated space, preventing interference with other applications on your system.
    • Scalability: With containerization, scaling your llama-server instances becomes much simpler, allowing you to handle increased demand by spinning up more containers as needed.
    • Simplified deployment: RamaLama handles the intricacies of setting up and configuring llama-server within a container, significantly reducing the complexity of deployment for developers.

    Acknowledging the foundations: llama.cpp

    It's crucial to acknowledge the foundational work that makes such powerful multimodal capabilities possible. The underlying technology often relies on community efforts. In this case, much credit goes to the impressive llama.cpp project, which has been instrumental in bringing these models to a wider audience with its efficient and flexible implementation.

    Furthermore, we extend our sincere gratitude to Xuan-Son Nguyen (Hugging Face) and the llama.cpp community for their invaluable contributions and dedicated efforts within the llama.cpp ecosystem. His work, and the work of many others in the open source community, are what truly drive innovation and empower developers to build incredible things.

    Join the multimodal journey!

    RamaLama's multimodal feature, powered by containerized llama-server and built upon the excellent work of projects like llama.cpp, represents a significant step forward for developers looking to integrate advanced AI capabilities into their applications. We encourage you to explore RamaLama, experiment with the smolvlm model, and start building the next generation of intelligent, multimodal experiences.

    Head over to RamaLama to learn more and get started today! We can't wait to see what you'll create.

    Related Posts

    • How RamaLama makes working with AI models boring

    • Introducing Podman AI Lab: Developer tooling for working with LLMs

    • Simplify AI data integration with RamaLama and RAG

    • Podman AI Lab and RamaLama unite for easier local AI

    • How RamaLama runs AI models in isolation by default

    • Getting started with Podman AI Lab

    Recent Posts

    • Create and enrich ServiceNow ITSM tickets with Ansible Automation Platform

    • Expand Model-as-a-Service for secure enterprise AI

    • OpenShift LACP bonding performance expectations

    • Build container images in CI/CD with Tekton and Buildpacks

    • How to deploy OpenShift AI & Service Mesh 3 on one cluster

    What’s up next?

    Learn how to access a large language model using Node.js and LangChain.js. You’ll also explore LangChain.js APIs that simplify common requirements like retrieval-augmented generation (RAG).

    Start the activity
    Red Hat Developers logo LinkedIn YouTube Twitter Facebook

    Products

    • Red Hat Enterprise Linux
    • Red Hat OpenShift
    • Red Hat Ansible Automation Platform

    Build

    • Developer Sandbox
    • Developer Tools
    • Interactive Tutorials
    • API Catalog

    Quicklinks

    • Learning Resources
    • E-books
    • Cheat Sheets
    • Blog
    • Events
    • Newsletter

    Communicate

    • About us
    • Contact sales
    • Find a partner
    • Report a website issue
    • Site Status Dashboard
    • Report a security problem

    RED HAT DEVELOPER

    Build here. Go anywhere.

    We serve the builders. The problem solvers who create careers with code.

    Join us if you’re a developer, software engineer, web designer, front-end designer, UX designer, computer scientist, architect, tester, product manager, project manager or team lead.

    Sign me up

    Red Hat legal and privacy links

    • About Red Hat
    • Jobs
    • Events
    • Locations
    • Contact Red Hat
    • Red Hat Blog
    • Inclusion at Red Hat
    • Cool Stuff Store
    • Red Hat Summit
    © 2025 Red Hat

    Red Hat legal and privacy links

    • Privacy statement
    • Terms of use
    • All policies and guidelines
    • Digital accessibility

    Report a website issue