Blog

Integrating vLLM with Google ADK: A High-Performance Local LLM Guide

Learn how to leverage vLLM to host high-performance local LLMs and integrate them seamlessly with Google ADK using LiteLLM.

Posted on: 2026-03-06 by AI Assistant


In the world of AI agents, latency and throughput are critical. While cloud-based models are powerful, many developers are turning to local hosting for better control, privacy, and performance. vLLM has emerged as one of the fastest and most memory-efficient serving engines for Large Language Models (LLMs).

In this post, we’ll explore how to integrate vLLM with the Google Agent Development Kit (ADK) to build high-performance agentic applications.

What is vLLM?

vLLM is a high-throughput and memory-efficient serving engine for LLMs. It uses PagedAttention, an algorithm that manages attention key and value memory more efficiently, allowing for significantly higher throughput than traditional libraries like Hugging Face Transformers.

One of the best features of vLLM is its built-in support for the OpenAI API protocol. This means you can host any supported model (like Llama 3, Mistral, or Gemma) and interact with it as if it were an OpenAI endpoint.

Hosting a Model with vLLM

To get started, you can run a vLLM server using Docker or direct Python installation. Here is a basic command to start a server with a model like Mistral:

python -m vllm.entrypoints.openai.api_server \
    --model mistralai/Mistral-7B-Instruct-v0.2 \
    --port 8000

Once the server is running, it will expose an OpenAI-compatible API at http://localhost:8000/v1.

Integrating with Google ADK

Google ADK leverages the LiteLLM library to provide a unified interface for various LLM providers. Since vLLM provides an OpenAI-compatible endpoint, integrating it into your ADK agent is straightforward.

Configuration

You need to configure your agent to use the LiteLlm model class and point it to your local vLLM instance.

from adk import Agent, LiteLlm

# Initialize the model pointing to your vLLM server
model = LiteLlm(
    model_name="openai/mistralai/Mistral-7B-Instruct-v0.2",
    api_base="http://localhost:8000/v1",
    api_key="token-not-needed"  # vLLM usually doesn't require a key unless configured
)

# Create your agent
agent = Agent(
    name="LocalAssistant",
    model=model,
    instructions="You are a helpful assistant running on a local vLLM instance."
)

Key Considerations

  1. Model Naming: When using LiteLLM with vLLM, use the openai/ prefix followed by the exact model name used when starting the vLLM server.
  2. API Base: Ensure the api_base includes the /v1 suffix.
  3. Performance: Hosting locally requires significant GPU memory (VRAM). Ensure your hardware can handle the model size and the KV cache required by vLLM.

Benefits of the vLLM + ADK Stack

Combining vLLM with Google ADK offers several advantages:

Conclusion

Integrating vLLM with Google ADK opens up powerful possibilities for building fast, private, and cost-effective AI agents. By following the steps above, you can bring the latest open-source models into your agentic workflows with minimal friction.

Happy building!