vLLM Recipes

This repo intends to host community maintained common recipes to run vLLM answering the question: How do I run model X on hardware Y for task Z?

Guides

DeepSeek

DeepSeek

Ernie

Ernie

GLM

GLM

inclusionAI

inclusionAI

InternVL

InternVL

InternLM

InternLM

Jina AI

Jina AI

Llama

MiniMax

minmax

Mistral AI

Mistral AI

Moonshotai

Moonshotai

OpenAI

OpenAI

PaddlePaddle

PaddlePaddle

Qwen

Qwen

Seed

Seed

Tencent-Hunyuan

Tencent-Hunyuan

Contributing

Please feel free to contribute by adding a new recipe or improving an existing one, just send us a PR!

While the repo is designed to be directly viewable in GitHub (Markdown files as first citizen), you can build the docs as web pages locally.

  1. uv venv
  2. source .venv/bin/activate
  3. uv pip install -r requirements.txt
  4. uv run mkdocs serve

License

This project is licensed under the Apache License 2.0 - see the LICENSE file for details.