Skip to content

vllm-project/recipes

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

282 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This repo intends to host community maintained common recipes to run vLLM answering the question: How do I run model X on hardware Y for task Z?

Guides

Arcee AI Arcee AI

DeepSeek DeepSeek

Ernie Ernie

GLM GLM

Google Google

inclusionAI inclusionAI

InternVL InternVL

InternLM InternLM

Jina AI Jina AI

Llama

Microsoft Microsoft

MiniMax minmax

Xiaomi MiMo Xiaomi MiMo

Mistral AI Mistral AI

Moonshotai Moonshotai

NVIDIA

OpenAI OpenAI

PaddlePaddle PaddlePaddle

Qwen Qwen

Seed Seed

StepFun StepFun

Tencent-Hunyuan Tencent-Hunyuan

Contributing

New recipes live as structured YAML at models/<hf_org>/<hf_repo>.yaml and render on recipes.vllm.ai. See CONTRIBUTING.md for the full schema, VRAM formula, and validation steps.

Quick loop:

pnpm install
pnpm dev                              # http://localhost:3000
node scripts/build-recipes-api.mjs    # validates every YAML + rebuilds the JSON API

Legacy MkDocs guides

The top-level Markdown directories (DeepSeek/, Qwen/, etc.) are the historical MkDocs site, kept as a reference during the YAML migration. To preview them:

uv venv
source .venv/bin/activate
uv pip install -r requirements.txt
uv run mkdocs serve --dev-addr 127.0.0.1:8001

License

This project is licensed under the Apache License 2.0 - see the LICENSE file for details.

About

Common recipes to run vLLM

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors