Skip to content

Latest commit

 

History

History
148 lines (90 loc) · 6.11 KB

File metadata and controls

148 lines (90 loc) · 6.11 KB

This repo intends to host community maintained common recipes to run vLLM answering the question: How do I run model X on hardware Y for task Z?

Guides

Arcee AI Arcee AI

DeepSeek DeepSeek

Ernie Ernie

GLM GLM

Google Google

inclusionAI inclusionAI

InternVL InternVL

InternLM InternLM

Jina AI Jina AI

Llama

Microsoft Microsoft

MiniMax minmax

Xiaomi MiMo Xiaomi MiMo

Mistral AI Mistral AI

Moonshotai Moonshotai

NVIDIA

OpenAI OpenAI

PaddlePaddle PaddlePaddle

Qwen Qwen

Seed Seed

StepFun StepFun

Tencent-Hunyuan Tencent-Hunyuan

Contributing

New recipes live as structured YAML at models/<hf_org>/<hf_repo>.yaml and render on recipes.vllm.ai. See CONTRIBUTING.md for the full schema, VRAM formula, and validation steps.

Quick loop:

pnpm install
pnpm dev                              # http://localhost:3000
node scripts/build-recipes-api.mjs    # validates every YAML + rebuilds the JSON API

Legacy MkDocs guides

The top-level Markdown directories (DeepSeek/, Qwen/, etc.) are the historical MkDocs site, kept as a reference during the YAML migration. To preview them:

uv venv
source .venv/bin/activate
uv pip install -r requirements.txt
uv run mkdocs serve --dev-addr 127.0.0.1:8001

License

This project is licensed under the Apache License 2.0 - see the LICENSE file for details.