Skip to content

Latest commit

 

History

History
129 lines (78 loc) · 5.24 KB

File metadata and controls

129 lines (78 loc) · 5.24 KB

This repo intends to host community maintained common recipes to run vLLM answering the question: How do I run model X on hardware Y for task Z?

Guides

DeepSeek DeepSeek

Ernie Ernie

GLM GLM

Google Google

inclusionAI inclusionAI

InternVL InternVL

InternLM InternLM

Jina AI Jina AI

Llama

MiniMax minmax

Xiaomi MiMo Xiaomi MiMo

Mistral AI Mistral AI

Moonshotai Moonshotai

NVIDIA

OpenAI OpenAI

PaddlePaddle PaddlePaddle

Qwen Qwen

Seed Seed

StepFun StepFun

Tencent-Hunyuan Tencent-Hunyuan

Contributing

Please feel free to contribute by adding a new recipe or improving an existing one, just send us a PR!

While the repo is designed to be directly viewable in GitHub (Markdown files as first citizen), you can build the docs as web pages locally.

uv venv
source .venv/bin/activate
uv pip install -r requirements.txt
uv run mkdocs serve

License

This project is licensed under the Apache License 2.0 - see the LICENSE file for details.