Skip to content

[Roadmap] vLLM Production Stack 2026 Roadmap #855

@ruizhang0101

Description

@ruizhang0101

Router

Advanced Deployment

Support for Modern Workloads

  • (P0) Support deployment with vLLM Omni
  • (P1) Intelligent routing for agent workloads and multimodal models
  • (P2) End-to-end deployment for multimodal models
  • (P2) First-class support for RL workloads

Safety and Security

  • (P0) Request migration when the vLLM instance fails
  • (P1) Ship router extension config with pre-defined safety policies, rate limits, and logging enabled
  • (P1) Add optional PII detection in the request path

CI/CD and Misc

  • (P0) Release bot for automated releases (Helm chart, K8s controller packages, Docker images)
  • (P2) GitHub Actions for building router Docker images for multiple architectures
  • (P2) Tutorials covering more cloud platforms, models, and features

If any of the items you wanted are not on the roadmap, your suggestion and contribution are strongly welcomed! Please feel free to comment in this thread, open a feature request, or create an RFC.

Happy vLLMing!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions