Stop using title frontmatter and fix doc that can only be reached by search (#20623)

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
Harry Mellor
2025-07-08 11:27:40 +01:00
committed by GitHub
parent b4bab81660
commit b942c094e3
81 changed files with 82 additions and 238 deletions

View File

@ -55,6 +55,7 @@ nav:
- contributing/model/registration.md
- contributing/model/tests.md
- contributing/model/multimodal.md
- CI: contributing/ci
- Design Documents:
- V0: design
- V1: design/v1

View File

@ -1,5 +1,3 @@
---
title: Contact Us
---
# Contact Us
--8<-- "README.md:contact-us"

View File

@ -1,6 +1,4 @@
---
title: Meetups
---
# Meetups
We host regular meetups in San Francisco Bay Area every 2 months. We will share the project updates from the vLLM team and have guest speakers from the industry to share their experience and insights. Please find the materials of our previous meetups below:

View File

@ -1,6 +1,4 @@
---
title: Engine Arguments
---
# Engine Arguments
Engine arguments control the behavior of the vLLM engine.

View File

@ -1,6 +1,4 @@
---
title: Server Arguments
---
# Server Arguments
The `vllm serve` command is used to launch the OpenAI-compatible server.

View File

@ -1,6 +1,4 @@
---
title: Benchmark Suites
---
# Benchmark Suites
vLLM contains two sets of benchmarks:

View File

@ -1,6 +1,4 @@
---
title: Update PyTorch version on vLLM OSS CI/CD
---
# Update PyTorch version on vLLM OSS CI/CD
vLLM's current policy is to always use the latest PyTorch stable
release in CI/CD. It is standard practice to submit a PR to update the

View File

@ -1,6 +1,4 @@
---
title: Summary
---
# Summary
!!! important
Many decoder language models can now be automatically loaded using the [Transformers backend][transformers-backend] without having to implement them in vLLM. See if `vllm serve <model>` works first!

View File

@ -1,6 +1,4 @@
---
title: Basic Model
---
# Basic Model
This guide walks you through the steps to implement a basic vLLM model.

View File

@ -1,6 +1,4 @@
---
title: Multi-Modal Support
---
# Multi-Modal Support
This document walks you through the steps to extend a basic model so that it accepts [multi-modal inputs](../../features/multimodal_inputs.md).

View File

@ -1,6 +1,4 @@
---
title: Registering a Model
---
# Registering a Model
vLLM relies on a model registry to determine how to run each model.
A list of pre-registered architectures can be found [here](../../models/supported_models.md).

View File

@ -1,6 +1,4 @@
---
title: Unit Testing
---
# Unit Testing
This page explains how to write unit tests to verify the implementation of your model.

View File

@ -1,6 +1,4 @@
---
title: Using Docker
---
# Using Docker
[](){ #deployment-docker-pre-built-image }

View File

@ -1,6 +1,5 @@
---
title: Anyscale
---
# Anyscale
[](){ #deployment-anyscale }
[Anyscale](https://www.anyscale.com) is a managed, multi-cloud platform developed by the creators of Ray.

View File

@ -1,6 +1,4 @@
---
title: Anything LLM
---
# Anything LLM
[Anything LLM](https://github.com/Mintplex-Labs/anything-llm) is a full-stack application that enables you to turn any document, resource, or piece of content into context that any LLM can use as references during chatting.

View File

@ -1,6 +1,4 @@
---
title: AutoGen
---
# AutoGen
[AutoGen](https://github.com/microsoft/autogen) is a framework for creating multi-agent AI applications that can act autonomously or work alongside humans.

View File

@ -1,6 +1,4 @@
---
title: BentoML
---
# BentoML
[BentoML](https://github.com/bentoml/BentoML) allows you to deploy a large language model (LLM) server with vLLM as the backend, which exposes OpenAI-compatible endpoints. You can serve the model locally or containerize it as an OCI-compliant image and deploy it on Kubernetes.

View File

@ -1,6 +1,4 @@
---
title: Cerebrium
---
# Cerebrium
<p align="center">
<img src="https://i.ibb.co/hHcScTT/Screenshot-2024-06-13-at-10-14-54.png" alt="vLLM_plus_cerebrium"/>

View File

@ -1,6 +1,4 @@
---
title: Chatbox
---
# Chatbox
[Chatbox](https://github.com/chatboxai/chatbox) is a desktop client for LLMs, available on Windows, Mac, Linux.

View File

@ -1,6 +1,4 @@
---
title: Dify
---
# Dify
[Dify](https://github.com/langgenius/dify) is an open-source LLM app development platform. Its intuitive interface combines agentic AI workflow, RAG pipeline, agent capabilities, model management, observability features, and more, allowing you to quickly move from prototype to production.

View File

@ -1,6 +1,4 @@
---
title: dstack
---
# dstack
<p align="center">
<img src="https://i.ibb.co/71kx6hW/vllm-dstack.png" alt="vLLM_plus_dstack"/>

View File

@ -1,6 +1,4 @@
---
title: Haystack
---
# Haystack
# Haystack

View File

@ -1,6 +1,4 @@
---
title: Helm
---
# Helm
A Helm chart to deploy vLLM for Kubernetes

View File

@ -1,6 +1,4 @@
---
title: LiteLLM
---
# LiteLLM
[LiteLLM](https://github.com/BerriAI/litellm) call all LLM APIs using the OpenAI format [Bedrock, Huggingface, VertexAI, TogetherAI, Azure, OpenAI, Groq etc.]

View File

@ -1,6 +1,4 @@
---
title: Lobe Chat
---
# Lobe Chat
[Lobe Chat](https://github.com/lobehub/lobe-chat) is an open-source, modern-design ChatGPT/LLMs UI/Framework.

View File

@ -1,6 +1,4 @@
---
title: LWS
---
# LWS
LeaderWorkerSet (LWS) is a Kubernetes API that aims to address common deployment patterns of AI/ML inference workloads.
A major use case is for multi-host/multi-node distributed inference.

View File

@ -1,6 +1,4 @@
---
title: Modal
---
# Modal
vLLM can be run on cloud GPUs with [Modal](https://modal.com), a serverless computing platform designed for fast auto-scaling.

View File

@ -1,6 +1,4 @@
---
title: Open WebUI
---
# Open WebUI
1. Install the [Docker](https://docs.docker.com/engine/install/)

View File

@ -1,6 +1,4 @@
---
title: Retrieval-Augmented Generation
---
# Retrieval-Augmented Generation
[Retrieval-augmented generation (RAG)](https://en.wikipedia.org/wiki/Retrieval-augmented_generation) is a technique that enables generative artificial intelligence (Gen AI) models to retrieve and incorporate new information. It modifies interactions with a large language model (LLM) so that the model responds to user queries with reference to a specified set of documents, using this information to supplement information from its pre-existing training data. This allows LLMs to use domain-specific and/or updated information. Use cases include providing chatbot access to internal company data or generating responses based on authoritative sources.

View File

@ -1,6 +1,4 @@
---
title: SkyPilot
---
# SkyPilot
<p align="center">
<img src="https://imgur.com/yxtzPEu.png" alt="vLLM"/>

View File

@ -1,6 +1,4 @@
---
title: Streamlit
---
# Streamlit
[Streamlit](https://github.com/streamlit/streamlit) lets you transform Python scripts into interactive web apps in minutes, instead of weeks. Build dashboards, generate reports, or create chat apps.

View File

@ -1,5 +1,3 @@
---
title: NVIDIA Triton
---
# NVIDIA Triton
The [Triton Inference Server](https://github.com/triton-inference-server) hosts a tutorial demonstrating how to quickly deploy a simple [facebook/opt-125m](https://huggingface.co/facebook/opt-125m) model using vLLM. Please see [Deploying a vLLM model in Triton](https://github.com/triton-inference-server/tutorials/blob/main/Quick_Deploy/vLLM/README.md#deploying-a-vllm-model-in-triton) for more details.

View File

@ -1,6 +1,4 @@
---
title: KServe
---
# KServe
vLLM can be deployed with [KServe](https://github.com/kserve/kserve) on Kubernetes for highly scalable distributed model serving.

View File

@ -1,6 +1,4 @@
---
title: KubeAI
---
# KubeAI
[KubeAI](https://github.com/substratusai/kubeai) is a Kubernetes operator that enables you to deploy and manage AI models on Kubernetes. It provides a simple and scalable way to deploy vLLM in production. Functionality such as scale-from-zero, load based autoscaling, model caching, and much more is provided out of the box with zero external dependencies.

View File

@ -1,6 +1,4 @@
---
title: Llama Stack
---
# Llama Stack
vLLM is also available via [Llama Stack](https://github.com/meta-llama/llama-stack) .

View File

@ -1,6 +1,4 @@
---
title: llmaz
---
# llmaz
[llmaz](https://github.com/InftyAI/llmaz) is an easy-to-use and advanced inference platform for large language models on Kubernetes, aimed for production use. It uses vLLM as the default model serving backend.

View File

@ -1,6 +1,4 @@
---
title: Production stack
---
# Production stack
Deploying vLLM on Kubernetes is a scalable and efficient way to serve machine learning models. This guide walks you through deploying vLLM using the [vLLM production stack](https://github.com/vllm-project/production-stack). Born out of a Berkeley-UChicago collaboration, [vLLM production stack](https://github.com/vllm-project/production-stack) is an officially released, production-optimized codebase under the [vLLM project](https://github.com/vllm-project), designed for LLM deployment with:

View File

@ -1,6 +1,4 @@
---
title: Using Kubernetes
---
# Using Kubernetes
Deploying vLLM on Kubernetes is a scalable and efficient way to serve machine learning models. This guide walks you through deploying vLLM using native Kubernetes.

View File

@ -1,6 +1,4 @@
---
title: Using Nginx
---
# Using Nginx
This document shows how to launch multiple vLLM serving containers and use Nginx to act as a load balancer between the servers.

View File

@ -1,6 +1,4 @@
---
title: Architecture Overview
---
# Architecture Overview
This document provides an overview of the vLLM architecture.

View File

@ -1,6 +1,4 @@
---
title: Automatic Prefix Caching
---
# Automatic Prefix Caching
The core idea of [PagedAttention](https://blog.vllm.ai/2023/06/20/vllm.html) is to partition the KV cache of each request into KV Blocks. Each block contains the attention keys and values for a fixed number of tokens. The PagedAttention algorithm allows these blocks to be stored in non-contiguous physical memory so that we can eliminate memory fragmentation by allocating the memory on demand.

View File

@ -1,6 +1,4 @@
---
title: Integration with HuggingFace
---
# Integration with HuggingFace
This document describes how vLLM integrates with HuggingFace libraries. We will explain step by step what happens under the hood when we run `vllm serve`.

View File

@ -1,6 +1,4 @@
---
title: vLLM Paged Attention
---
# vLLM Paged Attention
Currently, vLLM utilizes its own implementation of a multi-head query
attention kernel (`csrc/attention/attention_kernels.cu`).

View File

@ -1,6 +1,4 @@
---
title: Multi-Modal Data Processing
---
# Multi-Modal Data Processing
To enable various optimizations in vLLM such as [chunked prefill][chunked-prefill] and [prefix caching](../features/automatic_prefix_caching.md), we use [BaseMultiModalProcessor][vllm.multimodal.processing.BaseMultiModalProcessor] to provide the correspondence between placeholder feature tokens (e.g. `<image>`) and multi-modal inputs (e.g. the raw input image) based on the outputs of HF processor.

View File

@ -1,6 +1,4 @@
---
title: vLLM's Plugin System
---
# vLLM's Plugin System
The community frequently requests the ability to extend vLLM with custom features. To facilitate this, vLLM includes a plugin system that allows users to add custom features without modifying the vLLM codebase. This document explains how plugins work in vLLM and how to create a plugin for vLLM.

View File

@ -1,6 +1,4 @@
---
title: Automatic Prefix Caching
---
# Automatic Prefix Caching
## Introduction

View File

@ -1,6 +1,4 @@
---
title: Compatibility Matrix
---
# Compatibility Matrix
The tables below show mutually exclusive features and the support on some hardware.

View File

@ -1,6 +1,4 @@
---
title: Disaggregated Prefilling (experimental)
---
# Disaggregated Prefilling (experimental)
This page introduces you the disaggregated prefilling feature in vLLM.

View File

@ -1,6 +1,4 @@
---
title: LoRA Adapters
---
# LoRA Adapters
This document shows you how to use [LoRA adapters](https://arxiv.org/abs/2106.09685) with vLLM on top of a base model.

View File

@ -1,6 +1,4 @@
---
title: Multimodal Inputs
---
# Multimodal Inputs
This page teaches you how to pass multi-modal inputs to [multi-modal models][supported-mm-models] in vLLM.

View File

@ -1,6 +1,4 @@
---
title: Quantization
---
# Quantization
Quantization trades off model precision for smaller memory footprint, allowing large models to be run on a wider range of devices.

View File

@ -1,6 +1,4 @@
---
title: AutoAWQ
---
# AutoAWQ
To create a new 4-bit quantized model, you can leverage [AutoAWQ](https://github.com/casper-hansen/AutoAWQ).
Quantization reduces the model's precision from BF16/FP16 to INT4 which effectively reduces the total model memory footprint.

View File

@ -1,6 +1,4 @@
---
title: BitBLAS
---
# BitBLAS
vLLM now supports [BitBLAS](https://github.com/microsoft/BitBLAS) for more efficient and flexible model inference. Compared to other quantization frameworks, BitBLAS provides more precision combinations.

View File

@ -1,6 +1,4 @@
---
title: BitsAndBytes
---
# BitsAndBytes
vLLM now supports [BitsAndBytes](https://github.com/TimDettmers/bitsandbytes) for more efficient model inference.
BitsAndBytes quantizes models to reduce memory usage and enhance performance without significantly sacrificing accuracy.

View File

@ -1,6 +1,4 @@
---
title: FP8 W8A8
---
# FP8 W8A8
vLLM supports FP8 (8-bit floating point) weight and activation quantization using hardware acceleration on GPUs such as Nvidia H100 and AMD MI300x.
Currently, only Hopper and Ada Lovelace GPUs are officially supported for W8A8.

View File

@ -1,6 +1,4 @@
---
title: GGUF
---
# GGUF
!!! warning
Please note that GGUF support in vLLM is highly experimental and under-optimized at the moment, it might be incompatible with other features. Currently, you can use GGUF as a way to reduce memory footprint. If you encounter any issues, please report them to the vLLM team.

View File

@ -1,6 +1,4 @@
---
title: GPTQModel
---
# GPTQModel
To create a new 4-bit or 8-bit GPTQ quantized model, you can leverage [GPTQModel](https://github.com/ModelCloud/GPTQModel) from ModelCloud.AI.

View File

@ -1,6 +1,4 @@
---
title: INT4 W4A16
---
# INT4 W4A16
vLLM supports quantizing weights to INT4 for memory savings and inference acceleration. This quantization method is particularly useful for reducing model size and maintaining low latency in workloads with low queries per second (QPS).

View File

@ -1,6 +1,4 @@
---
title: INT8 W8A8
---
# INT8 W8A8
vLLM supports quantizing weights and activations to INT8 for memory savings and inference acceleration.
This quantization method is particularly useful for reducing model size while maintaining good performance.

View File

@ -1,6 +1,4 @@
---
title: Quantized KV Cache
---
# Quantized KV Cache
## FP8 KV Cache

View File

@ -1,6 +1,4 @@
---
title: AMD Quark
---
# AMD Quark
Quantization can effectively reduce memory and bandwidth usage, accelerate computation and improve
throughput while with minimal accuracy loss. vLLM can leverage [Quark](https://quark.docs.amd.com/latest/),

View File

@ -1,6 +1,4 @@
---
title: Supported Hardware
---
# Supported Hardware
The table below shows the compatibility of various quantization implementations with different hardware platforms in vLLM:

View File

@ -1,6 +1,4 @@
---
title: Reasoning Outputs
---
# Reasoning Outputs
vLLM offers support for reasoning models like [DeepSeek R1](https://huggingface.co/deepseek-ai/DeepSeek-R1), which are designed to generate outputs containing both reasoning steps and final conclusions.

View File

@ -1,6 +1,4 @@
---
title: Speculative Decoding
---
# Speculative Decoding
!!! warning
Please note that speculative decoding in vLLM is not yet optimized and does

View File

@ -1,6 +1,4 @@
---
title: Structured Outputs
---
# Structured Outputs
vLLM supports the generation of structured outputs using
[xgrammar](https://github.com/mlc-ai/xgrammar) or

View File

@ -1,6 +1,4 @@
---
title: Installation
---
# Installation
vLLM supports the following hardware platforms:

View File

@ -1,6 +1,4 @@
---
title: Quickstart
---
# Quickstart
This guide will help you quickly get started with vLLM to perform:

View File

@ -1,6 +1,4 @@
---
title: Loading models with Run:ai Model Streamer
---
# Loading models with Run:ai Model Streamer
Run:ai Model Streamer is a library to read tensors in concurrency, while streaming it to GPU memory.
Further reading can be found in [Run:ai Model Streamer Documentation](https://github.com/run-ai/runai-model-streamer/blob/master/docs/README.md).

View File

@ -1,6 +1,4 @@
---
title: Loading models with CoreWeave's Tensorizer
---
# Loading models with CoreWeave's Tensorizer
vLLM supports loading models with [CoreWeave's Tensorizer](https://docs.coreweave.com/coreweave-machine-learning-and-ai/inference/tensorizer).
vLLM model tensors that have been serialized to disk, an HTTP/HTTPS endpoint, or S3 endpoint can be deserialized

View File

@ -1,6 +1,4 @@
---
title: Generative Models
---
# Generative Models
vLLM provides first-class support for generative models, which covers most of LLMs.

View File

@ -1,6 +1,4 @@
---
title: TPU
---
# TPU
# TPU Supported Models
## Text-only Language Models

View File

@ -1,6 +1,4 @@
---
title: Pooling Models
---
# Pooling Models
vLLM also supports pooling models, including embedding, reranking and reward models.

View File

@ -1,6 +1,4 @@
---
title: Supported Models
---
# Supported Models
vLLM supports [generative](./generative_models.md) and [pooling](./pooling_models.md) models across various tasks.
If a model supports more than one task, you can set the task via the `--task` argument.

View File

@ -1,6 +1,4 @@
---
title: Distributed Inference and Serving
---
# Distributed Inference and Serving
## How to decide the distributed inference strategy?

View File

@ -1,6 +1,4 @@
---
title: LangChain
---
# LangChain
vLLM is also available via [LangChain](https://github.com/langchain-ai/langchain) .

View File

@ -1,6 +1,4 @@
---
title: LlamaIndex
---
# LlamaIndex
vLLM is also available via [LlamaIndex](https://github.com/run-llama/llama_index) .

View File

@ -1,6 +1,4 @@
---
title: Offline Inference
---
# Offline Inference
Offline inference is possible in your own code using vLLM's [`LLM`][vllm.LLM] class.
@ -23,7 +21,7 @@ The available APIs depend on the model type:
!!! info
[API Reference][offline-inference-api]
### Ray Data LLM API
## Ray Data LLM API
Ray Data LLM is an alternative offline inference API that uses vLLM as the underlying engine.
This API adds several batteries-included capabilities that simplify large-scale, GPU-efficient inference:

View File

@ -1,6 +1,4 @@
---
title: OpenAI-Compatible Server
---
# OpenAI-Compatible Server
vLLM provides an HTTP server that implements OpenAI's [Completions API](https://platform.openai.com/docs/api-reference/completions), [Chat API](https://platform.openai.com/docs/api-reference/chat), and more! This functionality lets you serve models and interact with them using an HTTP client.

View File

@ -1,6 +1,4 @@
---
title: Frequently Asked Questions
---
# Frequently Asked Questions
> Q: How can I serve multiple models on a single port using the OpenAI API?

View File

@ -1,6 +1,4 @@
---
title: Troubleshooting
---
# Troubleshooting
This document outlines some troubleshooting strategies you can consider. If you think you've discovered a bug, please [search existing issues](https://github.com/vllm-project/vllm/issues?q=is%3Aissue) first to see if it has already been reported. If not, please [file a new issue](https://github.com/vllm-project/vllm/issues/new/choose), providing as much relevant information as possible.