Skip to content

Commit 483ea64

Browse files
authored
[Docs] Replace all explicit anchors with real links (#27087)
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
1 parent e20eba7 commit 483ea64

28 files changed

+54
-139
lines changed

.markdownlint.yaml

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,6 @@ MD013: false
44
MD024:
55
siblings_only: true
66
MD033: false
7-
MD042: false
87
MD045: false
98
MD046: false
109
MD051: false

docs/api/README.md

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -20,8 +20,6 @@ API documentation for vLLM's configuration classes.
2020
- [vllm.config.CompilationConfig][]
2121
- [vllm.config.VllmConfig][]
2222

23-
[](){ #offline-inference-api }
24-
2523
## Offline Inference
2624

2725
LLM Class.
@@ -45,18 +43,14 @@ Engine classes for offline and online inference.
4543

4644
Inference parameters for vLLM APIs.
4745

48-
[](){ #sampling-params }
49-
5046
- [vllm.SamplingParams][]
5147
- [vllm.PoolingParams][]
5248

53-
[](){ #multi-modality }
54-
5549
## Multi-Modality
5650

5751
vLLM provides experimental support for multi-modal models through the [vllm.multimodal][] package.
5852

59-
Multi-modal inputs can be passed alongside text and token prompts to [supported models][supported-mm-models]
53+
Multi-modal inputs can be passed alongside text and token prompts to [supported models](../models/supported_models.md#list-of-multimodal-language-models)
6054
via the `multi_modal_data` field in [vllm.inputs.PromptType][].
6155

6256
Looking to add your own multi-modal model? Please follow the instructions listed [here](../contributing/model/multimodal.md).

docs/configuration/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,6 @@ This section lists the most common options for running vLLM.
44

55
There are three main levels of configuration, from highest priority to lowest priority:
66

7-
- [Request parameters][completions-api] and [input arguments][sampling-params]
7+
- [Request parameters](../serving/openai_compatible_server.md#completions-api) and [input arguments](../api/README.md#inference-parameters)
88
- [Engine arguments](./engine_args.md)
99
- [Environment variables](./env_vars.md)

docs/configuration/optimization.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -27,8 +27,6 @@ You can monitor the number of preemption requests through Prometheus metrics exp
2727

2828
In vLLM V1, the default preemption mode is `RECOMPUTE` rather than `SWAP`, as recomputation has lower overhead in the V1 architecture.
2929

30-
[](){ #chunked-prefill }
31-
3230
## Chunked Prefill
3331

3432
Chunked prefill allows vLLM to process large prefills in smaller chunks and batch them together with decode requests. This feature helps improve both throughput and latency by better balancing compute-bound (prefill) and memory-bound (decode) operations.

docs/contributing/benchmarks.md

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,8 @@ toc_depth: 4
77
vLLM provides comprehensive benchmarking tools for performance testing and evaluation:
88

99
- **[Benchmark CLI]**: `vllm bench` CLI tools and specialized benchmark scripts for interactive performance testing
10-
- **[Performance benchmarks][performance-benchmarks]**: Automated CI benchmarks for development
11-
- **[Nightly benchmarks][nightly-benchmarks]**: Comparative benchmarks against alternatives
10+
- **[Performance benchmarks](#performance-benchmarks)**: Automated CI benchmarks for development
11+
- **[Nightly benchmarks](#nightly-benchmarks)**: Comparative benchmarks against alternatives
1212

1313
[Benchmark CLI]: #benchmark-cli
1414

@@ -924,8 +924,6 @@ throughput numbers correctly is also adjusted.
924924

925925
</details>
926926

927-
[](){ #performance-benchmarks }
928-
929927
## Performance Benchmarks
930928

931929
The performance benchmarks are used for development to confirm whether new changes improve performance under various workloads. They are triggered on every commit with both the `perf-benchmarks` and `ready` labels, and when a PR is merged into vLLM.
@@ -988,8 +986,6 @@ The benchmarking currently runs on a predefined set of models configured in the
988986

989987
All continuous benchmarking results are automatically published to the public [vLLM Performance Dashboard](https://hud.pytorch.org/benchmark/llms?repoName=vllm-project%2Fvllm).
990988

991-
[](){ #nightly-benchmarks }
992-
993989
## Nightly Benchmarks
994990

995991
These compare vLLM's performance against alternatives (`tgi`, `trt-llm`, and `lmdeploy`) when there are major updates of vLLM (e.g., bumping up to a new version). They are primarily intended for consumers to evaluate when to choose vLLM over other options and are triggered on every commit with both the `perf-benchmarks` and `nightly-benchmarks` labels.

docs/contributing/model/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
# Summary
22

33
!!! important
4-
Many decoder language models can now be automatically loaded using the [Transformers backend][transformers-backend] without having to implement them in vLLM. See if `vllm serve <model>` works first!
4+
Many decoder language models can now be automatically loaded using the [Transformers backend](../../models/supported_models.md#transformers) without having to implement them in vLLM. See if `vllm serve <model>` works first!
55

66
vLLM models are specialized [PyTorch](https://pytorch.org/) models that take advantage of various [features](../../features/README.md#compatibility-matrix) to optimize their performance.
77

docs/contributing/model/registration.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ This page provides detailed instructions on how to do so.
88

99
## Built-in models
1010

11-
To add a model directly to the vLLM library, start by forking our [GitHub repository](https://github.com/vllm-project/vllm) and then [build it from source][build-from-source].
11+
To add a model directly to the vLLM library, start by forking our [GitHub repository](https://github.com/vllm-project/vllm) and then [build it from source](../../getting_started/installation/gpu.md#build-wheel-from-source).
1212
This gives you the ability to modify the codebase and test your model.
1313

1414
After you have implemented your model (see [tutorial](basic.md)), put it into the [vllm/model_executor/models](../../../vllm/model_executor/models) directory.

docs/contributing/model/tests.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -39,8 +39,6 @@ For [generative models](../../models/generative_models.md), there are two levels
3939

4040
For [pooling models](../../models/pooling_models.md), we simply check the cosine similarity, as defined in [tests/models/utils.py](../../../tests/models/utils.py).
4141

42-
[](){ #mm-processing-tests }
43-
4442
### Multi-modal processing
4543

4644
#### Common tests

docs/deployment/docker.md

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,5 @@
11
# Using Docker
22

3-
[](){ #deployment-docker-pre-built-image }
4-
53
## Use vLLM's Official Docker Image
64

75
vLLM offers an official Docker image for deployment.
@@ -62,8 +60,6 @@ You can add any other [engine-args](../configuration/engine_args.md) you need af
6260
RUN uv pip install --system git+https://github.com/huggingface/transformers.git
6361
```
6462

65-
[](){ #deployment-docker-build-image-from-source }
66-
6763
## Building vLLM's Docker Image from Source
6864

6965
You can build and run vLLM from source via the provided [docker/Dockerfile](../../docker/Dockerfile). To build vLLM:

docs/deployment/frameworks/anyscale.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,5 @@
11
# Anyscale
22

3-
[](){ #deployment-anyscale }
4-
53
[Anyscale](https://www.anyscale.com) is a managed, multi-cloud platform developed by the creators of Ray.
64

75
Anyscale automates the entire lifecycle of Ray clusters in your AWS, GCP, or Azure account, delivering the flexibility of open-source Ray

0 commit comments

Comments
 (0)