Skip to content

Commit 844f881

Browse files
authored
docs: Bring back some missed release/0.4.0 doc changes, fix broken links, add lychee link checker github action (#2482)
1 parent 41f095c commit 844f881

File tree

45 files changed

+202
-83
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

45 files changed

+202
-83
lines changed
Lines changed: 61 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,61 @@
1+
name: Docs link check
2+
3+
on:
4+
push:
5+
branches:
6+
- main
7+
pull_request:
8+
9+
permissions:
10+
contents: read
11+
12+
jobs:
13+
lychee:
14+
runs-on: ubuntu-latest
15+
steps:
16+
- name: Check out repository
17+
uses: actions/checkout@v4
18+
19+
# Cache lychee results (e.g. to avoid hitting rate limits)
20+
# https://lychee.cli.rs/github_action_recipes/caching/
21+
- name: Restore lychee cache
22+
uses: actions/cache@v4
23+
with:
24+
path: .lycheecache
25+
key: cache-lychee-${{ github.sha }}
26+
restore-keys: cache-lychee-
27+
28+
# https://github.com/lycheeverse/lychee/issues/1487
29+
- name: Install CA Certificates for lychee
30+
run: |
31+
sudo apt-get install ca-certificates
32+
33+
- name: Install lychee
34+
run: |
35+
set -euo pipefail
36+
mkdir -p "$HOME/.local/bin"
37+
cd "$RUNNER_TEMP"
38+
# TODO: Lychee v0.19.1 doesn't support regex in --exclude-path, so use nightly
39+
# release until there is a released version containing regex support.
40+
curl -sSL -o lychee.tar.gz \
41+
https://github.com/lycheeverse/lychee/releases/download/nightly/lychee-x86_64-unknown-linux-gnu.tar.gz
42+
tar -xzf lychee.tar.gz
43+
BIN_PATH=$(find . -maxdepth 2 -type f -name lychee | head -n1)
44+
install -m 0755 "$BIN_PATH" "$HOME/.local/bin/lychee"
45+
echo "$HOME/.local/bin" >> "$GITHUB_PATH"
46+
lychee --version
47+
48+
- name: Check documentation links with lychee
49+
env:
50+
# Set GITHUB_TOKEN to avoid github rate limits on URL checks
51+
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
52+
run: |
53+
set -euo pipefail
54+
# Run lychee against all files in repo
55+
lychee \
56+
--cache \
57+
--no-progress \
58+
--exclude-path "ATTRIBUTIONS.*" \
59+
--accept "200..=299, 403, 429" \
60+
--exclude-all-private --exclude 0.0.0.0 \
61+
.

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -183,7 +183,7 @@ Run the backend/worker like this:
183183
python -m dynamo.sglang.worker --help
184184
```
185185

186-
You can pass any sglang flags directly to this worker, see https://docs.sglang.ai/backend/server_arguments.html . See there to use multiple GPUs.
186+
You can pass any sglang flags directly to this worker, see https://docs.sglang.ai/advanced_features/server_arguments.html . See there to use multiple GPUs.
187187

188188
## TensorRT-LLM
189189

benchmarks/llm/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,4 +12,4 @@ See the License for the specific language governing permissions and
1212
limitations under the License.
1313
-->
1414

15-
[../../examples/llm/benchmarks/README.md](../../examples/llm/benchmarks/README.md)
15+
Coming soon.

components/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -77,4 +77,4 @@ To get started with Dynamo components:
7777
4. **Run deployment scripts** from the engine's launch directory
7878
5. **Monitor performance** using the metrics component
7979

80-
For detailed instructions, see the README files in each component directory and the main [Dynamo documentation](../../docs/).
80+
For detailed instructions, see the README files in each component directory and the main [Dynamo documentation](../docs/).

components/backends/sglang/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ git checkout $(git describe --tags $(git rev-list --tags --max-count=1))
5252

5353
## Quick Start
5454

55-
Below we provide a guide that lets you run all of our the common deployment patterns on a single node. See our different [architectures](../llm/README.md#deployment-architectures) for a high level overview of each pattern and the architecture diagram for each.
55+
Below we provide a guide that lets you run all of our common deployment patterns on a single node.
5656

5757
### Start NATS and ETCD in the background
5858

components/backends/sglang/deploy/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ extraPodSpec:
7474

7575
Before using these templates, ensure you have:
7676

77-
1. **Dynamo Cloud Platform installed** - See [Installing Dynamo Cloud](../../docs/guides/dynamo_deploy/dynamo_cloud.md)
77+
1. **Dynamo Cloud Platform installed** - See [Installing Dynamo Cloud](../../../../docs/guides/dynamo_deploy/dynamo_cloud.md)
7878
2. **Kubernetes cluster with GPU support**
7979
3. **Container registry access** for SGLang runtime images
8080
4. **HuggingFace token secret** (referenced as `envFromSecret: hf-token-secret`)
@@ -159,4 +159,4 @@ Common issues and solutions:
159159
3. **Health check failures**: Review model loading logs and increase `initialDelaySeconds`
160160
4. **Out of memory**: Increase memory limits or reduce model batch size
161161

162-
For additional support, refer to the [deployment troubleshooting guide](../../docs/guides/dynamo_deploy/quickstart.md#troubleshooting).
162+
For additional support, refer to the [deployment guide](../../../../docs/guides/dynamo_deploy/quickstart.md).

components/backends/sglang/docs/dsr1-wideep-h100.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ SPDX-License-Identifier: Apache-2.0
55

66
# Running DeepSeek-R1 Disaggregated with WideEP on H100s
77

8-
Dynamo supports SGLang's implementation of wide expert parallelism and large scale P/D for DeepSeek-R1! You can read their blog post [here](https://www.nvidia.com/en-us/technologies/ai/deepseek-r1-large-scale-p-d-with-wide-expert-parallelism/) for more details. We provide a Dockerfile for this in `container/Dockerfile.sglang-deepep` and configurations to deploy this at scale. In this example, we will run 1 prefill worker on 4 H100 nodes and 1 decode worker on 9 H100 nodes (104 total GPUs).
8+
Dynamo supports SGLang's implementation of wide expert parallelism and large scale P/D for DeepSeek-R1! You can read their blog post [here](https://lmsys.org/blog/2025-05-05-large-scale-ep/) for more details. We provide a Dockerfile for this in `container/Dockerfile.sglang-wideep` and configurations to deploy this at scale. In this example, we will run 1 prefill worker on 4 H100 nodes and 1 decode worker on 9 H100 nodes (104 total GPUs).
99

1010
## Instructions
1111

components/backends/sglang/slurm_jobs/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
# Example: Deploy Multi-node SGLang with Dynamo on SLURM
22

3-
This folder implements the example of [SGLang DeepSeek-R1 Disaggregated with WideEP](../dsr1-wideep.md) on a SLURM cluster.
3+
This folder implements the example of [SGLang DeepSeek-R1 Disaggregated with WideEP](../docs/dsr1-wideep-h100.md) on a SLURM cluster.
44

55
## Overview
66

7-
The scripts in this folder set up multiple cluster nodes to run the [SGLang DeepSeek-R1 Disaggregated with WideEP](../dsr1-wideep.md) example, with separate nodes handling prefill and decode.
7+
The scripts in this folder set up multiple cluster nodes to run the [SGLang DeepSeek-R1 Disaggregated with WideEP](../docs/dsr1-wideep-h100.md) example, with separate nodes handling prefill and decode.
88
The node setup is done using Python job submission scripts with Jinja2 templates for flexible configuration. The setup also includes GPU utilization monitoring capabilities to track performance during benchmarks.
99

1010
## Scripts
@@ -57,7 +57,7 @@ For simplicity of the example, we will make some assumptions about your SLURM cl
5757
If your cluster supports similar container based plugins, you may be able to
5858
modify the template to use that instead.
5959
3. We assume you have already built a recent Dynamo+SGLang container image as
60-
described [here](../dsr1-wideep.md#instructions).
60+
described [here](../docs/dsr1-wideep-h100.md#instructions).
6161
This is the image that can be passed to the `--container-image` argument in later steps.
6262

6363
## Usage

components/backends/trtllm/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -193,7 +193,7 @@ For complete Kubernetes deployment instructions, configurations, and troubleshoo
193193

194194
### Client
195195

196-
See [client](../llm/README.md#client) section to learn how to send request to the deployment.
196+
See [client](../sglang/README.md#testing-the-deployment) section to learn how to send request to the deployment.
197197

198198
NOTE: To send a request to a multi-node deployment, target the node which is running `python3 -m dynamo.frontend <args>`.
199199

@@ -218,7 +218,7 @@ DISAGGREGATION_STRATEGY="prefill_first" ./launch/disagg.sh
218218

219219
## KV Cache Transfer in Disaggregated Serving
220220

221-
Dynamo with TensorRT-LLM supports two methods for transferring KV cache in disaggregated serving: UCX (default) and NIXL (experimental). For detailed information and configuration instructions for each method, see the [KV cache transfer guide](./kv-cache-tranfer.md).
221+
Dynamo with TensorRT-LLM supports two methods for transferring KV cache in disaggregated serving: UCX (default) and NIXL (experimental). For detailed information and configuration instructions for each method, see the [KV cache transfer guide](./kv-cache-transfer.md).
222222

223223

224224
## Request Migration
@@ -233,7 +233,7 @@ This allows a request to be migrated up to 3 times before failing. See the [Requ
233233

234234
## Client
235235

236-
See [client](../llm/README.md#client) section to learn how to send request to the deployment.
236+
See [client](../sglang/README.md#testing-the-deployment) section to learn how to send request to the deployment.
237237

238238
NOTE: To send a request to a multi-node deployment, target the node which is running `python3 -m dynamo.frontend <args>`.
239239

components/backends/trtllm/deploy/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -211,7 +211,7 @@ envs:
211211
212212
## Testing the Deployment
213213
214-
Send a test request to verify your deployment. See the [client section](../../../../components/backends/llm/README.md#client) for detailed instructions.
214+
Send a test request to verify your deployment. See the [client section](../../../../components/backends/vllm/README.md#client) for detailed instructions.
215215
216216
**Note:** For multi-node deployments, target the node running `python3 -m dynamo.frontend <args>`.
217217

@@ -241,7 +241,7 @@ TensorRT-LLM supports two methods for KV cache transfer in disaggregated serving
241241
- **UCX** (default): Standard method for KV cache transfer
242242
- **NIXL** (experimental): Alternative transfer method
243243

244-
For detailed configuration instructions, see the [KV cache transfer guide](../kv-cache-tranfer.md).
244+
For detailed configuration instructions, see the [KV cache transfer guide](../kv-cache-transfer.md).
245245

246246
## Request Migration
247247

0 commit comments

Comments
 (0)