Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
29 commits
Select commit Hold shift + click to select a range
154be8d
Port DeepSeek-V4-Pro FP4 disaggregated vLLM sweep from gb200 to gb300-cr
Oseltamivir Apr 25, 2026
017b66a
Fill in PR link for gb300-cr changelog entry
Oseltamivir Apr 25, 2026
b91ca49
Rename gb300-cr to gb300-cw; fix model path to /mnt/vast/models/dsv4/
Oseltamivir Apr 25, 2026
b6ebbd3
Fix gb300-cw SLURM account and extend runner group to _2/_3
Oseltamivir Apr 25, 2026
c6b45fd
Pin runner-side uv to /tmp so x86 binary doesn't leak to ARM64 compute
Oseltamivir Apr 25, 2026
aaea407
Force --segment per recipe via sbatch_directives
Oseltamivir Apr 25, 2026
3bd82f1
Cap cargo parallelism via CARGO_BUILD_JOBS=4 in gb300 recipes
Oseltamivir Apr 25, 2026
b3d2b12
Force --mem=0 (use full node memory) on every gb300 recipe; fix hered…
Oseltamivir Apr 25, 2026
b3d2bd8
Merge branch 'main' into dsv4-fp4-gb300-dynamo-vllm-disagg
Oseltamivir Apr 25, 2026
33f6eb4
Update perf-changelog.yaml
Oseltamivir Apr 25, 2026
43c3bc4
Update gb300 recipe headers — segment is recipe-driven, not auto
Oseltamivir Apr 25, 2026
32aca3e
Set NVIDIA_VISIBLE_DEVICES + DRIVER_CAPABILITIES so enroot mounts lib…
Oseltamivir Apr 25, 2026
e66e667
Cache dynamo wheel build globally on /mnt/vast (gb300-cw)
Oseltamivir Apr 25, 2026
9cb8ee5
Switch dynamo cache lock from flock to mkdir (NFS doesn't honor flock)
Oseltamivir Apr 25, 2026
369b1ed
Pre-build dynamo wheel via single-node srun before sbatch (gb300-cw)
Oseltamivir Apr 25, 2026
f37eb70
Prebuild srun: add --mem=0, cap CARGO_BUILD_JOBS=8, drop rustc debuginfo
Oseltamivir Apr 25, 2026
86ac394
Mount /mnt/vast/dynamo_cache into worker containers (extra_mount)
Oseltamivir Apr 25, 2026
6997f95
Patch vllm HANDSHAKE_TIMEOUT_MINS 5->30 in setup script
Oseltamivir Apr 25, 2026
3900434
Drop NVL-only NCCL flags + add NCCL_DEBUG=INFO
Oseltamivir Apr 25, 2026
7851967
Re-add NCCL_MNNVL_ENABLE, add debug diagnostics, reduce to 1p1d repro…
Oseltamivir Apr 25, 2026
87bdf1f
Remove vLLM HANDSHAKE_TIMEOUT_MINS sed patch from setup script
Oseltamivir Apr 25, 2026
7f526db
Restore handshake timeout patch, add DP Coordinator logging, drop NCC…
Oseltamivir Apr 25, 2026
6415458
Rewrite coordinator patch to match actual vLLM source strings
Oseltamivir Apr 25, 2026
cedac56
Rewrite coordinator patch: regex matching + inspect.getsource verify
Oseltamivir Apr 25, 2026
ff4ab3a
Merge branch 'main' into dsv4-fp4-gb300-dynamo-vllm-disagg
Oseltamivir Apr 25, 2026
8570717
more
Oseltamivir Apr 26, 2026
b39f41e
Merge branch 'main' into dsv4-fp4-gb300-dynamo-vllm-disagg
Oseltamivir Apr 26, 2026
3f33f27
Merge branch 'main' into dsv4-fp4-gb300-dynamo-vllm-disagg
Oseltamivir Apr 26, 2026
df79838
configs
Oseltamivir Apr 26, 2026
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
99 changes: 99 additions & 0 deletions .github/configs/nvidia-master.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7666,3 +7666,102 @@ dsv4-fp4-gb200-dynamo-vllm:
tp: 16
ep: 16
dp-attn: true

dsv4-fp4-gb300-dynamo-vllm:
image: vllm/vllm-openai:deepseekv4-cu130
model: deepseek-ai/DeepSeek-V4-Pro
model-prefix: dsv4
runner: gb300-cw
precision: fp4
framework: dynamo-vllm
multinode: true
disagg: true
# Same topology + tuning as dsv4-fp4-gb300-dynamo-vllm's gb200 sibling, just
# pointed at the gb300 recipe variants. Cluster gb300-cw is 2x 18-node
# racks; each job is rack-pinned via srtctl's auto `#SBATCH --segment={N}`.
Comment on lines +7679 to +7681
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🟡 Comment at line 7666 reads "Same topology + tuning as dsv4-fp4-gb300-dynamo-vllm's gb200 sibling" but it's inside the dsv4-fp4-gb300-dynamo-vllm config block (declared at line 7657) — so the config is referring to itself as having a gb200 sibling. Almost certainly a copy-paste leftover from the GB200→GB300 port; should reference dsv4-fp4-gb200-dynamo-vllm instead. Pure comment-only nit, no runtime effect.

Extended reasoning...

What's wrong

In .github/configs/nvidia-master.yaml the new config block dsv4-fp4-gb300-dynamo-vllm: is declared at line 7657. The header comment for that block, lines 7666-7668, currently reads:

dsv4-fp4-gb300-dynamo-vllm:        # line 7657
  ...
  # Same topology + tuning as dsv4-fp4-gb300-dynamo-vllm's gb200 sibling, just  # line 7666
  # pointed at the gb300 recipe variants. Cluster gb300-cr is 2x 18-node
  # racks; each job is rack-pinned via srtctl's auto `#SBATCH --segment={N}`.

A config can't be its own sibling. The author clearly intended to point at dsv4-fp4-gb200-dynamo-vllm — that is the existing GB200 config defined immediately above (ending at line 7655) and the actual upstream this PR ports from.

Step-by-step proof

  1. Line 7657: dsv4-fp4-gb300-dynamo-vllm: — this opens the config block.
  2. Lines 7658-7665: scalar fields (image, model, model-prefix, runner, precision, framework, multinode, disagg) — all still inside the block opened at 7657.
  3. Line 7666: comment under that same key, which begins "Same topology + tuning as dsv4-fp4-gb300-dynamo-vllm's gb200 sibling…"
  4. The phrase "dsv4-fp4-gb300-dynamo-vllm's gb200 sibling" reads as "the gb200 sibling of dsv4-fp4-gb300-dynamo-vllm" — i.e. the current block's gb200 sibling, which is dsv4-fp4-gb200-dynamo-vllm (lines 7544-7655). Saying "X's gb200 sibling" while being X is a tautology with no referent.
  5. The PR description corroborates: "Same DSV4-Pro FP4 sweep we already run on gb200, ported to the gb300-cr cluster" — i.e. the sibling is gb200, not gb300.

Impact

None on runtime, parsing, generated artifacts, or sweep behavior — YAML comments are inert. This is purely a readability issue: a future reader following the comment will go looking for a non-existent reference.

Fix

Change dsv4-fp4-gb300-dynamo-vllm's gb200 sibling to dsv4-fp4-gb200-dynamo-vllm (or equivalent phrasing such as "as the gb200 sibling (dsv4-fp4-gb200-dynamo-vllm)"). One-token edit while the PR is still open.

seq-len-configs:
- isl: 1024
osl: 1024
search-space:
- conc-list: [1, 4, 8, 16, 32, 64]
prefill:
num-worker: 1
tp: 8
ep: 8
dp-attn: true
additional-settings:
- "CONFIG_FILE=recipes/vllm/deepseek-v4/1k1k/disagg-gb300-1p1d-dep8-tep8.yaml"
decode:
num-worker: 1
tp: 8
ep: 1
dp-attn: false
- conc-list: [128, 256, 1024, 2048, 4096]
prefill:
num-worker: 1
tp: 8
ep: 8
dp-attn: true
additional-settings:
- "CONFIG_FILE=recipes/vllm/deepseek-v4/1k1k/disagg-gb300-1p1d-dep8-dep16.yaml"
decode:
num-worker: 1
tp: 16
ep: 16
dp-attn: true
- conc-list: [4096, 8192]
prefill:
num-worker: 3
tp: 8
ep: 8
dp-attn: true
additional-settings:
- "CONFIG_FILE=recipes/vllm/deepseek-v4/1k1k/disagg-gb300-3p1d-dep8-dep16.yaml"
decode:
num-worker: 1
tp: 16
ep: 16
dp-attn: true

- isl: 8192
osl: 1024
search-space:
- conc-list: [1, 4, 8, 16, 32, 64]
prefill:
num-worker: 1
tp: 8
ep: 8
dp-attn: true
additional-settings:
- "CONFIG_FILE=recipes/vllm/deepseek-v4/8k1k/disagg-gb300-1p1d-dep8-tep8.yaml"
decode:
num-worker: 1
tp: 8
ep: 1
dp-attn: false
- conc-list: [512, 1024]
prefill:
num-worker: 3
tp: 8
ep: 8
dp-attn: true
additional-settings:
- "CONFIG_FILE=recipes/vllm/deepseek-v4/8k1k/disagg-gb300-3p1d-dep8-dep16.yaml"
decode:
num-worker: 1
tp: 16
ep: 16
dp-attn: true
- conc-list: [4096, 8192]
prefill:
num-worker: 7
tp: 8
ep: 8
dp-attn: true
additional-settings:
- "CONFIG_FILE=recipes/vllm/deepseek-v4/8k1k/disagg-gb300-7p1d-dep8-dep16.yaml"
decode:
num-worker: 1
tp: 16
ep: 16
dp-attn: true
5 changes: 5 additions & 0 deletions .github/configs/runners.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -131,3 +131,8 @@ gb300:
- 'gb300-nv_0'
- 'gb300-nv_1'
- 'gb300-nv_2'
gb300-cw:
- 'gb300-cw_0'
- 'gb300-cw_1'
- 'gb300-cw_2'
- 'gb300-cw_3'
Original file line number Diff line number Diff line change
@@ -0,0 +1,139 @@
name: "dsv4-vllm-disagg-gb300-1p1d-dep8-dep16"

# GB300 mirror of disagg-gb200-1p1d-dep8-dep16.yaml. Same tuning at FP4
# (288 GB HBM/GPU on GB300 vs 184 GB on GB200 — extra headroom for KV).
# Cluster: gb300-cw (2x 18-node racks); job pins to one rack via the
# explicit sbatch_directives.segment="6" below (cw's srtslurm.yaml turns
# off srtctl's auto-segment so each recipe owns its segment value).
#
# 1k/1k mid-to-high throughput topology. Single prefill worker feeding a
# wide DP=16 decode handles conc 256-4096 cleanly for 1k prompts.

model:
path: "deepseek-v4-pro"
container: "vllm/vllm-openai:deepseekv4-cu130"
precision: "fp4"

dynamo:
hash: 6a159fedd8e4a1563aa647c31f622aedbf254b5b
# Install handled by our custom vllm-container-deps.sh, which builds
# the dynamo wheel ONCE on /mnt/vast and lets every rank pip-install
# from cache. See runners/gb300-cw-vllm-container-deps.sh.
install: false

setup_script: vllm-container-deps.sh

# Mount /mnt/vast/dynamo_cache into every worker container so each
# rank can pip-install from the wheel that launch_gb300-cw.sh
# pre-built there. Without this only /mnt/vast/models/<model> is
# in scope and our setup script errors out with 'prebuilt cache
# missing'.
extra_mount:
- "/mnt/vast/dynamo_cache:/mnt/vast/dynamo_cache"

# Pin all 6 nodes to the same rack on cw.
sbatch_directives:
segment: "6"
# Use all node memory; cw default was too tight.
mem: "0"

slurm:
time_limit: "8:00:00"

health_check:
max_attempts: 1440
interval_seconds: 10

resources:
gpu_type: "gb300"
gpus_per_node: 4
prefill_nodes: 2
decode_nodes: 4
prefill_workers: 1
decode_workers: 1
gpus_per_prefill: 8
gpus_per_decode: 16

frontend:
type: dynamo
enable_multiple_frontends: false

backend:
type: vllm
connector: null

prefill_environment:
VLLM_ENGINE_READY_TIMEOUT_S: "3600"
TILELANG_CLEANUP_TEMP_FILES: "1"
CARGO_BUILD_JOBS: "4"
VLLM_USE_NCCL_SYMM_MEM: "1"
NCCL_CUMEM_ENABLE: "1"
NCCL_MNNVL_ENABLE: "1"
NCCL_NVLS_ENABLE: "1"
NCCL_DEBUG: "INFO"
NCCL_DEBUG_SUBSYS: "INIT,BOOTSTRAP,ENV,NET,GRAPH,NVLS"
VLLM_SERVER_DEV_MODE: "1"

decode_environment:
VLLM_ENGINE_READY_TIMEOUT_S: "3600"
TILELANG_CLEANUP_TEMP_FILES: "1"
CARGO_BUILD_JOBS: "4"
VLLM_USE_NCCL_SYMM_MEM: "1"
NCCL_CUMEM_ENABLE: "1"
NCCL_MNNVL_ENABLE: "1"
NCCL_NVLS_ENABLE: "1"
NCCL_DEBUG: "INFO"
NCCL_DEBUG_SUBSYS: "INIT,BOOTSTRAP,ENV,NET,GRAPH,NVLS"
VLLM_SERVER_DEV_MODE: "1"

vllm_config:
prefill:
kv-transfer-config: '{"kv_connector": "NixlConnector", "kv_role": "kv_both"}'
served-model-name: "deepseek-ai/DeepSeek-V4-Pro"
kv-cache-dtype: "fp8"
tensor-parallel-size: 1
pipeline-parallel-size: 1
data-parallel-size: 8
data-parallel-rpc-port: 13345
enable-expert-parallel: true
enforce-eager: true
max-model-len: 3072
max-num-seqs: 16
max-num-batched-tokens: 16384
trust-remote-code: true
no-enable-prefix-caching: true
no-enable-flashinfer-autotune: true
block-size: 256
gpu-memory-utilization: 0.88
no-disable-hybrid-kv-cache-manager: true
enable-sleep-mode: true

decode:
kv-transfer-config: '{"kv_connector": "NixlConnector", "kv_role": "kv_both"}'
served-model-name: "deepseek-ai/DeepSeek-V4-Pro"
kv-cache-dtype: "fp8"
tensor-parallel-size: 1
pipeline-parallel-size: 1
data-parallel-size: 16
data-parallel-rpc-port: 13345
enable-expert-parallel: true
max-model-len: 3072
max-num-seqs: 512
max-cudagraph-capture-size: 512
max-num-batched-tokens: 512
trust-remote-code: true
no-enable-prefix-caching: true
block-size: 256
compilation-config: '{"cudagraph_mode":"FULL_DECODE_ONLY","mode":0}'
gpu-memory-utilization: 0.9
stream-interval: 50
no-disable-hybrid-kv-cache-manager: true
enable-sleep-mode: true

benchmark:
type: "sa-bench"
isl: 1024
osl: 1024
concurrencies: "128x256x1024x2048x4096"
req_rate: "inf"
use_chat_template: false
Loading
Loading