Add MiniMax-M2.5 FP8 vLLM benchmark for B200#757
Add MiniMax-M2.5 FP8 vLLM benchmark for B200#757functionstackx wants to merge 17 commits intomainfrom
Conversation
- Add benchmark script benchmarks/minimaxm2.5_fp8_b200.sh - Add config entry minimaxm2.5-fp8-b200-vllm to nvidia-master.yaml - Image: vllm/vllm-openai:v0.15.1-cu130 - Model: MiniMaxAI/MiniMax-M2.5 - TP=2 and TP=4, concurrency 4-64 - Sequence lengths: 1k1k, 1k8k, 8k1k - Update perf-changelog.yaml Closes #756 Co-authored-by: functionstackx <functionstackx@users.noreply.github.com>
34f7756 to
cbc5d0b
Compare
|
@claude please review. i made a few changes |
|
Claude finished @kedarpotdar-nv's task in 2m 40s —— View job Review of PR #757
Bug: Wrong
|
|
@claude implement this fix in this PR's runners/launch_b200-dgxc-slurm.sh too
|
|
Claude finished @functionstackx's task in 1m 5s —— View job Implementing fix in
|
Change HF_HUB_CACHE_MOUNT from /home/sa-shared/gharunners/hf-hub-cache/ to /scratch/fsw/gharunners/hf-hub-cache Co-authored-by: functionstackx <functionstackx@users.noreply.github.com>
|
@ankursingh-nv i cancelled the workflow as it seems to be hard erroring out |
Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
66a55e0 to
b56773c
Compare

Add MiniMax-M2.5 FP8 vLLM benchmark configuration for B200 GPUs.
benchmarks/minimaxm2.5_fp8_b200.shminimaxm2.5-fp8-b200-vllmin nvidia-master.yamlvllm/vllm-openai:v0.15.1-cu130Closes #756
Generated with Claude Code