Skip to content

Commit cf5bdee

Browse files
authored
[deepseek v3.2] Update deepgemm version (#117)
Signed-off-by: Chen Zhang <[email protected]>
1 parent 494c567 commit cf5bdee

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

DeepSeek/DeepSeek-V3_2-Exp.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@
77

88
```bash
99
uv pip install vllm --extra-index-url https://wheels.vllm.ai/nightly
10-
uv pip install https://wheels.vllm.ai/dsv32/deep_gemm-2.1.0%2B594953a-cp312-cp312-linux_x86_64.whl
10+
uv pip install git+https://github.com/deepseek-ai/[email protected] --no-build-isolation # Other versions may also work. We recommend using the latest released version from https://github.com/deepseek-ai/DeepGEMM/releases
1111
```
1212

1313
Note: DeepGEMM is used in two places: MoE and MQA logits computation. It is necessary for MQA logits computation. If you want to disable the MoE part, you can set `VLLM_USE_DEEP_GEMM=0` in the environment variable. Some users reported that the performance is better with `VLLM_USE_DEEP_GEMM=0`, e.g. on H20 GPUs. It might be also beneficial to disable DeepGEMM if you want to skip the long warmup.

0 commit comments

Comments
 (0)