Skip to content

Commit 75e0ae1

Browse files
committed
Add remaining vLLM short long option name mapping, use 'vllm serve' for launch in place of the old command
1 parent 07a4973 commit 75e0ae1

File tree

1 file changed

+6
-1
lines changed

1 file changed

+6
-1
lines changed

vec_inf/client/_client_vars.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -78,7 +78,12 @@
7878
VLLM_SHORT_TO_LONG_MAP = {
7979
"-tp": "--tensor-parallel-size",
8080
"-pp": "--pipeline-parallel-size",
81+
"-dp": "--data-parallel-size",
82+
"-dpl": "--data-parallel-size-local",
83+
"-dpa": "--data-parallel-address",
84+
"-dpp": "--data-parallel-rpc-port",
8185
"-O": "--compilation-config",
86+
"-q": "--quantization",
8287
}
8388

8489

@@ -214,7 +219,7 @@ class SlurmScriptTemplate(TypedDict):
214219
' && mv temp.json "$json_path"',
215220
],
216221
"launch_cmd": [
217-
"python3.10 -m vllm.entrypoints.openai.api_server \\",
222+
"vllm serve \\",
218223
" --model {model_weights_path} \\",
219224
" --served-model-name {model_name} \\",
220225
' --host "0.0.0.0" \\',

0 commit comments

Comments
 (0)