Skip to content
This repository was archived by the owner on Oct 11, 2024. It is now read-only.

Commit 42abcff

Browse files
kerthcetrshaw@neuralmagic.com
authored and
rshaw@neuralmagic.com
committed
[Bugfix] Fix flag name for max_seq_len_to_capture (vllm-project#4935)
Signed-off-by: kerthcet <kerthcet@gmail.com>
1 parent f270b9c commit 42abcff

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

vllm/engine/arg_utils.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -356,9 +356,9 @@ def add_cli_args(
356356
help='Maximum context length covered by CUDA '
357357
'graphs. When a sequence has context length '
358358
'larger than this, we fall back to eager mode. '
359-
'(DEPRECATED. Use --max-seq_len-to-capture instead'
359+
'(DEPRECATED. Use --max-seq-len-to-capture instead'
360360
')')
361-
parser.add_argument('--max-seq_len-to-capture',
361+
parser.add_argument('--max-seq-len-to-capture',
362362
type=int,
363363
default=EngineArgs.max_seq_len_to_capture,
364364
help='Maximum sequence length covered by CUDA '

0 commit comments

Comments
 (0)