Skip to content

Commit 4d6b510

Browse files
AllenDoujimpang
authored andcommitted
[Misc] Fix arg names (vllm-project#5524)
1 parent eec9fc7 commit 4d6b510

File tree

3 files changed

+6
-6
lines changed

3 files changed

+6
-6
lines changed

benchmarks/kernels/benchmark_paged_attention.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -165,7 +165,7 @@ def run_cuda_benchmark(num_iters: int, profile: bool = False) -> float:
165165
choices=["v1", "v2"],
166166
default="v2")
167167
parser.add_argument("--batch-size", type=int, default=8)
168-
parser.add_argument("--seq_len", type=int, default=4096)
168+
parser.add_argument("--seq-len", type=int, default=4096)
169169
parser.add_argument("--num-query-heads", type=int, default=64)
170170
parser.add_argument("--num-kv-heads", type=int, default=8)
171171
parser.add_argument("--head-size",

examples/aqlm_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ def main():
1717
type=int,
1818
default=0,
1919
help='known good models by index, [0-4]')
20-
parser.add_argument('--tensor_parallel_size',
20+
parser.add_argument('--tensor-parallel-size',
2121
'-t',
2222
type=int,
2323
default=1,

examples/fp8/extract_scales.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -327,7 +327,7 @@ def main(args):
327327
"--quantization-param-path <filename>). This is only used "
328328
"if the KV cache dtype is FP8 and on ROCm (AMD GPU).")
329329
parser.add_argument(
330-
"--quantized_model",
330+
"--quantized-model",
331331
help="Specify the directory containing a single quantized HF model. "
332332
"It is expected that the quantization format is FP8_E4M3, for use "
333333
"on ROCm (AMD GPU).",
@@ -339,18 +339,18 @@ def main(args):
339339
choices=["auto", "safetensors", "npz", "pt"],
340340
default="auto")
341341
parser.add_argument(
342-
"--output_dir",
342+
"--output-dir",
343343
help="Optionally specify the output directory. By default the "
344344
"KV cache scaling factors will be saved in the model directory, "
345345
"however you can override this behavior here.",
346346
default=None)
347347
parser.add_argument(
348-
"--output_name",
348+
"--output-name",
349349
help="Optionally specify the output filename.",
350350
# TODO: Change this once additional scaling factors are enabled
351351
default="kv_cache_scales.json")
352352
parser.add_argument(
353-
"--tp_size",
353+
"--tp-size",
354354
help="Optionally specify the tensor-parallel (TP) size that the "
355355
"quantized model should correspond to. If specified, during KV "
356356
"cache scaling factor extraction the observed TP size will be "

0 commit comments

Comments
 (0)