Skip to content

Commit 02d411f

Browse files
[Doc]: fix typos in Python comments (vllm-project#24115)
Signed-off-by: Didier Durand <[email protected]>
1 parent d7e1e59 commit 02d411f

File tree

11 files changed

+11
-11
lines changed

11 files changed

+11
-11
lines changed

.buildkite/nightly-benchmarks/scripts/compare-json-results.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -218,7 +218,7 @@ def split_json_by_tp_pp(
218218
"--xaxis",
219219
type=str,
220220
default="# of max concurrency.",
221-
help="column name to use as X Axis in comparision graph",
221+
help="column name to use as X Axis in comparison graph",
222222
)
223223
args = parser.parse_args()
224224

benchmarks/benchmark_serving.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1104,7 +1104,7 @@ def create_argument_parser():
11041104
"--percentile-metrics",
11051105
type=str,
11061106
default="ttft,tpot,itl",
1107-
help="Comma-separated list of selected metrics to report percentils. "
1107+
help="Comma-separated list of selected metrics to report percentiles. "
11081108
"This argument specifies the metrics to report percentiles. "
11091109
'Allowed metric names are "ttft", "tpot", "itl", "e2el". '
11101110
'Default value is "ttft,tpot,itl".',

benchmarks/benchmark_serving_structured_output.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -998,7 +998,7 @@ def create_argument_parser():
998998
"--percentile-metrics",
999999
type=str,
10001000
default="ttft,tpot,itl",
1001-
help="Comma-separated list of selected metrics to report percentils. "
1001+
help="Comma-separated list of selected metrics to report percentiles. "
10021002
"This argument specifies the metrics to report percentiles. "
10031003
'Allowed metric names are "ttft", "tpot", "itl", "e2el". '
10041004
'Default value is "ttft,tpot,itl".',

benchmarks/benchmark_throughput.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -719,7 +719,7 @@ def create_argument_parser():
719719
"[length * (1 - range_ratio), length * (1 + range_ratio)].",
720720
)
721721

722-
# hf dtaset
722+
# hf dataset
723723
parser.add_argument(
724724
"--hf-subset", type=str, default=None, help="Subset of the HF dataset."
725725
)

tools/profiler/visualize_layerwise_profile.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -119,7 +119,7 @@ def all_the_same(items) -> bool:
119119
if not all_the_same(trace_eles)), None)
120120

121121
if first_trace_difference is None:
122-
# can't create a unique name, leave them names as the
122+
# can't create a unique name, leave the names as they
123123
# are they will get aggregated by the pivot_table call
124124
continue
125125

vllm/compilation/collective_fusion.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -513,7 +513,7 @@ def call_trtllm_fused_allreduce_norm(
513513
torch.ops._C.static_scaled_fp8_quant(
514514
quant_out, norm_out, scale_factor)
515515
if scale_factor is None or norm_out is not None:
516-
# we need to return allreduce outpput
516+
# we need to return allreduce output
517517
# in cases of non quant fused AR + RMS norm
518518
# and fused AR + RMS norm + quant without fused add
519519
allreduce_in.copy_(allreduce_out)

vllm/engine/multiprocessing/engine.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ class MQLLMEngine:
4949
5050
This class is used to wrap the
5151
[`LLMEngine`][vllm.engine.llm_engine.LLMEngine] class to enable use
52-
in concurrnet manner. It runs a background loop and uses zeromq to
52+
in concurrent manner. It runs a background loop and uses zeromq to
5353
receive new requests and stream outputs incrementally via ipc.
5454
5555
The [`LLMEngine`][vllm.engine.llm_engine.LLMEngine] generate or encode

vllm/model_executor/layers/quantization/utils/w8a8_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@
2323
# The condition to determine if it is on a platform that supports
2424
# torch._scaled_mm rowwise feature.
2525
# The condition is determined once as the operations
26-
# are time consuming.
26+
# are time-consuming.
2727
USE_ROWWISE_TORCH_SCALED_MM = (current_platform.is_rocm() and version.parse(
2828
torch.__version__) >= version.parse("2.7")
2929
and current_platform.has_device_capability(94))

vllm/model_executor/model_loader/default_loader.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -211,7 +211,7 @@ def _get_weights_iterator(
211211

212212
if not USE_TPU_COMMONS:
213213
# In PyTorch XLA, we should call `xm.mark_step`
214-
# requently so that not too many ops are accumulated
214+
# frequently so that not too many ops are accumulated
215215
# in the XLA program. import torch_xla.core.xla_model
216216
# as xm
217217
import torch_xla.core.xla_model as xm

vllm/v1/worker/xpu_worker.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ def determine_available_memory(self) -> int:
8484
"""Profiles the peak memory usage of the model to determine how many
8585
KV blocks may be allocated without OOMs.
8686
The engine will first conduct a profiling of the existing memory usage.
87-
Then, it calculate the maximum possible number of GPU and CPU blocks
87+
Then, it calculates the maximum possible number of GPU and CPU blocks
8888
that can be allocated with the remaining free memory.
8989
.. tip::
9090
You may limit the usage of GPU memory

0 commit comments

Comments
 (0)