Skip to content

Commit bcb06d7

Browse files
[Doc]: fix typos in various files (vllm-project#24726)
Signed-off-by: Didier Durand <[email protected]>
1 parent 0377802 commit bcb06d7

File tree

11 files changed

+11
-11
lines changed

11 files changed

+11
-11
lines changed

benchmarks/kernels/benchmark_w8a8_block_fp8.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -56,7 +56,7 @@ def w8a8_block_matmul(
5656
Bs: The per-block quantization scale for `B`.
5757
block_size: The block size for per-block quantization.
5858
It should be 2-dim, e.g., [128, 128].
59-
output_dytpe: The dtype of the returned tensor.
59+
output_dtype: The dtype of the returned tensor.
6060
6161
Returns:
6262
torch.Tensor: The result of matmul.

csrc/cpu/cpu_types_vxe.hpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ namespace vec_op {
1212
#define vec_sub(a, b) ((a) - (b))
1313
#define vec_mul(a, b) ((a) * (b))
1414
#define vec_div(a, b) ((a) / (b))
15-
#define vec_sr(a, b) ((a) >> (b)) // Vector Shift Right Algebaic
15+
#define vec_sr(a, b) ((a) >> (b)) // Vector Shift Right Algebraic
1616
#define vec_sl(a, b) ((a) << (b)) // Vector Shift Left
1717

1818
// FIXME: FP16 is not fully supported in Torch-CPU

csrc/cpu/sgl-kernels/moe.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -215,7 +215,7 @@ int moe_align_block_size(
215215
offsets[mb + 1] = sorted_id_size(sorted_ids + mb * BLOCK_M);
216216
}
217217
});
218-
// TODO: do we need to vecterize this ?
218+
// TODO: do we need to vectorize this ?
219219
for (int mb = 0; mb < num_token_blocks; ++mb) {
220220
offsets[mb + 1] += offsets[mb];
221221
}

docs/design/multiprocessing.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ page for information on known issues and how to solve them.
88
## Introduction
99

1010
!!! important
11-
The source code references are to the state of the code at the time of writing in December, 2024.
11+
The source code references are to the state of the code at the time of writing in December 2024.
1212

1313
The use of Python multiprocessing in vLLM is complicated by:
1414

vllm/attention/backends/flash_attn.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -901,7 +901,7 @@ def _get_query_key_seq_metadata(
901901
attn_metadata.encoder_seq_start_loc,
902902
attn_metadata.max_encoder_seq_len)
903903
elif attn_type == AttentionType.ENCODER:
904-
# For encoder attention both the query and the key are same i.e the
904+
# For encoder attention both the query and the key are same i.e. the
905905
# encoder sequence.
906906
return (attn_metadata.encoder_seq_start_loc,
907907
attn_metadata.max_encoder_seq_len,

vllm/benchmarks/datasets.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -551,7 +551,7 @@ def generate_token_sequence(
551551
[6880, 6881] -> ['Ġcalls', 'here'] ->
552552
[1650, 939, 486] -> ['Ġcall', 'sh', 'ere']
553553
To avoid uncontrolled change of the prompt length,
554-
the encoded sequence is truncated before being decode again.
554+
the encoded sequence is truncated before being decoded again.
555555
"""
556556
# Build the inner sequence by sampling sequentially from the vocab
557557
inner_seq = ((offset + index + np.arange(input_len))

vllm/entrypoints/openai/protocol.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -242,7 +242,7 @@ def get_logits_processors(processors: Optional[LogitsProcessors],
242242
elif processors:
243243
raise ValueError(
244244
"The `logits_processors` argument is not supported by this "
245-
"server. See --logits-processor-pattern engine argugment "
245+
"server. See --logits-processor-pattern engine argument "
246246
"for more information.")
247247
return None
248248

vllm/model_executor/layers/mamba/mamba_mixer2.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -324,7 +324,7 @@ def __init__(self,
324324
# - the weight already has a "weight_loader" attribute
325325
# which set_weight_attrs will raise if we do not
326326
# delete before trying to override it
327-
# - ditto for the otther two weights below
327+
# - ditto for the other two weights below
328328
delattr(self.conv1d.bias, "weight_loader")
329329
set_weight_attrs(
330330
self.conv1d.bias,

vllm/model_executor/models/minicpmv.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1117,7 +1117,7 @@ def _process_vision_input(
11171117

11181118
def _process_multimodal_inputs(self, modalities: dict):
11191119
# The result multimodal_embeddings is tuple of tensors, with each
1120-
# tensor correspoending to a multimodal data item (image or video).
1120+
# tensor corresponding to a multimodal data item (image or video).
11211121
multimodal_embeddings: tuple[torch.Tensor, ...] = ()
11221122

11231123
# NOTE: It is important to iterate over the keys in this dictionary

vllm/v1/worker/gpu_model_runner.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2659,7 +2659,7 @@ def _dummy_run(
26592659
num_tokens += num_pad
26602660

26612661
# If cudagraph_mode.decode_mode() == FULL and
2662-
# cudagraph_mode.seperate_routine(). This means that we are using
2662+
# cudagraph_mode.separate_routine(). This means that we are using
26632663
# different graphs and/or modes for mixed prefill-decode batches vs.
26642664
# uniform decode batches. A uniform decode batch means that all
26652665
# requests have identical query length, except a potential virtual

0 commit comments

Comments
 (0)