Skip to content

Commit 166228b

Browse files
didier-durandeicherseiji
authored andcommitted
[Doc]: fix typos in Python comments (vllm-project#24026)
Signed-off-by: Didier Durand <[email protected]>
1 parent 8116399 commit 166228b

File tree

14 files changed

+17
-17
lines changed

14 files changed

+17
-17
lines changed

examples/offline_inference/multilora_inference.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ def create_test_prompts(
2323
2 requests for base model, 4 requests for the LoRA. We define 2
2424
different LoRA adapters (using the same model for demo purposes).
2525
Since we also set `max_loras=1`, the expectation is that the requests
26-
with the second LoRA adapter will be ran after all requests with the
26+
with the second LoRA adapter will be run after all requests with the
2727
first adapter have finished.
2828
"""
2929
return [

vllm/distributed/device_communicators/pynccl.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ def __init__(
3131
group: the process group to work on. If None, it will use the
3232
default process group.
3333
device: the device to bind the PyNcclCommunicator to. If None,
34-
it will be bind to f"cuda:{local_rank}".
34+
it will be bound to f"cuda:{local_rank}".
3535
library_path: the path to the NCCL library. If None, it will
3636
use the default library path.
3737
It is the caller's responsibility to make sure each communicator

vllm/distributed/parallel_state.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -939,8 +939,8 @@ def get_pipeline_model_parallel_group():
939939
def graph_capture(device: torch.device):
940940
"""
941941
`graph_capture` is a context manager which should surround the code that
942-
is capturing the CUDA graph. Its main purpose is to ensure that the
943-
some operations will be run after the graph is captured, before the graph
942+
is capturing the CUDA graph. Its main purpose is to ensure that some
943+
operations will be run after the graph is captured, before the graph
944944
is replayed. It returns a `GraphCaptureContext` object which contains the
945945
necessary data for the graph capture. Currently, it only contains the
946946
stream that the graph capture is running on. This stream is set to the

vllm/entrypoints/openai/tool_parsers/pythonic_tool_parser.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -165,7 +165,7 @@ def extract_tool_calls_streaming(
165165
index] += delta.function.arguments
166166

167167
# HACK: serving_chat.py inspects the internal state of tool parsers
168-
# when determining it's final streaming delta, automatically
168+
# when determining its final streaming delta, automatically
169169
# adding autocompleted JSON.
170170
# These two lines avoid that nonsense while ensuring finish_reason
171171
# is set to tool_calls when at least one tool is called.

vllm/model_executor/layers/fused_moe/moe_pallas.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@
77

88
def _histogram(input: torch.Tensor, min: int, max: int) -> torch.Tensor:
99
"""
10-
Compute the histogram of a int32 tensor. The bin edges are defined by the
10+
Compute the histogram of an int32 tensor. The bin edges are defined by the
1111
min and max values, with step = 1.
1212
"""
1313
assert input.dtype == torch.int32, "input must be of torch.int32 dtype."

vllm/model_executor/models/ovis.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -544,7 +544,7 @@ def forward(
544544
vision_embeddings)
545545
input_ids = None
546546

547-
# up until here we have a inputs_embeds 100% numerical identity
547+
# up until here we have an inputs_embeds 100% numerical identity
548548
# between the OG HF Transformers implementation and ours
549549
hidden_states = self.llm(
550550
input_ids=input_ids,

vllm/model_executor/models/phi4mm_audio.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ class ConformerEncoderLayer(nn.Module):
4343
if set different to 0, the number of
4444
depthwise_seperable_out_channel will be used as a
4545
channel_out of the second conv1d layer.
46-
otherwise, it equal to 0, the second conv1d layer is skipped.
46+
otherwise, it equals to 0, the second conv1d layer is skipped.
4747
depthwise_multiplier: int
4848
number of input_dim channels duplication. this value
4949
will be used to compute the hidden channels of the Conv1D.
@@ -115,7 +115,7 @@ class ConformerEncoderLayer(nn.Module):
115115
we recalculate activation in backward.
116116
default "".
117117
export: bool, optional
118-
if set to True, it remove the padding from convolutional layers
118+
if set to True, it removes the padding from convolutional layers
119119
and allow the onnx conversion for inference.
120120
default False.
121121
use_pt_scaled_dot_product_attention: bool, optional
@@ -686,7 +686,7 @@ class ConformerEncoder(TransformerEncoderBase):
686686
only work for glu_in_attention !=0
687687
default "swish".
688688
export: bool, optional
689-
if set to True, it remove the padding from convolutional layers
689+
if set to True, it removes the padding from convolutional layers
690690
and allow the onnx conversion for inference.
691691
default False.
692692
activation_checkpointing: str, optional

vllm/model_executor/models/phi4mm_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -258,7 +258,7 @@ class DepthWiseSeperableConv1d(nn.Module):
258258
if set different to 0, the number of
259259
depthwise_seperable_out_channel will be used as a channel_out
260260
of the second conv1d layer.
261-
otherwise, it equal to 0, the second conv1d layer is skipped.
261+
otherwise, it equals to 0, the second conv1d layer is skipped.
262262
kernel_size: int
263263
kernel_size
264264
depthwise_multiplier: int

vllm/third_party/pynvml.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1022,7 +1022,7 @@ def _extractNVMLErrorsAsClasses():
10221022
Each NVML Error gets a new NVMLError subclass. This way try,except blocks can filter appropriate
10231023
exceptions more easily.
10241024
1025-
NVMLError is a parent class. Each NVML_ERROR_* gets it's own subclass.
1025+
NVMLError is a parent class. Each NVML_ERROR_* gets its own subclass.
10261026
e.g. NVML_ERROR_ALREADY_INITIALIZED will be turned into NVMLError_AlreadyInitialized
10271027
'''
10281028
this_module = sys.modules[__name__]

vllm/transformers_utils/configs/nemotron.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626
class NemotronConfig(PretrainedConfig):
2727
r"""
2828
This is the configuration class to store the configuration of a
29-
[`NemotronModel`]. It is used to instantiate an Nemotron model
29+
[`NemotronModel`]. It is used to instantiate a Nemotron model
3030
according to the specified arguments, defining the model architecture.
3131
Instantiating a configuration with the defaults will yield a similar
3232
configuration to that of the Nemotron-8B.

0 commit comments

Comments
 (0)