Skip to content

Commit fad73be

Browse files
[Doc]: fix typos in Python comments (vllm-project#24077)
Signed-off-by: Didier Durand <durand.didier@gmail.com>
1 parent 56d0408 commit fad73be

File tree

14 files changed

+19
-19
lines changed

14 files changed

+19
-19
lines changed

tests/async_engine/test_api_server.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -98,7 +98,7 @@ def test_api_server(api_server, distributed_executor_backend: str):
9898
pool.join()
9999

100100
# check cancellation stats
101-
# give it some times to update the stats
101+
# give it some time to update the stats
102102
time.sleep(1)
103103

104104
num_aborted_requests = requests.get(

tests/core/block/e2e/test_correctness.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -439,10 +439,10 @@ def test_auto_prefix_caching_with_preemption(baseline_llm_generator,
439439
@pytest.mark.parametrize("seed", [1])
440440
def test_auto_prefix_caching_after_eviction_start(baseline_llm_generator,
441441
test_llm_generator):
442-
"""Verify block manager v2 with auto prefix caching could works normal
442+
"""Verify block manager v2 with auto prefix caching could work normally
443443
even when eviction started.
444444
With APC enabled, all blocks are held by native block at the beginning.
445-
Then blocks are managed by evictor instead. If cache hit at the evitor's
445+
Then blocks are managed by evictor instead. If cache hit at the evictor's
446446
block, then it could be reused, or we need to recompute its kv cache.
447447
"""
448448
output_len = 10

tests/engine/test_arg_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -167,7 +167,7 @@ def test_get_kwargs():
167167
# dict should have json tip in help
168168
json_tip = "Should either be a valid JSON string or JSON keys"
169169
assert json_tip in kwargs["json_tip"]["help"]
170-
# nested config should should construct the nested config
170+
# nested config should construct the nested config
171171
assert kwargs["nested_config"]["type"]('{"field": 2}') == NestedConfig(2)
172172

173173

tests/kernels/moe/test_deepep_deepgemm_moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -282,7 +282,7 @@ def triton_impl(a: torch.Tensor, topk_ids: torch.Tensor,
282282
a1_scale=a1_scale,
283283
block_shape=block_shape,
284284
# Make sure this is set to False so we
285-
# dont end up comparing the same implementation.
285+
# don't end up comparing the same implementation.
286286
allow_deep_gemm=False)
287287

288288

tests/lora/test_add_lora.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -59,10 +59,10 @@ async def requests_processing_time(llm,
5959
@pytest.mark.asyncio
6060
async def test_add_lora(chatglm3_lora_files):
6161
"""
62-
The add_lora function is used to pre-load some LoRA adapters into the
62+
The add_lora function is used to preload some LoRA adapters into the
6363
engine in anticipation of future requests using these adapters. To test
6464
this functionality, we use the async engine to process some requests - We
65-
do it twice, once with add_lora() pre-loading and once without.
65+
do it twice, once with add_lora() preloading and once without.
6666
6767
We measure the request processing time in both cases and expect the time
6868
to be lesser in the case with add_lora() calls.

tests/lora/test_lora_allowed_token_ids.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ def test_allowed_token_ids_with_lora_vocab(llama_2_7b_base_huggingface_id,
1818
adapters that define additional tokens.
1919
"""
2020

21-
# Setup a base model compatible with the sql_lora_files adapter and
21+
# Set up a base model compatible with the sql_lora_files adapter and
2222
# a known number of tokens in the base model.
2323
model_config = ModelConfig(
2424
model=llama_2_7b_base_huggingface_id,
@@ -84,7 +84,7 @@ def test_allowed_token_ids_with_lora_adapter_no_vocab(
8484
adapters that do not define additional tokens.
8585
"""
8686

87-
# Setup a base model compatible with the qwen25vl_lora_files adapter and
87+
# Set up a base model compatible with the qwen25vl_lora_files adapter and
8888
# a known number of tokens in the base model.
8989
model_config = ModelConfig(
9090
model=qwen25vl_base_huggingface_id,

tests/models/language/generation/test_common.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
from ...utils import check_logprobs_close
1414

1515
# These have unsupported head_dim for FA. We do not
16-
# not have a clean way to fall back, so we fail with
16+
# have a clean way to fall back, so we fail with
1717
# a clear msg when it happens.
1818
# https://github.com/vllm-project/vllm/issues/14524
1919
REQUIRES_V0 = ["microsoft/phi-2", "stabilityai/stablelm-3b-4e1t"]

tests/models/language/generation/test_mistral.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@
2020
"mistralai/Mistral-7B-Instruct-v0.3",
2121
# uses the v3-Tekken tokenizer
2222
"mistralai/Ministral-8B-Instruct-2410",
23-
# Mistral-Nemo is to big for CI, but passes locally
23+
# Mistral-Nemo is too big for CI, but passes locally
2424
# "mistralai/Mistral-Nemo-Instruct-2407"
2525
]
2626

@@ -273,7 +273,7 @@ def test_mistral_function_calling(vllm_runner, model: str, dtype: str) -> None:
273273

274274

275275
def test_mistral_function_call_nested_json():
276-
"""Ensure that the function-name regex captures the entire outer-most
276+
"""Ensure that the function-name regex captures the entire outermost
277277
JSON block, including nested braces."""
278278

279279
# Create a minimal stub tokenizer that provides the few attributes the

tests/models/multimodal/generation/test_qwen2_vl.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -154,7 +154,7 @@ def get_image_embeds(model):
154154
embed_counter += cur_batch_embed_len
155155
image_counter += cur_batch_image_count
156156

157-
# ensure we don't lost any images or embeddings
157+
# ensure we don't lose any images or embeddings
158158
assert embed_counter == image_embeds.size(0)
159159
assert image_counter == image_grid_thw.size(0)
160160
assert len(image_batches) == len(result)
@@ -238,7 +238,7 @@ def get_image_embeds(model):
238238
embed_counter += cur_batch_embed_len
239239
video_counter += cur_batch_video_count
240240

241-
# ensure we don't lost any videos or embeddings
241+
# ensure we don't lose any videos or embeddings
242242
assert embed_counter == video_embeds.size(0)
243243
assert video_counter == video_grid_thw.size(0)
244244
assert len(video_batches) == len(result)

tests/v1/core/test_kv_cache_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -247,7 +247,7 @@ def test_free_kv_cache_block_queue_append_n():
247247

248248
def test_free_kv_cache_block_queue_popleft_n():
249249
blocks = [KVCacheBlock(block_id=i) for i in range(6)]
250-
# Create a empty FreeKVCacheBlockQueue with these blocks
250+
# Create an empty FreeKVCacheBlockQueue with these blocks
251251
queue = FreeKVCacheBlockQueue(
252252
[blocks[1], blocks[3], blocks[5], blocks[4], blocks[0], blocks[2]])
253253
assert queue.num_free_blocks == 6

0 commit comments

Comments
 (0)