Skip to content

Commit c4830ef

Browse files
authored
fix typos (#4176)
* fix typos * fix
1 parent 0b62648 commit c4830ef

File tree

10 files changed

+34
-50
lines changed

10 files changed

+34
-50
lines changed

benchmarks/benchmark_mtp.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -98,7 +98,7 @@ def main(args):
9898
raise ValueError("--max_concurrency should be same length as --s_itl_base_model")
9999

100100
for max_concurrency, s_itl in zip(args.max_concurrency, args.s_itl_base_model):
101-
# Wramup
101+
# Warmup
102102
print("Starting warmup...")
103103
with open(os.devnull, "w") as f:
104104
with contextlib.redirect_stdout(f):

custom_ops/gpu_ops/custom_all_reduce/all_reduce.cuh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -303,7 +303,7 @@ class CustomAllreduce {
303303
bool full_nvlink_;
304304

305305
RankSignals sg_;
306-
// Stores an map from a pointer to its peer pointters from all ranks.
306+
// Stores an map from a pointer to its peer pointers from all ranks.
307307
std::unordered_map<void*, RankData*> buffers_;
308308
Signal* self_sg_;
309309

docs/get_started/installation/nvidia_gpu.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ The following installation methods are available when your environment meets the
1010

1111
## 1. Pre-built Docker Installation (Recommended)
1212

13-
**Notice**: The pre-built image only supports SM80/90 GPU(e.g. H800/A800),if you are deploying on SM86/89GPU(L40/4090/L20), please reinstall ```fastdpeloy-gpu``` after you create the container.
13+
**Notice**: The pre-built image only supports SM80/90 GPU(e.g. H800/A800),if you are deploying on SM86/89GPU(L40/4090/L20), please reinstall ```fastdeploy-gpu``` after you create the container.
1414

1515
```shell
1616
docker pull ccr-2vdh3abv-pub.cnc.bj.baidubce.com/paddlepaddle/fastdeploy-cuda-12.6:2.2.0

docs/usage/code_overview.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,6 @@ Below is an overview of the FastDeploy code structure and functionality organize
2020
- ```platforms```: Platform-specific modules for underlying hardware support.
2121
- ```scheduler```: Request scheduling module for large models.
2222
- ```metrics```: Core component for collecting, managing, and exporting Prometheus metrics, tracking key runtime performance data (e.g., request latency, resource utilization, successful request counts).
23-
- ```splitwise```: Modules related to PD disaggragation deployment.
23+
- ```splitwise```: Modules related to PD disaggregation deployment.
2424
- ```scripts```/```tools```: Utility scripts for FastDeploy operations (e.g., compilation, unit testing, code style fixes).
2525
- ```test```: Code for unit testing and validation.

docs/usage/log.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ By default, logs are stored in the `log` directory under the execution path. To
3030
* `cache_transfer_manager.log` : Logs startup parameters and received request information.
3131
* `launch_cache_manager.log` : Records cache transfer startup parameters and error messages.
3232

33-
## PD Disaggragation Logs
33+
## PD Disaggregation Logs
3434
* `cache_messager.log` : Logs transmission protocols and messages used by the P instance.
3535
* `splitwise_connector.log` : Records data received from P/D instances and connection establishment details.
3636

mkdocs.yml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
site_name: 'FastDeploy : Large Language Model Deployement'
1+
site_name: 'FastDeploy : Large Language Model Deployment'
22
repo_url: https://github.com/PaddlePaddle/FastDeploy
33
repo_name: FastDeploy
44

@@ -36,7 +36,7 @@ plugins:
3636
- locale: en
3737
default: true
3838
name: English
39-
site_name: 'FastDeploy: Large Language Model Deployement'
39+
site_name: 'FastDeploy: Large Language Model Deployment'
4040
build: true
4141
link: /FastDeploy/
4242
- locale: zh
@@ -59,7 +59,7 @@ plugins:
5959
ERNIE-4.5-VL-424B-A47B: ERNIE-4.5-VL-424B-A47B快速部署
6060
Quick Deployment For QWEN: Qwen3-0.6b快速部署
6161
Online Serving: 在线服务
62-
OpenAI-Compitable API Server: 兼容 OpenAI 协议的服务化部署
62+
OpenAI-Compatible API Server: 兼容 OpenAI 协议的服务化部署
6363
Monitor Metrics: 监控Metrics
6464
Scheduler: 调度器
6565
Graceful Shutdown: 服务优雅关闭
@@ -114,7 +114,7 @@ nav:
114114
- ERNIE-4.5-VL-424B-A47B: get_started/ernie-4.5-vl.md
115115
- Quick Deployment For QWEN: get_started/quick_start_qwen.md
116116
- Online Serving:
117-
- OpenAI-Compitable API Server: online_serving/README.md
117+
- OpenAI-Compatible API Server: online_serving/README.md
118118
- Monitor Metrics: online_serving/metrics.md
119119
- Scheduler: online_serving/scheduler.md
120120
- Graceful Shutdown: online_serving/graceful_shutdown_service.md

tests/ce/server/test_return_token_ids.py

Lines changed: 17 additions & 33 deletions
Original file line numberDiff line numberDiff line change
@@ -9,12 +9,7 @@
99

1010
import json
1111

12-
from core import (
13-
TEMPLATE,
14-
URL,
15-
build_request_payload,
16-
send_request,
17-
)
12+
from core import TEMPLATE, URL, build_request_payload, send_request
1813

1914
COMPLETIONS_URL = URL.replace("/v1/chat/completions", "/v1/completions")
2015

@@ -29,17 +24,17 @@ def test_completion_stream_text_after_process_raw_prediction():
2924
"stream": True,
3025
"stream_options": {"include_usage": True, "continuous_usage_stats": True},
3126
"max_tokens": 50,
32-
"return_token_ids": True
27+
"return_token_ids": True,
3328
}
34-
29+
3530
payload = build_request_payload(TEMPLATE, data)
3631
resp = send_request(COMPLETIONS_URL, payload, stream=True)
3732
for line in resp.iter_lines(decode_unicode=True):
3833
if line.strip() == "data: [DONE]":
3934
break
4035
if line.strip() == "" or not line.startswith("data: "):
4136
continue
42-
line = line[len("data: "):]
37+
line = line[len("data: ") :]
4338
response_data = json.loads(line)
4439

4540
choice = response_data["choices"][0]
@@ -51,21 +46,16 @@ def test_completion_stream_text_after_process_raw_prediction():
5146
reasoning_content = choice["reasoning_content"]
5247
text = choice["text"]
5348
assert reasoning_content or text in raw_prediction, "raw_prediction取值结果不正确"
54-
if "finish_reason" in line.strip() :
49+
if "finish_reason" in line.strip():
5550
break
5651

57-
58-
def test_completion_text_after_process_raw_predictio_return_tokrn_ids():
52+
53+
def test_completion_text_after_process_raw_predictio_return_token_ids():
5954
"""
6055
/v1/completions接口,非流式接口
6156
返回属性"text_after_process"和"reasoning_content"
6257
"""
63-
data = {
64-
"stream": False,
65-
"prompt": "你是谁",
66-
"max_tokens": 50,
67-
"return_token_ids": True
68-
}
58+
data = {"stream": False, "prompt": "你是谁", "max_tokens": 50, "return_token_ids": True}
6959
payload = build_request_payload(TEMPLATE, data)
7060
resp = send_request(COMPLETIONS_URL, payload).json()
7161

@@ -80,14 +70,10 @@ def test_completion_text_after_process_raw_predictio_return_tokrn_ids():
8070

8171
def test_completion_text_after_process_raw_prediction():
8272
"""
83-
/v1/completions接口,无return_tokrn_ids参数
73+
/v1/completions接口,无return_token_ids参数
8474
非流式接口中,无return token ids 属性"text_after_process"和"reasoning_content"值为null
8575
"""
86-
data = {
87-
"stream": False,
88-
"prompt": "你是谁",
89-
"max_tokens": 50
90-
}
76+
data = {"stream": False, "prompt": "你是谁", "max_tokens": 50}
9177
payload = build_request_payload(TEMPLATE, data)
9278
resp = send_request(COMPLETIONS_URL, payload).json()
9379

@@ -108,17 +94,17 @@ def test_stream_text_after_process_raw_prediction():
10894
"stream": True,
10995
"stream_options": {"include_usage": True, "continuous_usage_stats": True},
11096
"max_tokens": 50,
111-
"return_token_ids": True
97+
"return_token_ids": True,
11298
}
11399

114100
payload = build_request_payload(TEMPLATE, data)
115101
resp = send_request(URL, payload, stream=True)
116102
for line in resp.iter_lines(decode_unicode=True):
117-
if line.strip() == "data: [DONE]" :
103+
if line.strip() == "data: [DONE]":
118104
break
119105
if line.strip() == "" or not line.startswith("data: "):
120106
continue
121-
line = line[len("data: "):]
107+
line = line[len("data: ") :]
122108
response_data = json.loads(line)
123109

124110
choice = response_data["choices"][0]
@@ -130,11 +116,11 @@ def test_stream_text_after_process_raw_prediction():
130116
reasoning_content = choice["delta"]["reasoning_content"]
131117
content = choice["delta"]["content"]
132118
assert reasoning_content or content in raw_prediction, "raw_prediction取值结果不正确"
133-
if "finish_reason" in line.strip() :
119+
if "finish_reason" in line.strip():
134120
break
135121

136-
137-
def test_text_after_process_raw_prediction_return_tokrn_ids():
122+
123+
def test_text_after_process_raw_prediction_return_token_ids():
138124
"""
139125
/v1/chat/completions接口,非流式接口
140126
返回属性"text_after_process"和"reasoning_content"
@@ -161,7 +147,7 @@ def test_text_after_process_raw_prediction_return_tokrn_ids():
161147

162148
def test_text_after_process_raw_prediction():
163149
"""
164-
/v1/chat/completions接口,无return_tokrn_ids参数
150+
/v1/chat/completions接口,无return_token_ids参数
165151
无return token ids 属性"text_after_process"和"reasoning_content"值为null
166152
"""
167153
data = {
@@ -179,5 +165,3 @@ def test_text_after_process_raw_prediction():
179165

180166
raw_prediction = resp["choices"][0]["message"]["raw_prediction"]
181167
assert raw_prediction is None, "raw_prediction取值结果不正确"
182-
183-

tests/ci_use/EB_Lite_with_adapter/zmq_client.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ def consume_results(self, result_queue):
5050
if self.need_exit:
5151
break
5252
except Exception as e:
53-
print(f"zmq client occured error {e} type: {type(e)} frames: {frames}")
53+
print(f"zmq client occurred error {e} type: {type(e)} frames: {frames}")
5454

5555
def start(self, result_queue):
5656
threading.Thread(target=self.consume_results, args=(result_queue,), daemon=True).start()
@@ -118,4 +118,4 @@ def recv_results(self):
118118
self.result[task_id] = result["result"]
119119
self.task_event[task_id].set()
120120
except Exception as e:
121-
print(f"zmq client occured error {e} type: {type(e)} frames: {frames}")
121+
print(f"zmq client occurred error {e} type: {type(e)} frames: {frames}")

tests/distributed/custom_all_reduce.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -54,14 +54,14 @@ def test_case(self):
5454
fa = CustomAllreduce(model_parallel_group)
5555

5656
for m, n in mns:
57-
data_cusom_ar = paddle.rand([m, n], dtype="bfloat16")
58-
data_paddle = data_cusom_ar.clone()
59-
if fa.should_custom_ar(data_cusom_ar):
60-
fa.custom_all_reduce(data_cusom_ar)
57+
data_custom_ar = paddle.rand([m, n], dtype="bfloat16")
58+
data_paddle = data_custom_ar.clone()
59+
if fa.should_custom_ar(data_custom_ar):
60+
fa.custom_all_reduce(data_custom_ar)
6161
dist.all_reduce(data_paddle)
6262
if dist.get_rank() == 0:
6363
np.testing.assert_allclose(
64-
data_cusom_ar.numpy(),
64+
data_custom_ar.numpy(),
6565
data_paddle.numpy(),
6666
rtol=1e-04,
6767
atol=1e-04,

tools/deep_gemm_pre-compile/pre_compile.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@ def pre_compile_from_config(config_file: str, num_threads: int, expert_parallel:
158158

159159
pbar.close()
160160

161-
logger.info(f"Total compliation time: {time() - start_time:.2f} seconds")
161+
logger.info(f"Total compilation time: {time() - start_time:.2f} seconds")
162162

163163

164164
def main(args):

0 commit comments

Comments
 (0)