Skip to content

Commit 559da40

Browse files
authored
[llm-d] Keep working (#912)
<!-- This is an auto-generated comment: release notes by coderabbit.ai --> ## Summary by CodeRabbit * **Chores** * Removed InferenceService resource capture from the state capture process. InferenceService resources will no longer be included in state snapshots. <!-- end of auto-generated comment: release notes by coderabbit.ai -->
2 parents c64063b + 974c4ca commit 559da40

File tree

3 files changed

+32
-10
lines changed

3 files changed

+32
-10
lines changed

projects/llm-d/toolbox/llmd_capture_isvc_state/tasks/main.yml

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -40,13 +40,6 @@
4040
-oyaml > "{{ artifact_extra_logs_dir }}/artifacts/llminferenceservice.pods.yaml"
4141
ignore_errors: true
4242

43-
- name: Capture regular InferenceServices that may be created
44-
shell:
45-
oc get inferenceservice \
46-
-n "{{ target_namespace }}" \
47-
-oyaml > "{{ artifact_extra_logs_dir }}/artifacts/inferenceservices.yaml"
48-
ignore_errors: true
49-
5043
- name: Capture deployments related to the LLMInferenceService
5144
shell:
5245
oc get deployments \

projects/llm-d/visualizations/llmd_inference/plotting/error_report.py

Lines changed: 29 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -103,6 +103,35 @@ def do_plot(self, *args):
103103
if hasattr(results, 'test_name'):
104104
header.append(html.P(f"Test name: {results.test_name}"))
105105

106+
# Show test entry labels
107+
labels_list = []
108+
109+
for key, value in entry.settings.__dict__.items():
110+
labels_list.append(html.Span(f"{key}={value}", style={
111+
"background-color": "#f0f0f0",
112+
"padding": "2px 6px",
113+
"margin": "2px",
114+
"border-radius": "3px",
115+
"font-family": "monospace",
116+
"font-size": "0.9em"
117+
}))
118+
labels_list.append(" ")
119+
120+
header.append(html.P([
121+
"Settings: ",
122+
html.Span(labels_list)
123+
]))
124+
125+
header.append(html.P([
126+
"Directory: ",
127+
html.Code(str(entry.location).strip("./"), style={
128+
"background-color": "#f8f8f8",
129+
"padding": "2px 4px",
130+
"border-radius": "3px",
131+
"font-size": "0.9em"
132+
})
133+
]))
134+
106135
if hasattr(results, 'test_failure_reason') and results.test_failure_reason:
107136
header.append(html.P([
108137
"Failure reason: ",

projects/llm-d/visualizations/llmd_inference/store/parsers.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -343,9 +343,9 @@ def get_metric_value(metric_name, stat_type='median', default=0.0):
343343
tpot_p95 = get_metric_value('time_per_output_token_ms', 'p95') / 1000.0
344344

345345
# Extract throughput metrics
346-
request_rate = get_metric_value('requests_per_second', 'median')
347-
input_tokens_per_second = get_metric_value('input_tokens_per_second', 'median')
348-
output_tokens_per_second = get_metric_value('output_tokens_per_second', 'median')
346+
request_rate = get_metric_value('requests_per_second', 'mean')
347+
input_tokens_per_second = get_metric_value('input_tokens_per_second', 'mean')
348+
output_tokens_per_second = get_metric_value('output_tokens_per_second', 'mean')
349349
total_tokens_per_second = input_tokens_per_second + output_tokens_per_second
350350

351351
# Calculate requests completed and tokens per request

0 commit comments

Comments
 (0)