Skip to content

Commit 5c82ecf

Browse files
Update generated content (#242)
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: mathieu-benoit <11720844+mathieu-benoit@users.noreply.github.com>
1 parent f505d19 commit 5c82ecf

File tree

7 files changed

+75
-8
lines changed

7 files changed

+75
-8
lines changed
Lines changed: 26 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,26 @@
1+
---
2+
title: "dmr-llm-model-for-openai-clients"
3+
draft: false
4+
mermaid: true
5+
type: examples
6+
source: "community"
7+
implementation: "score-compose"
8+
resourceType: "llm-model"
9+
provisionerType: "template"
10+
flavor: "dmr"
11+
excerpt: ''
12+
description: 'Generates the LLM model via the Docker Model Runner (DMR). Outputs the url for OpenAI SDK/clients.'
13+
expectedOutputs:
14+
- model
15+
- url
16+
- api-key
17+
supportedParams:
18+
- model
19+
- context_size
20+
hasMore: false
21+
22+
---
23+
24+
{{% resource-provisioner-content description="Generates the LLM model via the Docker Model Runner (DMR). Outputs the url for OpenAI SDK/clients." type="llm-model" supportedParams="model,context_size" expectedOutputs="model,url,api-key" %}}
25+
26+
{{% example-file filename="10-dmr-llm-model-for-openai-clients.provisioners.yaml" dir="resource-provisioners/community/llm-model/score-compose" githubUrl="https://github.com/score-spec/community-provisioners/blob/main" %}}

content/en/examples/resource-provisioners/community/llm-model/score-compose/template/dmr-llm-model.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ resourceType: "llm-model"
99
provisionerType: "template"
1010
flavor: "dmr"
1111
excerpt: ''
12-
description: 'Generates the LLM model via the Docker Model Runner (DMR).'
12+
description: 'Generates the LLM model via the Docker Model Runner (DMR). Outputs the url for Ollama-compatible clients.'
1313
expectedOutputs:
1414
- model
1515
- url
@@ -21,6 +21,6 @@ hasMore: false
2121

2222
---
2323

24-
{{% resource-provisioner-content description="Generates the LLM model via the Docker Model Runner (DMR)." type="llm-model" supportedParams="model,context_size" expectedOutputs="model,url,api-key" %}}
24+
{{% resource-provisioner-content description="Generates the LLM model via the Docker Model Runner (DMR). Outputs the url for Ollama-compatible clients." type="llm-model" supportedParams="model,context_size" expectedOutputs="model,url,api-key" %}}
2525

2626
{{% example-file filename="10-dmr-llm-model.provisioners.yaml" dir="resource-provisioners/community/llm-model/score-compose" githubUrl="https://github.com/score-spec/community-provisioners/blob/main" %}}

content/en/examples/score/resources/community-provisioners/llm-model.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -14,6 +14,7 @@ flavor: "Resources"
1414

1515
{{< tabs >}}
1616
{{% tab name="score-compose" %}}
17+
{{% example-file filename="10-dmr-llm-model-for-openai-clients.provisioners.yaml" dir="score/resources/community-provisioners/llm-model/score-compose" githubUrl="https://github.com/score-spec/community-provisioners/blob/main" %}}
1718
{{% example-file filename="10-dmr-llm-model-via-curl-cmd.provisioners.yaml" dir="score/resources/community-provisioners/llm-model/score-compose" githubUrl="https://github.com/score-spec/community-provisioners/blob/main" %}}
1819
{{% example-file filename="10-dmr-llm-model-via-curl-service.provisioners.yaml" dir="score/resources/community-provisioners/llm-model/score-compose" githubUrl="https://github.com/score-spec/community-provisioners/blob/main" %}}
1920
{{% example-file filename="10-dmr-llm-model-via-service-provider.provisioners.yaml" dir="score/resources/community-provisioners/llm-model/score-compose" githubUrl="https://github.com/score-spec/community-provisioners/blob/main" %}}
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,20 @@
1+
- uri: template://community-provisioners/dmr-llm-model-for-openai-clients
2+
type: llm-model
3+
description: Generates the LLM model via the Docker Model Runner (DMR). Outputs the url for OpenAI SDK/clients.
4+
supported_params:
5+
- model
6+
- context_size
7+
outputs: |
8+
model: {{ .Init.model }}
9+
url: "http://172.17.0.1:12434/engines/v1/"
10+
api-key: "not-needed"
11+
expected_outputs:
12+
- model
13+
- url
14+
- api-key
15+
init: |
16+
model: {{ .Params.model | default "ai/smollm2:135M-Q4_0" }}
17+
models: |
18+
{{ .Id }}:
19+
model: {{ .Init.model }}
20+
context_size: {{ .Params.context_size | default 2048 }}
Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
11
- uri: template://community-provisioners/dmr-llm-model
22
type: llm-model
3-
description: Generates the LLM model via the Docker Model Runner (DMR).
3+
description: Generates the LLM model via the Docker Model Runner (DMR). Outputs the url for Ollama-compatible clients.
44
supported_params:
55
- model
66
- context_size
77
outputs: |
88
model: {{ .Init.model }}
9-
url: "http://172.17.0.1:12434/engines/v1/"
9+
url: "http://172.17.0.1:12434/"
1010
api-key: "not-needed"
1111
expected_outputs:
1212
- model
@@ -17,4 +17,4 @@
1717
models: |
1818
{{ .Id }}:
1919
model: {{ .Init.model }}
20-
context_size: {{ .Params.context_size | default 2048 }}
20+
context_size: {{ .Params.context_size | default 2048 }}
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,20 @@
1+
- uri: template://community-provisioners/dmr-llm-model-for-openai-clients
2+
type: llm-model
3+
description: Generates the LLM model via the Docker Model Runner (DMR). Outputs the url for OpenAI SDK/clients.
4+
supported_params:
5+
- model
6+
- context_size
7+
outputs: |
8+
model: {{ .Init.model }}
9+
url: "http://172.17.0.1:12434/engines/v1/"
10+
api-key: "not-needed"
11+
expected_outputs:
12+
- model
13+
- url
14+
- api-key
15+
init: |
16+
model: {{ .Params.model | default "ai/smollm2:135M-Q4_0" }}
17+
models: |
18+
{{ .Id }}:
19+
model: {{ .Init.model }}
20+
context_size: {{ .Params.context_size | default 2048 }}
Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
11
- uri: template://community-provisioners/dmr-llm-model
22
type: llm-model
3-
description: Generates the LLM model via the Docker Model Runner (DMR).
3+
description: Generates the LLM model via the Docker Model Runner (DMR). Outputs the url for Ollama-compatible clients.
44
supported_params:
55
- model
66
- context_size
77
outputs: |
88
model: {{ .Init.model }}
9-
url: "http://172.17.0.1:12434/engines/v1/"
9+
url: "http://172.17.0.1:12434/"
1010
api-key: "not-needed"
1111
expected_outputs:
1212
- model
@@ -17,4 +17,4 @@
1717
models: |
1818
{{ .Id }}:
1919
model: {{ .Init.model }}
20-
context_size: {{ .Params.context_size | default 2048 }}
20+
context_size: {{ .Params.context_size | default 2048 }}

0 commit comments

Comments
 (0)