Skip to content

Commit 9ec6b49

Browse files
committed
readme : update [no ci]
1 parent f6dd38c commit 9ec6b49

File tree

1 file changed

+1
-5
lines changed

1 file changed

+1
-5
lines changed

examples/server/README.md

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -149,7 +149,6 @@ The project is under active development, and we are [looking for feedback and co
149149
| `--ssl-cert-file FNAME` | path to file a PEM-encoded SSL certificate<br/>(env: LLAMA_ARG_SSL_CERT_FILE) |
150150
| `-to, --timeout N` | server read/write timeout in seconds (default: 600)<br/>(env: LLAMA_ARG_TIMEOUT) |
151151
| `--threads-http N` | number of threads used to process HTTP requests (default: -1)<br/>(env: LLAMA_ARG_THREADS_HTTP) |
152-
| `-spf, --system-prompt-file FNAME` | set a file to load a system prompt (initial prompt of all slots), this is useful for chat applications |
153152
| `--metrics` | enable prometheus compatible metrics endpoint (default: disabled)<br/>(env: LLAMA_ARG_ENDPOINT_METRICS) |
154153
| `--slots` | enable slots monitoring endpoint (default: disabled)<br/>(env: LLAMA_ARG_ENDPOINT_SLOTS) |
155154
| `--props` | enable changing global properties via POST /props (default: disabled)<br/>(env: LLAMA_ARG_ENDPOINT_PROPS) |
@@ -320,7 +319,6 @@ node index.js
320319

321320
- The prompt is a string or an array with the first element given as a string
322321
- The model's `tokenizer.ggml.add_bos_token` metadata is `true`
323-
- The system prompt is empty
324322

325323
`temperature`: Adjust the randomness of the generated text. Default: `0.8`
326324

@@ -536,14 +534,12 @@ This endpoint is public (no API key check). By default, it is read-only. To make
536534

537535
```json
538536
{
539-
"system_prompt": "",
540537
"default_generation_settings": { ... },
541538
"total_slots": 1,
542539
"chat_template": ""
543540
}
544541
```
545542

546-
- `system_prompt` - the system prompt (initial prompt of all slots). Please note that this does not take into account the chat template. It will append the prompt at the beginning of formatted prompt.
547543
- `default_generation_settings` - the default generation settings for the `/completion` endpoint, which has the same fields as the `generation_settings` response object from the `/completion` endpoint.
548544
- `total_slots` - the total number of slots for process requests (defined by `--parallel` option)
549545
- `chat_template` - the model's original Jinja2 prompt template
@@ -554,7 +550,7 @@ To use this endpoint with POST method, you need to start server with `--props`
554550

555551
*Options:*
556552

557-
- `system_prompt`: Change the system prompt (initial prompt of all slots). Please note that this does not take into account the chat template. It will append the prompt at the beginning of formatted prompt.
553+
- None yet
558554

559555
### POST `/v1/chat/completions`: OpenAI-compatible Chat Completions API
560556

0 commit comments

Comments
 (0)