You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: CONTRIBUTING.md
+1Lines changed: 1 addition & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,6 +5,7 @@
5
5
- Execute [the full CI locally on your machine](ci/README.md) before publishing
6
6
- Please rate the complexity of your PR (i.e. `Review Complexity : Low`, `Review Complexity : Medium`, `Review Complexity : High`). This makes it easier for maintainers to triage the PRs.
7
7
- The PR template has a series of review complexity checkboxes `[ ]` that [you can mark as](https://docs.github.com/en/get-started/writing-on-github/working-with-advanced-formatting/about-task-lists)`[X]` for your convenience
8
+
- Consider allowing write access to your branch for faster review
8
9
- If your PR becomes stale, don't hesitate to ping the maintainers in the comments
"how much the prompt of a request must match the prompt of a slot in order to use that slot (default: %.2f, 0.0 = disabled)\n", params.slot_prompt_similarity });
1667
+
options.push_back({ "server", " --lora-init-without-apply", "load LoRA adapters without applying them (apply later via POST /lora-adapters) (default: %s)", params.lora_init_without_apply ? "enabled" : "disabled"});
std::vector<std::tuple<std::string, float>> lora_adapter; // lora adapter path with user defined scale
138
+
bool lora_init_without_apply = false; //only load lora to memory, but do not apply it to ctx (user can manually apply lora later using llama_lora_adapter_apply)
139
+
std::vector<llama_lora_adapter_info> lora_adapters; // lora adapter path with user defined scale
131
140
132
141
std::vector<llama_control_vector_load_info> control_vectors; // control vector with user defined scale
0 commit comments