You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+5Lines changed: 5 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -12,3 +12,8 @@ If you care, **please contribute to [this discussion](https://github.com/ggergan
12
12
- Leave main.cpp UNTOUCHED, We want to be able to update the repo and pull any changes automatically.
13
13
- No dynamic memory allocation! Setup structs with FIXED (known) shapes and sizes for ALL output fields. Python will ALWAYS provide the memory, we just write to it.
14
14
- No external libraries or dependencies. That means no Flask, Pybind and whatever. All You Need Is Python.
15
+
16
+
## Usage
17
+
- Windows binaries are provided in the form of **llamalib.dll** but if you feel worried go ahead and rebuild it yourself.
18
+
- Weights are not included, you can use the llama.cpp quantize.exe to generate them from your official weight files (or download them from...places).
19
+
- To run, simply clone the repo and run `llama_for_kobold.py [ggml_quant_model.bin] [port]`, and then connect with Kobold or Kobold Lite.
0 commit comments