You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+2-1Lines changed: 2 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -74,4 +74,5 @@ In no particular order of importance, these are the things that bother me:
74
74
- The args thing is cute, but it might be easier as a yaml file. Key value pairs. Flags can be keys with null values.
75
75
- The `--ctx-size` parameter doesn't seem quite right given that new models have the training (or max) context size in their metadata. That size should be used subject to a maximum in a passed parameter. E.g. So a 128K model can run comfortably on a smaller device.
76
76
- Write docs for a Deploying step. It should address the args file, removing the extra executable depending on platform, models, host, port. context size.
77
-
- Make a `.gitattributes` file so we can set the default file to be displayed and keep the README.md from llama.cpp. This will help in syncing changes continually from upstream. Reference: https://git-scm.com/docs/gitattributes
77
+
-~~Make a `.gitattributes` file so we can set the default file to be displayed and keep the README.md from llama.cpp. This will help in syncing changes continually from upstream. Reference: https://git-scm.com/docs/gitattributes~~ -- This doesn't actually work.
78
+
- Cosmo needs libssl and libcrypto. Building these from scratch gets an error about Cosco not liking assembly files. Sort this out.
printf "\n**********\n*\n* FINISHED: Add Certs to Archive.\n*\n**********\n\n"
76
+
```
77
+
78
+
#### Verify certs Directory in Archive
79
+
80
+
Verify that the archive has your certs:
81
+
```
82
+
unzip -l $LLAMA_SERVER_ONE_ZIP
83
+
printf "\n**********\n*\n* FINISHED: Verify certs Directory in Archive.\n*\n**********\n\n"
84
+
```
85
+
66
86
---
67
87
### Create website Directory in Archive
68
88
69
89
`llama.cpp` has a built in chat UI. If you'd like to provide a custom UI, you should add a `website` directory to the `llama-server-one` archive. `llama.cpp`'s chat UI is optimized for serving inside the project's source code. But we can copy the unoptimized source:
printf "\n**********\n*\n* FINISHED: Create website Directory in Archive.\n*\n**********\n\n"
78
99
```
@@ -84,8 +105,9 @@ Verify that the archive has your website:
84
105
unzip -l $LLAMA_SERVER_ONE_ZIP
85
106
printf "\n**********\n*\n* FINISHED: Verify website Directory in Archive.\n*\n**********\n\n"
86
107
```
108
+
87
109
---
88
-
### Create default-args File
110
+
### Create default-args File in Archive
89
111
90
112
A `default-args` file in the archive can specify sane default parameters. The format of the file is parameter name on a line, parameter value on a line, rinse, repeat. End the file with a `...` line to include user specified parameters.
0 commit comments