We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b66d4df commit a13ec57Copy full SHA for a13ec57
src/llama.cpp
@@ -4273,13 +4273,9 @@ namespace GGUFMeta {
4273
4274
using llama_buf_map = std::unordered_map<uint32_t, ggml_backend_buffer_t>;
4275
4276
-// TODO: update when needed or think of some clever automatic way to do this
4277
-static size_t llama_model_max_nodes(const llama_model & /*model*/) {
4278
- //if (model.arch == LLM_ARCH_LLAMA && model.hparams.n_layer > ??) { // llama-3 405B
4279
- // return 32768;
4280
- //}
+static size_t llama_model_max_nodes(const llama_model & model) {
+ return std::max<size_t>(8192, model.tensors_by_name.size()*5);
4281
4282
- return 65536;
4283
}
4284
4285
struct llama_model_loader {
0 commit comments