Skip to content

Commit 78981b8

Browse files
committed
Remove debug-associated and other misceallanous changes
1 parent 8783f42 commit 78981b8

File tree

6 files changed

+2
-15
lines changed

6 files changed

+2
-15
lines changed

common/arg.cpp

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1951,13 +1951,6 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
19511951
params.kv_unified = true;
19521952
}
19531953
).set_env("LLAMA_ARG_KV_SPLIT"));
1954-
add_opt(common_arg(
1955-
{"--dump-cache"},
1956-
"dump cache statistics after each token generation",
1957-
[](common_params & params) {
1958-
params.dump_cache = true;
1959-
}
1960-
).set_examples({LLAMA_EXAMPLE_MAIN}));
19611954
add_opt(common_arg(
19621955
{"--no-context-shift"},
19631956
string_format("disables context shift on infinite text generation (default: %s)", params.ctx_shift ? "disabled" : "enabled"),

common/common.h

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -399,8 +399,6 @@ struct common_params {
399399
ggml_type cache_type_k = GGML_TYPE_F16; // KV cache data type for the K
400400
ggml_type cache_type_v = GGML_TYPE_F16; // KV cache data type for the V
401401

402-
bool dump_cache = false; // dump cache statistics after each token
403-
404402
common_conversation_mode conversation_mode = COMMON_CONVERSATION_MODE_AUTO;
405403

406404
// multimodal models (see tools/mtmd)

examples/eval-callback/eval-callback.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -154,7 +154,7 @@ static bool ggml_debug(struct ggml_tensor * t, bool ask, void * user_data) {
154154

155155
if (!ggml_is_quantized(t->type)) {
156156
uint8_t * data = is_host ? (uint8_t *) t->data : cb_data->data.data();
157-
ggml_print_tensor(data, t->type, t->ne, t->nb, 8);
157+
ggml_print_tensor(data, t->type, t->ne, t->nb, 3);
158158
}
159159

160160
return true;

examples/model-conversion/qwen3stories.sh

Lines changed: 0 additions & 3 deletions
This file was deleted.

pyrightconfig.json

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,6 @@
66
"reportDuplicateImport": "error",
77
"reportDeprecated": "warning",
88
"reportUnnecessaryTypeIgnoreComment": "information",
9-
"reportAttributeAccessIssue": "warning",
109
"disableBytesTypePromotions": false, // TODO: change once Python 3.12 is the minimum
1110
"executionEnvironments": [
1211
{

src/llama-context.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1362,7 +1362,7 @@ void llama_context::output_reorder() {
13621362
//
13631363

13641364
uint32_t llama_context::graph_max_nodes() const {
1365-
return std::max<uint32_t>(16384, 512u*model.n_tensors());
1365+
return std::max<uint32_t>(8192, 128u*model.n_tensors());
13661366
}
13671367

13681368
llm_graph_result * llama_context::get_gf_res_reserve() const {

0 commit comments

Comments
 (0)