Skip to content

Commit fe35f72

Browse files
committed
Remove unused debug code
1 parent 1dffdae commit fe35f72

File tree

2 files changed

+2
-28
lines changed

2 files changed

+2
-28
lines changed

ggml/src/ggml-backend.cpp

Lines changed: 1 addition & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -243,15 +243,6 @@ void ggml_backend_tensor_set_async(ggml_backend_t backend, struct ggml_tensor *
243243
}
244244

245245
void ggml_backend_tensor_get_async(ggml_backend_t backend, const struct ggml_tensor * tensor, void * data, size_t offset, size_t size) {
246-
if (tensor->data == NULL) {
247-
// For output tensors that may not have been properly allocated
248-
if (tensor->flags & GGML_TENSOR_FLAG_OUTPUT) {
249-
fprintf(stderr, " Output tensor detected - this may indicate scheduling issue\n");
250-
// Return zeros for now to prevent crash
251-
memset(data, 0, size);
252-
return;
253-
}
254-
}
255246
GGML_ASSERT(tensor->data != NULL && "tensor not allocated");
256247
GGML_ASSERT(offset + size <= ggml_nbytes(tensor) && "tensor read out of bounds");
257248

@@ -270,14 +261,6 @@ void ggml_backend_tensor_set(struct ggml_tensor * tensor, const void * data, siz
270261
return;
271262
}
272263

273-
if (buf == NULL) {
274-
// For input tensors, buffer allocation may happen later by the scheduler
275-
if (tensor->flags & GGML_TENSOR_FLAG_INPUT) {
276-
// fprintf(stderr, "WARNING: Skipping tensor_set for input tensor '%s' - buffer will be allocated by scheduler\n",
277-
// tensor->name ? tensor->name : "unnamed");
278-
return;
279-
}
280-
}
281264
GGML_ASSERT(buf != NULL && "tensor buffer not set");
282265
GGML_ASSERT(tensor->data != NULL && "tensor not allocated");
283266
GGML_ASSERT(offset + size <= ggml_nbytes(tensor) && "tensor write out of bounds");
@@ -1666,17 +1649,7 @@ void ggml_backend_sched_set_tensor_backend(ggml_backend_sched_t sched, struct gg
16661649
ggml_backend_t ggml_backend_sched_get_tensor_backend(ggml_backend_sched_t sched, struct ggml_tensor * node) {
16671650
int backend_index = tensor_backend_id(node);
16681651
if (backend_index == -1) {
1669-
// Try to assign to CPU backend as fallback for output tensors
1670-
if (node->flags & GGML_TENSOR_FLAG_OUTPUT) {
1671-
fprintf(stderr, " Attempting to assign output tensor to CPU backend\n");
1672-
backend_index = sched->n_backends - 1; // CPU backend
1673-
tensor_backend_id(node) = backend_index;
1674-
SET_CAUSE(node, "out.cpu");
1675-
}
1676-
1677-
if (backend_index == -1) {
1678-
return NULL;
1679-
}
1652+
return NULL;
16801653
}
16811654
return sched->backends[backend_index];
16821655
}

src/llama-context.cpp

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,7 @@
99

1010
#include "ggml.h"
1111

12+
#include <cmath>
1213
#include <cinttypes>
1314
#include <cstring>
1415
#include <limits>

0 commit comments

Comments
 (0)