@@ -674,7 +674,7 @@ struct ggml_backend_sched {
674674    char  * context_buffer;
675675    size_t  context_buffer_size;
676676
677-     bool  disable_op_offload ;
677+     bool  op_offload ;
678678
679679    int  debug;
680680};
@@ -768,7 +768,7 @@ static int ggml_backend_sched_backend_id_from_cur(ggml_backend_sched_t sched, st
768768        if  (tensor->op  != GGML_OP_ROPE && src->buffer  != NULL  && src->buffer ->usage  == GGML_BACKEND_BUFFER_USAGE_WEIGHTS) {
769769            int  src_backend_id = ggml_backend_sched_backend_from_buffer (sched, src, tensor);
770770            //  check if a backend with higher prio wants to offload the op
771-             if  (! sched->disable_op_offload  && src_backend_id == sched->n_backends  - 1  && ggml_backend_buffer_is_host (src->buffer )) {
771+             if  (sched->op_offload  && src_backend_id == sched->n_backends  - 1  && ggml_backend_buffer_is_host (src->buffer )) {
772772                for  (int  b = 0 ; b < src_backend_id; b++) {
773773                    if  (ggml_backend_supports_op (sched->backends [b], tensor) && ggml_backend_offload_op (sched->backends [b], tensor)) {
774774                        SET_CAUSE (tensor, " 1.off"  );
@@ -1455,7 +1455,7 @@ ggml_backend_sched_t ggml_backend_sched_new(
14551455        int  n_backends,
14561456        size_t  graph_size,
14571457        bool  parallel,
1458-         bool  disable_op_offload ) {
1458+         bool  op_offload ) {
14591459    GGML_ASSERT (n_backends > 0 );
14601460    GGML_ASSERT (n_backends <= GGML_SCHED_MAX_BACKENDS);
14611461    GGML_ASSERT (ggml_backend_dev_type (ggml_backend_get_device (backends[n_backends - 1 ])) == GGML_BACKEND_DEVICE_TYPE_CPU);
@@ -1500,7 +1500,7 @@ ggml_backend_sched_t ggml_backend_sched_new(
15001500    }
15011501
15021502    sched->galloc  = ggml_gallocr_new_n (sched->bufts , n_backends);
1503-     sched->disable_op_offload  = disable_op_offload ;
1503+     sched->op_offload  = op_offload ;
15041504
15051505    ggml_backend_sched_reset (sched);
15061506
0 commit comments