|
14 | 14 | // #include <utility> |
15 | 15 | // #include <vector> |
16 | 16 |
|
17 | | -// #define DIRECT_OR_POOL_BY(what, threshold, var, fn) \ |
18 | | -// do { \ |
19 | | -// if ((what) < (threshold)) { \ |
20 | | -// for (int i = 0; i < (var); i++) { \ |
21 | | -// (fn)(i); \ |
22 | | -// } \ |
23 | | -// } else { \ |
24 | | -// pool->do_work_stealing_job((var), nullptr, (fn), nullptr); \ |
25 | | -// } \ |
| 17 | +// #define DIRECT_OR_POOL_BY(what, threshold, var, fn) \ |
| 18 | +// do { \ |
| 19 | +// if ((what) < (threshold)) { \ |
| 20 | +// for (int i = 0; i < (var); i++) { \ |
| 21 | +// (fn)(i); \ |
| 22 | +// } \ |
| 23 | +// } else { \ |
| 24 | +// pool->do_work_stealing_job((var), nullptr, (fn), nullptr); \ |
| 25 | +// } \ |
26 | 26 | // } while (0) |
27 | 27 |
|
28 | 28 | // #define VEC_DOT_TYPE(type) (ggml_internal_get_type_traits((ggml_type)(type)).vec_dot_type) |
|
31 | 31 | // #define QUANT_OFFSET(ptr, type, n, n_elements) \ |
32 | 32 | // (offset_pointer((ptr), (size_t)(n) * QUANT_BLCK_SIZE((n_elements), (type)))) |
33 | 33 |
|
34 | | -// #define LLAMAFILE_SGEMM_QUANT_FULL_MATMUL(m, n, k, a, a_type, b, b_col, c, c_col) \ |
35 | | -// do { \ |
36 | | -// llamafile_sgemm((m), (n), QUANT_BLCK_COUNT((k), (a_type)), (a), QUANT_BLCK_COUNT((k), (a_type)), \ |
37 | | -// QUANT_OFFSET((b), VEC_DOT_TYPE((a_type)), (b_col), (k)), \ |
38 | | -// QUANT_BLCK_COUNT((k), VEC_DOT_TYPE((a_type))), offset_pointer((c), (c_col) * (m) * sizeof(float)), \ |
39 | | -// (k), 0, 1, GGML_TASK_TYPE_COMPUTE, (a_type), VEC_DOT_TYPE((a_type)), GGML_TYPE_F32, \ |
40 | | -// GGML_PREC_DEFAULT); \ |
| 34 | +// #define LLAMAFILE_SGEMM_QUANT_FULL_MATMUL(m, n, k, a, a_type, b, b_col, c, c_col) \ |
| 35 | +// do { \ |
| 36 | +// llamafile_sgemm((m), (n), QUANT_BLCK_COUNT((k), (a_type)), (a), QUANT_BLCK_COUNT((k), (a_type)), \ |
| 37 | +// QUANT_OFFSET((b), VEC_DOT_TYPE((a_type)), (b_col), (k)), \ |
| 38 | +// QUANT_BLCK_COUNT((k), VEC_DOT_TYPE((a_type))), offset_pointer((c), (c_col) * (m) * |
| 39 | +// sizeof(float)), \ |
| 40 | +// (k), 0, 1, GGML_TASK_TYPE_COMPUTE, (a_type), VEC_DOT_TYPE((a_type)), GGML_TYPE_F32, \ |
| 41 | +// GGML_PREC_DEFAULT); \ |
41 | 42 | // } while (0) |
42 | 43 |
|
43 | | -// #define LLAMAFILE_SGEMM_MATMUL_F32(m, n, k, a, lda, b, ldb, c, ldc) \ |
44 | | -// do { \ |
45 | | -// llamafile_sgemm((m), (n), (k), (a), (lda), (b), (ldb), (c), (ldc), 0, 1, GGML_TASK_TYPE_COMPUTE, GGML_TYPE_F32, \ |
46 | | -// GGML_TYPE_F32, GGML_TYPE_F32, GGML_PREC_DEFAULT); \ |
| 44 | +// #define LLAMAFILE_SGEMM_MATMUL_F32(m, n, k, a, lda, b, ldb, c, ldc) \ |
| 45 | +// do { \ |
| 46 | +// llamafile_sgemm((m), (n), (k), (a), (lda), (b), (ldb), (c), (ldc), 0, 1, GGML_TASK_TYPE_COMPUTE, GGML_TYPE_F32, \ |
| 47 | +// GGML_TYPE_F32, GGML_TYPE_F32, GGML_PREC_DEFAULT); \ |
47 | 48 | // } while (0) |
48 | 49 |
|
49 | 50 | // // bool decide_absorb(size_t a,int a_type,size_t b,int b_type,size_t c,int c_type,size_t d,int d_type){ |
|
0 commit comments